11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * S390 version 3a53c8fabSHeiko Carstens * Copyright IBM Corp. 1999 41da177e4SLinus Torvalds * Author(s): Hartmut Penner (hp@de.ibm.com) 51da177e4SLinus Torvalds * 61da177e4SLinus Torvalds * Derived from "arch/i386/mm/init.c" 71da177e4SLinus Torvalds * Copyright (C) 1995 Linus Torvalds 81da177e4SLinus Torvalds */ 91da177e4SLinus Torvalds 101da177e4SLinus Torvalds #include <linux/signal.h> 111da177e4SLinus Torvalds #include <linux/sched.h> 121da177e4SLinus Torvalds #include <linux/kernel.h> 131da177e4SLinus Torvalds #include <linux/errno.h> 141da177e4SLinus Torvalds #include <linux/string.h> 151da177e4SLinus Torvalds #include <linux/types.h> 161da177e4SLinus Torvalds #include <linux/ptrace.h> 171da177e4SLinus Torvalds #include <linux/mman.h> 181da177e4SLinus Torvalds #include <linux/mm.h> 191da177e4SLinus Torvalds #include <linux/swap.h> 201da177e4SLinus Torvalds #include <linux/smp.h> 211da177e4SLinus Torvalds #include <linux/init.h> 221da177e4SLinus Torvalds #include <linux/pagemap.h> 231da177e4SLinus Torvalds #include <linux/bootmem.h> 24e5d709bbSHeiko Carstens #include <linux/memory.h> 25d882b172SHeiko Carstens #include <linux/pfn.h> 26028d9b3cSHeiko Carstens #include <linux/poison.h> 272b67fc46SHeiko Carstens #include <linux/initrd.h> 283a4c5d59SHeiko Carstens #include <linux/export.h> 295a0e3ad6STejun Heo #include <linux/gfp.h> 301da177e4SLinus Torvalds #include <asm/processor.h> 311da177e4SLinus Torvalds #include <asm/uaccess.h> 321da177e4SLinus Torvalds #include <asm/pgtable.h> 331da177e4SLinus Torvalds #include <asm/pgalloc.h> 341da177e4SLinus Torvalds #include <asm/dma.h> 351da177e4SLinus Torvalds #include <asm/lowcore.h> 361da177e4SLinus Torvalds #include <asm/tlb.h> 371da177e4SLinus Torvalds #include <asm/tlbflush.h> 38d882b172SHeiko Carstens #include <asm/sections.h> 39a0616cdeSDavid Howells #include <asm/ctl_reg.h> 40e5d709bbSHeiko Carstens #include <asm/sclp.h> 411da177e4SLinus Torvalds 421da177e4SLinus Torvalds pgd_t swapper_pg_dir[PTRS_PER_PGD] __attribute__((__aligned__(PAGE_SIZE))); 431485c5c8SHeiko Carstens 44238ec4efSMartin Schwidefsky unsigned long empty_zero_page, zero_page_mask; 451485c5c8SHeiko Carstens EXPORT_SYMBOL(empty_zero_page); 460b70068eSArd Biesheuvel EXPORT_SYMBOL(zero_page_mask); 471da177e4SLinus Torvalds 480999f119SJiang Liu static void __init setup_zero_pages(void) 49238ec4efSMartin Schwidefsky { 50238ec4efSMartin Schwidefsky struct cpuid cpu_id; 51238ec4efSMartin Schwidefsky unsigned int order; 52238ec4efSMartin Schwidefsky struct page *page; 53238ec4efSMartin Schwidefsky int i; 54238ec4efSMartin Schwidefsky 55238ec4efSMartin Schwidefsky get_cpu_id(&cpu_id); 56238ec4efSMartin Schwidefsky switch (cpu_id.machine) { 57238ec4efSMartin Schwidefsky case 0x9672: /* g5 */ 58238ec4efSMartin Schwidefsky case 0x2064: /* z900 */ 59238ec4efSMartin Schwidefsky case 0x2066: /* z900 */ 60238ec4efSMartin Schwidefsky case 0x2084: /* z990 */ 61238ec4efSMartin Schwidefsky case 0x2086: /* z990 */ 62238ec4efSMartin Schwidefsky case 0x2094: /* z9-109 */ 63238ec4efSMartin Schwidefsky case 0x2096: /* z9-109 */ 64238ec4efSMartin Schwidefsky order = 0; 65238ec4efSMartin Schwidefsky break; 66238ec4efSMartin Schwidefsky case 0x2097: /* z10 */ 67238ec4efSMartin Schwidefsky case 0x2098: /* z10 */ 687919e91bSMartin Schwidefsky case 0x2817: /* z196 */ 697919e91bSMartin Schwidefsky case 0x2818: /* z196 */ 70238ec4efSMartin Schwidefsky order = 2; 71238ec4efSMartin Schwidefsky break; 727919e91bSMartin Schwidefsky case 0x2827: /* zEC12 */ 7359471227SHeiko Carstens case 0x2828: /* zEC12 */ 747919e91bSMartin Schwidefsky order = 5; 757919e91bSMartin Schwidefsky break; 761f6b83e5SMartin Schwidefsky case 0x2964: /* z13 */ 771f6b83e5SMartin Schwidefsky default: 781f6b83e5SMartin Schwidefsky order = 7; 791f6b83e5SMartin Schwidefsky break; 80238ec4efSMartin Schwidefsky } 817919e91bSMartin Schwidefsky /* Limit number of empty zero pages for small memory sizes */ 821f6b83e5SMartin Schwidefsky while (order > 2 && (totalram_pages >> 10) < (1UL << order)) 831f6b83e5SMartin Schwidefsky order--; 84238ec4efSMartin Schwidefsky 85238ec4efSMartin Schwidefsky empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); 86238ec4efSMartin Schwidefsky if (!empty_zero_page) 87238ec4efSMartin Schwidefsky panic("Out of memory in setup_zero_pages"); 88238ec4efSMartin Schwidefsky 89238ec4efSMartin Schwidefsky page = virt_to_page((void *) empty_zero_page); 90238ec4efSMartin Schwidefsky split_page(page, order); 91238ec4efSMartin Schwidefsky for (i = 1 << order; i > 0; i--) { 920999f119SJiang Liu mark_page_reserved(page); 93238ec4efSMartin Schwidefsky page++; 94238ec4efSMartin Schwidefsky } 95238ec4efSMartin Schwidefsky 960999f119SJiang Liu zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; 97238ec4efSMartin Schwidefsky } 98238ec4efSMartin Schwidefsky 991da177e4SLinus Torvalds /* 1001da177e4SLinus Torvalds * paging_init() sets up the page tables 1011da177e4SLinus Torvalds */ 1021da177e4SLinus Torvalds void __init paging_init(void) 1031da177e4SLinus Torvalds { 10439b742f9SHeiko Carstens unsigned long max_zone_pfns[MAX_NR_ZONES]; 10514045ebfSMartin Schwidefsky unsigned long pgd_type, asce_bits; 1061da177e4SLinus Torvalds 1073610cce8SMartin Schwidefsky init_mm.pgd = swapper_pg_dir; 10814045ebfSMartin Schwidefsky if (VMALLOC_END > (1UL << 42)) { 10914045ebfSMartin Schwidefsky asce_bits = _ASCE_TYPE_REGION2 | _ASCE_TABLE_LENGTH; 11014045ebfSMartin Schwidefsky pgd_type = _REGION2_ENTRY_EMPTY; 11114045ebfSMartin Schwidefsky } else { 11214045ebfSMartin Schwidefsky asce_bits = _ASCE_TYPE_REGION3 | _ASCE_TABLE_LENGTH; 1136252d702SMartin Schwidefsky pgd_type = _REGION3_ENTRY_EMPTY; 11414045ebfSMartin Schwidefsky } 11514045ebfSMartin Schwidefsky S390_lowcore.kernel_asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits; 1163610cce8SMartin Schwidefsky clear_table((unsigned long *) init_mm.pgd, pgd_type, 1173610cce8SMartin Schwidefsky sizeof(unsigned long)*2048); 118f4eb07c1SHeiko Carstens vmem_map_init(); 1191da177e4SLinus Torvalds 1201da177e4SLinus Torvalds /* enable virtual mapping in kernel mode */ 1213610cce8SMartin Schwidefsky __ctl_load(S390_lowcore.kernel_asce, 1, 1); 1223610cce8SMartin Schwidefsky __ctl_load(S390_lowcore.kernel_asce, 7, 7); 1233610cce8SMartin Schwidefsky __ctl_load(S390_lowcore.kernel_asce, 13, 13); 124df9ee292SDavid Howells arch_local_irq_restore(4UL << (BITS_PER_LONG - 8)); 1251da177e4SLinus Torvalds 12617f34580SHeiko Carstens sparse_memory_present_with_active_regions(MAX_NUMNODES); 12717f34580SHeiko Carstens sparse_init(); 12839b742f9SHeiko Carstens memset(max_zone_pfns, 0, sizeof(max_zone_pfns)); 12939b742f9SHeiko Carstens max_zone_pfns[ZONE_DMA] = PFN_DOWN(MAX_DMA_ADDRESS); 13039b742f9SHeiko Carstens max_zone_pfns[ZONE_NORMAL] = max_low_pfn; 13139b742f9SHeiko Carstens free_area_init_nodes(max_zone_pfns); 1321da177e4SLinus Torvalds } 1331da177e4SLinus Torvalds 1341da177e4SLinus Torvalds void __init mem_init(void) 1351da177e4SLinus Torvalds { 1361b948d6cSMartin Schwidefsky if (MACHINE_HAS_TLB_LC) 1371b948d6cSMartin Schwidefsky cpumask_set_cpu(0, &init_mm.context.cpu_attach_mask); 1381b948d6cSMartin Schwidefsky cpumask_set_cpu(0, mm_cpumask(&init_mm)); 1391b948d6cSMartin Schwidefsky atomic_set(&init_mm.context.attach_count, 1); 1401b948d6cSMartin Schwidefsky 141a18d0e2dSJiang Liu max_mapnr = max_low_pfn; 1421da177e4SLinus Torvalds high_memory = (void *) __va(max_low_pfn * PAGE_SIZE); 1431da177e4SLinus Torvalds 14445e576b1SMartin Schwidefsky /* Setup guest page hinting */ 14545e576b1SMartin Schwidefsky cmma_init(); 14645e576b1SMartin Schwidefsky 1471da177e4SLinus Torvalds /* this will put all low memory onto the freelists */ 1480c988534SJiang Liu free_all_bootmem(); 1490999f119SJiang Liu setup_zero_pages(); /* Setup zeroed pages. */ 1501da177e4SLinus Torvalds 151a18d0e2dSJiang Liu mem_init_print_info(NULL); 152d882b172SHeiko Carstens printk("Write protected kernel read-only data: %#lx - %#lx\n", 153162e006eSHeiko Carstens (unsigned long)&_stext, 154162e006eSHeiko Carstens PFN_ALIGN((unsigned long)&_eshared) - 1); 1551da177e4SLinus Torvalds } 1561da177e4SLinus Torvalds 157d96221abSHeiko Carstens void free_initmem(void) 158d96221abSHeiko Carstens { 159dbe67df4SJiang Liu free_initmem_default(POISON_FREE_INITMEM); 1601da177e4SLinus Torvalds } 1611da177e4SLinus Torvalds 1621da177e4SLinus Torvalds #ifdef CONFIG_BLK_DEV_INITRD 1635e249d6eSHeiko Carstens void __init free_initrd_mem(unsigned long start, unsigned long end) 1641da177e4SLinus Torvalds { 16511199692SJiang Liu free_reserved_area((void *)start, (void *)end, POISON_FREE_INITMEM, 16611199692SJiang Liu "initrd"); 1671da177e4SLinus Torvalds } 1681da177e4SLinus Torvalds #endif 169421c175cSHeiko Carstens 170421c175cSHeiko Carstens #ifdef CONFIG_MEMORY_HOTPLUG 171421c175cSHeiko Carstens int arch_add_memory(int nid, u64 start, u64 size) 172421c175cSHeiko Carstens { 173892365abSGerald Schaefer unsigned long zone_start_pfn, zone_end_pfn, nr_pages; 174892365abSGerald Schaefer unsigned long start_pfn = PFN_DOWN(start); 175892365abSGerald Schaefer unsigned long size_pages = PFN_DOWN(size); 176421c175cSHeiko Carstens struct zone *zone; 177421c175cSHeiko Carstens int rc; 178421c175cSHeiko Carstens 179421c175cSHeiko Carstens rc = vmem_add_mapping(start, size); 180421c175cSHeiko Carstens if (rc) 181421c175cSHeiko Carstens return rc; 182892365abSGerald Schaefer for_each_zone(zone) { 183892365abSGerald Schaefer if (zone_idx(zone) != ZONE_MOVABLE) { 184892365abSGerald Schaefer /* Add range within existing zone limits */ 185892365abSGerald Schaefer zone_start_pfn = zone->zone_start_pfn; 186892365abSGerald Schaefer zone_end_pfn = zone->zone_start_pfn + 187892365abSGerald Schaefer zone->spanned_pages; 188892365abSGerald Schaefer } else { 189892365abSGerald Schaefer /* Add remaining range to ZONE_MOVABLE */ 190892365abSGerald Schaefer zone_start_pfn = start_pfn; 191892365abSGerald Schaefer zone_end_pfn = start_pfn + size_pages; 192892365abSGerald Schaefer } 193892365abSGerald Schaefer if (start_pfn < zone_start_pfn || start_pfn >= zone_end_pfn) 194892365abSGerald Schaefer continue; 195892365abSGerald Schaefer nr_pages = (start_pfn + size_pages > zone_end_pfn) ? 196892365abSGerald Schaefer zone_end_pfn - start_pfn : size_pages; 197892365abSGerald Schaefer rc = __add_pages(nid, zone, start_pfn, nr_pages); 198892365abSGerald Schaefer if (rc) 199892365abSGerald Schaefer break; 200892365abSGerald Schaefer start_pfn += nr_pages; 201892365abSGerald Schaefer size_pages -= nr_pages; 202892365abSGerald Schaefer if (!size_pages) 203892365abSGerald Schaefer break; 204892365abSGerald Schaefer } 205421c175cSHeiko Carstens if (rc) 206421c175cSHeiko Carstens vmem_remove_mapping(start, size); 207421c175cSHeiko Carstens return rc; 208421c175cSHeiko Carstens } 20924d335caSWen Congyang 210e5d709bbSHeiko Carstens unsigned long memory_block_size_bytes(void) 211e5d709bbSHeiko Carstens { 212e5d709bbSHeiko Carstens /* 213e5d709bbSHeiko Carstens * Make sure the memory block size is always greater 214e5d709bbSHeiko Carstens * or equal than the memory increment size. 215e5d709bbSHeiko Carstens */ 21637c5f6c8SDavid Hildenbrand return max_t(unsigned long, MIN_MEMORY_BLOCK_SIZE, sclp.rzm); 217e5d709bbSHeiko Carstens } 218e5d709bbSHeiko Carstens 21924d335caSWen Congyang #ifdef CONFIG_MEMORY_HOTREMOVE 22024d335caSWen Congyang int arch_remove_memory(u64 start, u64 size) 22124d335caSWen Congyang { 22224d335caSWen Congyang /* 22324d335caSWen Congyang * There is no hardware or firmware interface which could trigger a 22424d335caSWen Congyang * hot memory remove on s390. So there is nothing that needs to be 22524d335caSWen Congyang * implemented. 22624d335caSWen Congyang */ 22724d335caSWen Congyang return -EBUSY; 22824d335caSWen Congyang } 22924d335caSWen Congyang #endif 230421c175cSHeiko Carstens #endif /* CONFIG_MEMORY_HOTPLUG */ 231