11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * S390 version 3a53c8fabSHeiko Carstens * Copyright IBM Corp. 1999 41da177e4SLinus Torvalds * Author(s): Hartmut Penner (hp@de.ibm.com) 51da177e4SLinus Torvalds * 61da177e4SLinus Torvalds * Derived from "arch/i386/mm/init.c" 71da177e4SLinus Torvalds * Copyright (C) 1995 Linus Torvalds 81da177e4SLinus Torvalds */ 91da177e4SLinus Torvalds 101da177e4SLinus Torvalds #include <linux/signal.h> 111da177e4SLinus Torvalds #include <linux/sched.h> 121da177e4SLinus Torvalds #include <linux/kernel.h> 131da177e4SLinus Torvalds #include <linux/errno.h> 141da177e4SLinus Torvalds #include <linux/string.h> 151da177e4SLinus Torvalds #include <linux/types.h> 161da177e4SLinus Torvalds #include <linux/ptrace.h> 171da177e4SLinus Torvalds #include <linux/mman.h> 181da177e4SLinus Torvalds #include <linux/mm.h> 191da177e4SLinus Torvalds #include <linux/swap.h> 201da177e4SLinus Torvalds #include <linux/smp.h> 211da177e4SLinus Torvalds #include <linux/init.h> 221da177e4SLinus Torvalds #include <linux/pagemap.h> 231da177e4SLinus Torvalds #include <linux/bootmem.h> 24e5d709bbSHeiko Carstens #include <linux/memory.h> 25d882b172SHeiko Carstens #include <linux/pfn.h> 26028d9b3cSHeiko Carstens #include <linux/poison.h> 272b67fc46SHeiko Carstens #include <linux/initrd.h> 283a4c5d59SHeiko Carstens #include <linux/export.h> 295a0e3ad6STejun Heo #include <linux/gfp.h> 301da177e4SLinus Torvalds #include <asm/processor.h> 311da177e4SLinus Torvalds #include <asm/uaccess.h> 321da177e4SLinus Torvalds #include <asm/pgtable.h> 331da177e4SLinus Torvalds #include <asm/pgalloc.h> 341da177e4SLinus Torvalds #include <asm/dma.h> 351da177e4SLinus Torvalds #include <asm/lowcore.h> 361da177e4SLinus Torvalds #include <asm/tlb.h> 371da177e4SLinus Torvalds #include <asm/tlbflush.h> 38d882b172SHeiko Carstens #include <asm/sections.h> 39a0616cdeSDavid Howells #include <asm/ctl_reg.h> 40e5d709bbSHeiko Carstens #include <asm/sclp.h> 411da177e4SLinus Torvalds 421da177e4SLinus Torvalds pgd_t swapper_pg_dir[PTRS_PER_PGD] __attribute__((__aligned__(PAGE_SIZE))); 431485c5c8SHeiko Carstens 44238ec4efSMartin Schwidefsky unsigned long empty_zero_page, zero_page_mask; 451485c5c8SHeiko Carstens EXPORT_SYMBOL(empty_zero_page); 460b70068eSArd Biesheuvel EXPORT_SYMBOL(zero_page_mask); 471da177e4SLinus Torvalds 480999f119SJiang Liu static void __init setup_zero_pages(void) 49238ec4efSMartin Schwidefsky { 50238ec4efSMartin Schwidefsky struct cpuid cpu_id; 51238ec4efSMartin Schwidefsky unsigned int order; 52238ec4efSMartin Schwidefsky struct page *page; 53238ec4efSMartin Schwidefsky int i; 54238ec4efSMartin Schwidefsky 55238ec4efSMartin Schwidefsky get_cpu_id(&cpu_id); 56238ec4efSMartin Schwidefsky switch (cpu_id.machine) { 57238ec4efSMartin Schwidefsky case 0x9672: /* g5 */ 58238ec4efSMartin Schwidefsky case 0x2064: /* z900 */ 59238ec4efSMartin Schwidefsky case 0x2066: /* z900 */ 60238ec4efSMartin Schwidefsky case 0x2084: /* z990 */ 61238ec4efSMartin Schwidefsky case 0x2086: /* z990 */ 62238ec4efSMartin Schwidefsky case 0x2094: /* z9-109 */ 63238ec4efSMartin Schwidefsky case 0x2096: /* z9-109 */ 64238ec4efSMartin Schwidefsky order = 0; 65238ec4efSMartin Schwidefsky break; 66238ec4efSMartin Schwidefsky case 0x2097: /* z10 */ 67238ec4efSMartin Schwidefsky case 0x2098: /* z10 */ 687919e91bSMartin Schwidefsky case 0x2817: /* z196 */ 697919e91bSMartin Schwidefsky case 0x2818: /* z196 */ 70238ec4efSMartin Schwidefsky order = 2; 71238ec4efSMartin Schwidefsky break; 727919e91bSMartin Schwidefsky case 0x2827: /* zEC12 */ 7359471227SHeiko Carstens case 0x2828: /* zEC12 */ 747919e91bSMartin Schwidefsky default: 757919e91bSMartin Schwidefsky order = 5; 767919e91bSMartin Schwidefsky break; 77238ec4efSMartin Schwidefsky } 787919e91bSMartin Schwidefsky /* Limit number of empty zero pages for small memory sizes */ 797919e91bSMartin Schwidefsky if (order > 2 && totalram_pages <= 16384) 807919e91bSMartin Schwidefsky order = 2; 81238ec4efSMartin Schwidefsky 82238ec4efSMartin Schwidefsky empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); 83238ec4efSMartin Schwidefsky if (!empty_zero_page) 84238ec4efSMartin Schwidefsky panic("Out of memory in setup_zero_pages"); 85238ec4efSMartin Schwidefsky 86238ec4efSMartin Schwidefsky page = virt_to_page((void *) empty_zero_page); 87238ec4efSMartin Schwidefsky split_page(page, order); 88238ec4efSMartin Schwidefsky for (i = 1 << order; i > 0; i--) { 890999f119SJiang Liu mark_page_reserved(page); 90238ec4efSMartin Schwidefsky page++; 91238ec4efSMartin Schwidefsky } 92238ec4efSMartin Schwidefsky 930999f119SJiang Liu zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; 94238ec4efSMartin Schwidefsky } 95238ec4efSMartin Schwidefsky 961da177e4SLinus Torvalds /* 971da177e4SLinus Torvalds * paging_init() sets up the page tables 981da177e4SLinus Torvalds */ 991da177e4SLinus Torvalds void __init paging_init(void) 1001da177e4SLinus Torvalds { 10139b742f9SHeiko Carstens unsigned long max_zone_pfns[MAX_NR_ZONES]; 10214045ebfSMartin Schwidefsky unsigned long pgd_type, asce_bits; 1031da177e4SLinus Torvalds 1043610cce8SMartin Schwidefsky init_mm.pgd = swapper_pg_dir; 105f4eb07c1SHeiko Carstens #ifdef CONFIG_64BIT 10614045ebfSMartin Schwidefsky if (VMALLOC_END > (1UL << 42)) { 10714045ebfSMartin Schwidefsky asce_bits = _ASCE_TYPE_REGION2 | _ASCE_TABLE_LENGTH; 10814045ebfSMartin Schwidefsky pgd_type = _REGION2_ENTRY_EMPTY; 10914045ebfSMartin Schwidefsky } else { 11014045ebfSMartin Schwidefsky asce_bits = _ASCE_TYPE_REGION3 | _ASCE_TABLE_LENGTH; 1116252d702SMartin Schwidefsky pgd_type = _REGION3_ENTRY_EMPTY; 11214045ebfSMartin Schwidefsky } 113f4eb07c1SHeiko Carstens #else 11414045ebfSMartin Schwidefsky asce_bits = _ASCE_TABLE_LENGTH; 1153610cce8SMartin Schwidefsky pgd_type = _SEGMENT_ENTRY_EMPTY; 116f4eb07c1SHeiko Carstens #endif 11714045ebfSMartin Schwidefsky S390_lowcore.kernel_asce = (__pa(init_mm.pgd) & PAGE_MASK) | asce_bits; 1183610cce8SMartin Schwidefsky clear_table((unsigned long *) init_mm.pgd, pgd_type, 1193610cce8SMartin Schwidefsky sizeof(unsigned long)*2048); 120f4eb07c1SHeiko Carstens vmem_map_init(); 1211da177e4SLinus Torvalds 1221da177e4SLinus Torvalds /* enable virtual mapping in kernel mode */ 1233610cce8SMartin Schwidefsky __ctl_load(S390_lowcore.kernel_asce, 1, 1); 1243610cce8SMartin Schwidefsky __ctl_load(S390_lowcore.kernel_asce, 7, 7); 1253610cce8SMartin Schwidefsky __ctl_load(S390_lowcore.kernel_asce, 13, 13); 126df9ee292SDavid Howells arch_local_irq_restore(4UL << (BITS_PER_LONG - 8)); 1271da177e4SLinus Torvalds 12817f34580SHeiko Carstens sparse_memory_present_with_active_regions(MAX_NUMNODES); 12917f34580SHeiko Carstens sparse_init(); 13039b742f9SHeiko Carstens memset(max_zone_pfns, 0, sizeof(max_zone_pfns)); 13139b742f9SHeiko Carstens max_zone_pfns[ZONE_DMA] = PFN_DOWN(MAX_DMA_ADDRESS); 13239b742f9SHeiko Carstens max_zone_pfns[ZONE_NORMAL] = max_low_pfn; 13339b742f9SHeiko Carstens free_area_init_nodes(max_zone_pfns); 1341da177e4SLinus Torvalds } 1351da177e4SLinus Torvalds 1361da177e4SLinus Torvalds void __init mem_init(void) 1371da177e4SLinus Torvalds { 1381b948d6cSMartin Schwidefsky if (MACHINE_HAS_TLB_LC) 1391b948d6cSMartin Schwidefsky cpumask_set_cpu(0, &init_mm.context.cpu_attach_mask); 1401b948d6cSMartin Schwidefsky cpumask_set_cpu(0, mm_cpumask(&init_mm)); 1411b948d6cSMartin Schwidefsky atomic_set(&init_mm.context.attach_count, 1); 1421b948d6cSMartin Schwidefsky 143a18d0e2dSJiang Liu max_mapnr = max_low_pfn; 1441da177e4SLinus Torvalds high_memory = (void *) __va(max_low_pfn * PAGE_SIZE); 1451da177e4SLinus Torvalds 14645e576b1SMartin Schwidefsky /* Setup guest page hinting */ 14745e576b1SMartin Schwidefsky cmma_init(); 14845e576b1SMartin Schwidefsky 1491da177e4SLinus Torvalds /* this will put all low memory onto the freelists */ 1500c988534SJiang Liu free_all_bootmem(); 1510999f119SJiang Liu setup_zero_pages(); /* Setup zeroed pages. */ 1521da177e4SLinus Torvalds 153a18d0e2dSJiang Liu mem_init_print_info(NULL); 154d882b172SHeiko Carstens printk("Write protected kernel read-only data: %#lx - %#lx\n", 155162e006eSHeiko Carstens (unsigned long)&_stext, 156162e006eSHeiko Carstens PFN_ALIGN((unsigned long)&_eshared) - 1); 1571da177e4SLinus Torvalds } 1581da177e4SLinus Torvalds 159d96221abSHeiko Carstens void free_initmem(void) 160d96221abSHeiko Carstens { 161dbe67df4SJiang Liu free_initmem_default(POISON_FREE_INITMEM); 1621da177e4SLinus Torvalds } 1631da177e4SLinus Torvalds 1641da177e4SLinus Torvalds #ifdef CONFIG_BLK_DEV_INITRD 1655e249d6eSHeiko Carstens void __init free_initrd_mem(unsigned long start, unsigned long end) 1661da177e4SLinus Torvalds { 16711199692SJiang Liu free_reserved_area((void *)start, (void *)end, POISON_FREE_INITMEM, 16811199692SJiang Liu "initrd"); 1691da177e4SLinus Torvalds } 1701da177e4SLinus Torvalds #endif 171421c175cSHeiko Carstens 172421c175cSHeiko Carstens #ifdef CONFIG_MEMORY_HOTPLUG 173421c175cSHeiko Carstens int arch_add_memory(int nid, u64 start, u64 size) 174421c175cSHeiko Carstens { 175892365abSGerald Schaefer unsigned long zone_start_pfn, zone_end_pfn, nr_pages; 176892365abSGerald Schaefer unsigned long start_pfn = PFN_DOWN(start); 177892365abSGerald Schaefer unsigned long size_pages = PFN_DOWN(size); 178421c175cSHeiko Carstens struct zone *zone; 179421c175cSHeiko Carstens int rc; 180421c175cSHeiko Carstens 181421c175cSHeiko Carstens rc = vmem_add_mapping(start, size); 182421c175cSHeiko Carstens if (rc) 183421c175cSHeiko Carstens return rc; 184892365abSGerald Schaefer for_each_zone(zone) { 185892365abSGerald Schaefer if (zone_idx(zone) != ZONE_MOVABLE) { 186892365abSGerald Schaefer /* Add range within existing zone limits */ 187892365abSGerald Schaefer zone_start_pfn = zone->zone_start_pfn; 188892365abSGerald Schaefer zone_end_pfn = zone->zone_start_pfn + 189892365abSGerald Schaefer zone->spanned_pages; 190892365abSGerald Schaefer } else { 191892365abSGerald Schaefer /* Add remaining range to ZONE_MOVABLE */ 192892365abSGerald Schaefer zone_start_pfn = start_pfn; 193892365abSGerald Schaefer zone_end_pfn = start_pfn + size_pages; 194892365abSGerald Schaefer } 195892365abSGerald Schaefer if (start_pfn < zone_start_pfn || start_pfn >= zone_end_pfn) 196892365abSGerald Schaefer continue; 197892365abSGerald Schaefer nr_pages = (start_pfn + size_pages > zone_end_pfn) ? 198892365abSGerald Schaefer zone_end_pfn - start_pfn : size_pages; 199892365abSGerald Schaefer rc = __add_pages(nid, zone, start_pfn, nr_pages); 200892365abSGerald Schaefer if (rc) 201892365abSGerald Schaefer break; 202892365abSGerald Schaefer start_pfn += nr_pages; 203892365abSGerald Schaefer size_pages -= nr_pages; 204892365abSGerald Schaefer if (!size_pages) 205892365abSGerald Schaefer break; 206892365abSGerald Schaefer } 207421c175cSHeiko Carstens if (rc) 208421c175cSHeiko Carstens vmem_remove_mapping(start, size); 209421c175cSHeiko Carstens return rc; 210421c175cSHeiko Carstens } 21124d335caSWen Congyang 212e5d709bbSHeiko Carstens unsigned long memory_block_size_bytes(void) 213e5d709bbSHeiko Carstens { 214e5d709bbSHeiko Carstens /* 215e5d709bbSHeiko Carstens * Make sure the memory block size is always greater 216e5d709bbSHeiko Carstens * or equal than the memory increment size. 217e5d709bbSHeiko Carstens */ 218e5d709bbSHeiko Carstens return max_t(unsigned long, MIN_MEMORY_BLOCK_SIZE, sclp_get_rzm()); 219e5d709bbSHeiko Carstens } 220e5d709bbSHeiko Carstens 22124d335caSWen Congyang #ifdef CONFIG_MEMORY_HOTREMOVE 22224d335caSWen Congyang int arch_remove_memory(u64 start, u64 size) 22324d335caSWen Congyang { 22424d335caSWen Congyang /* 22524d335caSWen Congyang * There is no hardware or firmware interface which could trigger a 22624d335caSWen Congyang * hot memory remove on s390. So there is nothing that needs to be 22724d335caSWen Congyang * implemented. 22824d335caSWen Congyang */ 22924d335caSWen Congyang return -EBUSY; 23024d335caSWen Congyang } 23124d335caSWen Congyang #endif 232421c175cSHeiko Carstens #endif /* CONFIG_MEMORY_HOTPLUG */ 233