11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * linux/arch/arm/mm/init.c 31da177e4SLinus Torvalds * 490072059SRussell King * Copyright (C) 1995-2005 Russell King 51da177e4SLinus Torvalds * 61da177e4SLinus Torvalds * This program is free software; you can redistribute it and/or modify 71da177e4SLinus Torvalds * it under the terms of the GNU General Public License version 2 as 81da177e4SLinus Torvalds * published by the Free Software Foundation. 91da177e4SLinus Torvalds */ 101da177e4SLinus Torvalds #include <linux/kernel.h> 111da177e4SLinus Torvalds #include <linux/errno.h> 121da177e4SLinus Torvalds #include <linux/ptrace.h> 131da177e4SLinus Torvalds #include <linux/swap.h> 141da177e4SLinus Torvalds #include <linux/init.h> 151da177e4SLinus Torvalds #include <linux/bootmem.h> 161da177e4SLinus Torvalds #include <linux/mman.h> 171da177e4SLinus Torvalds #include <linux/nodemask.h> 181da177e4SLinus Torvalds #include <linux/initrd.h> 191da177e4SLinus Torvalds 201da177e4SLinus Torvalds #include <asm/mach-types.h> 211da177e4SLinus Torvalds #include <asm/setup.h> 2274d02fb9SRussell King #include <asm/sizes.h> 231da177e4SLinus Torvalds #include <asm/tlb.h> 241da177e4SLinus Torvalds 251da177e4SLinus Torvalds #include <asm/mach/arch.h> 261da177e4SLinus Torvalds #include <asm/mach/map.h> 271da177e4SLinus Torvalds 281b2e2b73SRussell King #include "mm.h" 291b2e2b73SRussell King 301da177e4SLinus Torvalds DEFINE_PER_CPU(struct mmu_gather, mmu_gathers); 311da177e4SLinus Torvalds 321da177e4SLinus Torvalds extern pgd_t swapper_pg_dir[PTRS_PER_PGD]; 331da177e4SLinus Torvalds extern void _stext, _text, _etext, __data_start, _end, __init_begin, __init_end; 341da177e4SLinus Torvalds extern unsigned long phys_initrd_start; 351da177e4SLinus Torvalds extern unsigned long phys_initrd_size; 361da177e4SLinus Torvalds 371da177e4SLinus Torvalds /* 381da177e4SLinus Torvalds * The sole use of this is to pass memory configuration 391da177e4SLinus Torvalds * data from paging_init to mem_init. 401da177e4SLinus Torvalds */ 411da177e4SLinus Torvalds static struct meminfo meminfo __initdata = { 0, }; 421da177e4SLinus Torvalds 431da177e4SLinus Torvalds /* 441da177e4SLinus Torvalds * empty_zero_page is a special page that is used for 451da177e4SLinus Torvalds * zero-initialized data and COW. 461da177e4SLinus Torvalds */ 471da177e4SLinus Torvalds struct page *empty_zero_page; 481da177e4SLinus Torvalds 491b2e2b73SRussell King /* 501b2e2b73SRussell King * The pmd table for the upper-most set of pages. 511b2e2b73SRussell King */ 521b2e2b73SRussell King pmd_t *top_pmd; 531b2e2b73SRussell King 541da177e4SLinus Torvalds void show_mem(void) 551da177e4SLinus Torvalds { 561da177e4SLinus Torvalds int free = 0, total = 0, reserved = 0; 571da177e4SLinus Torvalds int shared = 0, cached = 0, slab = 0, node; 581da177e4SLinus Torvalds 591da177e4SLinus Torvalds printk("Mem-info:\n"); 601da177e4SLinus Torvalds show_free_areas(); 611da177e4SLinus Torvalds printk("Free swap: %6ldkB\n", nr_swap_pages<<(PAGE_SHIFT-10)); 621da177e4SLinus Torvalds 631da177e4SLinus Torvalds for_each_online_node(node) { 641da177e4SLinus Torvalds struct page *page, *end; 651da177e4SLinus Torvalds 661da177e4SLinus Torvalds page = NODE_MEM_MAP(node); 671da177e4SLinus Torvalds end = page + NODE_DATA(node)->node_spanned_pages; 681da177e4SLinus Torvalds 691da177e4SLinus Torvalds do { 701da177e4SLinus Torvalds total++; 711da177e4SLinus Torvalds if (PageReserved(page)) 721da177e4SLinus Torvalds reserved++; 731da177e4SLinus Torvalds else if (PageSwapCache(page)) 741da177e4SLinus Torvalds cached++; 751da177e4SLinus Torvalds else if (PageSlab(page)) 761da177e4SLinus Torvalds slab++; 771da177e4SLinus Torvalds else if (!page_count(page)) 781da177e4SLinus Torvalds free++; 791da177e4SLinus Torvalds else 801da177e4SLinus Torvalds shared += page_count(page) - 1; 811da177e4SLinus Torvalds page++; 821da177e4SLinus Torvalds } while (page < end); 831da177e4SLinus Torvalds } 841da177e4SLinus Torvalds 851da177e4SLinus Torvalds printk("%d pages of RAM\n", total); 861da177e4SLinus Torvalds printk("%d free pages\n", free); 871da177e4SLinus Torvalds printk("%d reserved pages\n", reserved); 881da177e4SLinus Torvalds printk("%d slab pages\n", slab); 891da177e4SLinus Torvalds printk("%d pages shared\n", shared); 901da177e4SLinus Torvalds printk("%d pages swap cached\n", cached); 911da177e4SLinus Torvalds } 921da177e4SLinus Torvalds 9390072059SRussell King #define for_each_nodebank(iter,mi,no) \ 9490072059SRussell King for (iter = 0; iter < mi->nr_banks; iter++) \ 9590072059SRussell King if (mi->bank[iter].node == no) 961da177e4SLinus Torvalds 971da177e4SLinus Torvalds /* 981da177e4SLinus Torvalds * FIXME: We really want to avoid allocating the bootmap bitmap 991da177e4SLinus Torvalds * over the top of the initrd. Hopefully, this is located towards 1001da177e4SLinus Torvalds * the start of a bank, so if we allocate the bootmap bitmap at 1011da177e4SLinus Torvalds * the end, we won't clash. 1021da177e4SLinus Torvalds */ 1031da177e4SLinus Torvalds static unsigned int __init 1041da177e4SLinus Torvalds find_bootmap_pfn(int node, struct meminfo *mi, unsigned int bootmap_pages) 1051da177e4SLinus Torvalds { 1061da177e4SLinus Torvalds unsigned int start_pfn, bank, bootmap_pfn; 1071da177e4SLinus Torvalds 10890072059SRussell King start_pfn = PAGE_ALIGN(__pa(&_end)) >> PAGE_SHIFT; 1091da177e4SLinus Torvalds bootmap_pfn = 0; 1101da177e4SLinus Torvalds 11190072059SRussell King for_each_nodebank(bank, mi, node) { 1121da177e4SLinus Torvalds unsigned int start, end; 1131da177e4SLinus Torvalds 11492a8cbedSRussell King start = mi->bank[bank].start >> PAGE_SHIFT; 11592a8cbedSRussell King end = (mi->bank[bank].size + 11692a8cbedSRussell King mi->bank[bank].start) >> PAGE_SHIFT; 1171da177e4SLinus Torvalds 1181da177e4SLinus Torvalds if (end < start_pfn) 1191da177e4SLinus Torvalds continue; 1201da177e4SLinus Torvalds 1211da177e4SLinus Torvalds if (start < start_pfn) 1221da177e4SLinus Torvalds start = start_pfn; 1231da177e4SLinus Torvalds 1241da177e4SLinus Torvalds if (end <= start) 1251da177e4SLinus Torvalds continue; 1261da177e4SLinus Torvalds 1271da177e4SLinus Torvalds if (end - start >= bootmap_pages) { 1281da177e4SLinus Torvalds bootmap_pfn = start; 1291da177e4SLinus Torvalds break; 1301da177e4SLinus Torvalds } 1311da177e4SLinus Torvalds } 1321da177e4SLinus Torvalds 1331da177e4SLinus Torvalds if (bootmap_pfn == 0) 1341da177e4SLinus Torvalds BUG(); 1351da177e4SLinus Torvalds 1361da177e4SLinus Torvalds return bootmap_pfn; 1371da177e4SLinus Torvalds } 1381da177e4SLinus Torvalds 1391da177e4SLinus Torvalds static int __init check_initrd(struct meminfo *mi) 1401da177e4SLinus Torvalds { 1411da177e4SLinus Torvalds int initrd_node = -2; 1421da177e4SLinus Torvalds #ifdef CONFIG_BLK_DEV_INITRD 1431da177e4SLinus Torvalds unsigned long end = phys_initrd_start + phys_initrd_size; 1441da177e4SLinus Torvalds 1451da177e4SLinus Torvalds /* 1461da177e4SLinus Torvalds * Make sure that the initrd is within a valid area of 1471da177e4SLinus Torvalds * memory. 1481da177e4SLinus Torvalds */ 1491da177e4SLinus Torvalds if (phys_initrd_size) { 1501da177e4SLinus Torvalds unsigned int i; 1511da177e4SLinus Torvalds 1521da177e4SLinus Torvalds initrd_node = -1; 1531da177e4SLinus Torvalds 1541da177e4SLinus Torvalds for (i = 0; i < mi->nr_banks; i++) { 1551da177e4SLinus Torvalds unsigned long bank_end; 1561da177e4SLinus Torvalds 1571da177e4SLinus Torvalds bank_end = mi->bank[i].start + mi->bank[i].size; 1581da177e4SLinus Torvalds 1591da177e4SLinus Torvalds if (mi->bank[i].start <= phys_initrd_start && 1601da177e4SLinus Torvalds end <= bank_end) 1611da177e4SLinus Torvalds initrd_node = mi->bank[i].node; 1621da177e4SLinus Torvalds } 1631da177e4SLinus Torvalds } 1641da177e4SLinus Torvalds 1651da177e4SLinus Torvalds if (initrd_node == -1) { 1661da177e4SLinus Torvalds printk(KERN_ERR "initrd (0x%08lx - 0x%08lx) extends beyond " 1671da177e4SLinus Torvalds "physical memory - disabling initrd\n", 1681da177e4SLinus Torvalds phys_initrd_start, end); 1691da177e4SLinus Torvalds phys_initrd_start = phys_initrd_size = 0; 1701da177e4SLinus Torvalds } 1711da177e4SLinus Torvalds #endif 1721da177e4SLinus Torvalds 1731da177e4SLinus Torvalds return initrd_node; 1741da177e4SLinus Torvalds } 1751da177e4SLinus Torvalds 1761da177e4SLinus Torvalds /* 1771da177e4SLinus Torvalds * Reserve the various regions of node 0 1781da177e4SLinus Torvalds */ 17990072059SRussell King static __init void reserve_node_zero(pg_data_t *pgdat) 1801da177e4SLinus Torvalds { 1811da177e4SLinus Torvalds unsigned long res_size = 0; 1821da177e4SLinus Torvalds 1831da177e4SLinus Torvalds /* 1841da177e4SLinus Torvalds * Register the kernel text and data with bootmem. 1851da177e4SLinus Torvalds * Note that this can only be in node 0. 1861da177e4SLinus Torvalds */ 1871da177e4SLinus Torvalds #ifdef CONFIG_XIP_KERNEL 1881da177e4SLinus Torvalds reserve_bootmem_node(pgdat, __pa(&__data_start), &_end - &__data_start); 1891da177e4SLinus Torvalds #else 1901da177e4SLinus Torvalds reserve_bootmem_node(pgdat, __pa(&_stext), &_end - &_stext); 1911da177e4SLinus Torvalds #endif 1921da177e4SLinus Torvalds 1931da177e4SLinus Torvalds /* 1941da177e4SLinus Torvalds * Reserve the page tables. These are already in use, 1951da177e4SLinus Torvalds * and can only be in node 0. 1961da177e4SLinus Torvalds */ 1971da177e4SLinus Torvalds reserve_bootmem_node(pgdat, __pa(swapper_pg_dir), 1981da177e4SLinus Torvalds PTRS_PER_PGD * sizeof(pgd_t)); 1991da177e4SLinus Torvalds 2001da177e4SLinus Torvalds /* 2011da177e4SLinus Torvalds * Hmm... This should go elsewhere, but we really really need to 2021da177e4SLinus Torvalds * stop things allocating the low memory; ideally we need a better 2031da177e4SLinus Torvalds * implementation of GFP_DMA which does not assume that DMA-able 2041da177e4SLinus Torvalds * memory starts at zero. 2051da177e4SLinus Torvalds */ 2061da177e4SLinus Torvalds if (machine_is_integrator() || machine_is_cintegrator()) 2071da177e4SLinus Torvalds res_size = __pa(swapper_pg_dir) - PHYS_OFFSET; 2081da177e4SLinus Torvalds 2091da177e4SLinus Torvalds /* 2101da177e4SLinus Torvalds * These should likewise go elsewhere. They pre-reserve the 2111da177e4SLinus Torvalds * screen memory region at the start of main system memory. 2121da177e4SLinus Torvalds */ 2131da177e4SLinus Torvalds if (machine_is_edb7211()) 2141da177e4SLinus Torvalds res_size = 0x00020000; 2151da177e4SLinus Torvalds if (machine_is_p720t()) 2161da177e4SLinus Torvalds res_size = 0x00014000; 2171da177e4SLinus Torvalds 2181da177e4SLinus Torvalds #ifdef CONFIG_SA1111 2191da177e4SLinus Torvalds /* 2201da177e4SLinus Torvalds * Because of the SA1111 DMA bug, we want to preserve our 2211da177e4SLinus Torvalds * precious DMA-able memory... 2221da177e4SLinus Torvalds */ 2231da177e4SLinus Torvalds res_size = __pa(swapper_pg_dir) - PHYS_OFFSET; 2241da177e4SLinus Torvalds #endif 2251da177e4SLinus Torvalds if (res_size) 2261da177e4SLinus Torvalds reserve_bootmem_node(pgdat, PHYS_OFFSET, res_size); 2271da177e4SLinus Torvalds } 2281da177e4SLinus Torvalds 22990072059SRussell King static unsigned long __init 23090072059SRussell King bootmem_init_node(int node, int initrd_node, struct meminfo *mi) 2311da177e4SLinus Torvalds { 23290072059SRussell King unsigned long zone_size[MAX_NR_ZONES], zhole_size[MAX_NR_ZONES]; 23390072059SRussell King unsigned long start_pfn, end_pfn, boot_pfn; 23490072059SRussell King unsigned int boot_pages; 23590072059SRussell King pg_data_t *pgdat; 23690072059SRussell King int i; 2371da177e4SLinus Torvalds 23890072059SRussell King start_pfn = -1UL; 23990072059SRussell King end_pfn = 0; 24090072059SRussell King 24190072059SRussell King /* 24290072059SRussell King * Calculate the pfn range, and map the memory banks for this node. 24390072059SRussell King */ 24490072059SRussell King for_each_nodebank(i, mi, node) { 24590072059SRussell King unsigned long start, end; 24690072059SRussell King struct map_desc map; 24790072059SRussell King 24890072059SRussell King start = mi->bank[i].start >> PAGE_SHIFT; 24990072059SRussell King end = (mi->bank[i].start + mi->bank[i].size) >> PAGE_SHIFT; 25090072059SRussell King 25190072059SRussell King if (start_pfn > start) 25290072059SRussell King start_pfn = start; 25390072059SRussell King if (end_pfn < end) 25490072059SRussell King end_pfn = end; 25590072059SRussell King 2569769c246SDeepak Saxena map.pfn = __phys_to_pfn(mi->bank[i].start); 2579769c246SDeepak Saxena map.virtual = __phys_to_virt(mi->bank[i].start); 25890072059SRussell King map.length = mi->bank[i].size; 25990072059SRussell King map.type = MT_MEMORY; 26090072059SRussell King 26190072059SRussell King create_mapping(&map); 2621da177e4SLinus Torvalds } 2631da177e4SLinus Torvalds 2641da177e4SLinus Torvalds /* 26590072059SRussell King * If there is no memory in this node, ignore it. 2661da177e4SLinus Torvalds */ 26790072059SRussell King if (end_pfn == 0) 26890072059SRussell King return end_pfn; 2691da177e4SLinus Torvalds 2701da177e4SLinus Torvalds /* 27190072059SRussell King * Allocate the bootmem bitmap page. 2721da177e4SLinus Torvalds */ 27390072059SRussell King boot_pages = bootmem_bootmap_pages(end_pfn - start_pfn); 27490072059SRussell King boot_pfn = find_bootmap_pfn(node, mi, boot_pages); 2751da177e4SLinus Torvalds 2761da177e4SLinus Torvalds /* 27790072059SRussell King * Initialise the bootmem allocator for this node, handing the 27890072059SRussell King * memory banks over to bootmem. 2791da177e4SLinus Torvalds */ 28090072059SRussell King node_set_online(node); 28190072059SRussell King pgdat = NODE_DATA(node); 28290072059SRussell King init_bootmem_node(pgdat, boot_pfn, start_pfn, end_pfn); 28390072059SRussell King 28490072059SRussell King for_each_nodebank(i, mi, node) 28590072059SRussell King free_bootmem_node(pgdat, mi->bank[i].start, mi->bank[i].size); 2861da177e4SLinus Torvalds 2871da177e4SLinus Torvalds /* 28890072059SRussell King * Reserve the bootmem bitmap for this node. 2891da177e4SLinus Torvalds */ 29090072059SRussell King reserve_bootmem_node(pgdat, boot_pfn << PAGE_SHIFT, 29190072059SRussell King boot_pages << PAGE_SHIFT); 2921da177e4SLinus Torvalds 2931da177e4SLinus Torvalds #ifdef CONFIG_BLK_DEV_INITRD 29490072059SRussell King /* 29590072059SRussell King * If the initrd is in this node, reserve its memory. 29690072059SRussell King */ 29790072059SRussell King if (node == initrd_node) { 29890072059SRussell King reserve_bootmem_node(pgdat, phys_initrd_start, 2991da177e4SLinus Torvalds phys_initrd_size); 3001da177e4SLinus Torvalds initrd_start = __phys_to_virt(phys_initrd_start); 3011da177e4SLinus Torvalds initrd_end = initrd_start + phys_initrd_size; 3021da177e4SLinus Torvalds } 3031da177e4SLinus Torvalds #endif 3041da177e4SLinus Torvalds 30590072059SRussell King /* 30690072059SRussell King * Finally, reserve any node zero regions. 30790072059SRussell King */ 30890072059SRussell King if (node == 0) 30990072059SRussell King reserve_node_zero(pgdat); 31090072059SRussell King 31190072059SRussell King /* 31290072059SRussell King * initialise the zones within this node. 31390072059SRussell King */ 31490072059SRussell King memset(zone_size, 0, sizeof(zone_size)); 31590072059SRussell King memset(zhole_size, 0, sizeof(zhole_size)); 31690072059SRussell King 31790072059SRussell King /* 31890072059SRussell King * The size of this node has already been determined. If we need 31990072059SRussell King * to do anything fancy with the allocation of this memory to the 32090072059SRussell King * zones, now is the time to do it. 32190072059SRussell King */ 32290072059SRussell King zone_size[0] = end_pfn - start_pfn; 32390072059SRussell King 32490072059SRussell King /* 32590072059SRussell King * For each bank in this node, calculate the size of the holes. 32690072059SRussell King * holes = node_size - sum(bank_sizes_in_node) 32790072059SRussell King */ 32890072059SRussell King zhole_size[0] = zone_size[0]; 32990072059SRussell King for_each_nodebank(i, mi, node) 33090072059SRussell King zhole_size[0] -= mi->bank[i].size >> PAGE_SHIFT; 33190072059SRussell King 33290072059SRussell King /* 33390072059SRussell King * Adjust the sizes according to any special requirements for 33490072059SRussell King * this machine type. 33590072059SRussell King */ 33690072059SRussell King arch_adjust_zones(node, zone_size, zhole_size); 33790072059SRussell King 33890072059SRussell King free_area_init_node(node, pgdat, zone_size, start_pfn, zhole_size); 33990072059SRussell King 34090072059SRussell King return end_pfn; 34190072059SRussell King } 34290072059SRussell King 34390072059SRussell King static void __init bootmem_init(struct meminfo *mi) 34490072059SRussell King { 34590072059SRussell King unsigned long addr, memend_pfn = 0; 34690072059SRussell King int node, initrd_node, i; 34790072059SRussell King 34890072059SRussell King /* 34990072059SRussell King * Invalidate the node number for empty or invalid memory banks 35090072059SRussell King */ 35190072059SRussell King for (i = 0; i < mi->nr_banks; i++) 35290072059SRussell King if (mi->bank[i].size == 0 || mi->bank[i].node >= MAX_NUMNODES) 35390072059SRussell King mi->bank[i].node = -1; 35490072059SRussell King 35590072059SRussell King memcpy(&meminfo, mi, sizeof(meminfo)); 35690072059SRussell King 35790072059SRussell King /* 35890072059SRussell King * Clear out all the mappings below the kernel image. 35990072059SRussell King */ 3601a47ebc0SNicolas Pitre for (addr = 0; addr < MODULE_START; addr += PGDIR_SIZE) 3611a47ebc0SNicolas Pitre pmd_clear(pmd_off_k(addr)); 3621a47ebc0SNicolas Pitre #ifdef CONFIG_XIP_KERNEL 3631a47ebc0SNicolas Pitre /* The XIP kernel is mapped in the module area -- skip over it */ 3641a47ebc0SNicolas Pitre addr = ((unsigned long)&_etext + PGDIR_SIZE - 1) & PGDIR_MASK; 3651a47ebc0SNicolas Pitre #endif 3661a47ebc0SNicolas Pitre for ( ; addr < PAGE_OFFSET; addr += PGDIR_SIZE) 36790072059SRussell King pmd_clear(pmd_off_k(addr)); 36890072059SRussell King 36990072059SRussell King /* 37090072059SRussell King * Clear out all the kernel space mappings, except for the first 37190072059SRussell King * memory bank, up to the end of the vmalloc region. 37290072059SRussell King */ 37390072059SRussell King for (addr = __phys_to_virt(mi->bank[0].start + mi->bank[0].size); 37490072059SRussell King addr < VMALLOC_END; addr += PGDIR_SIZE) 37590072059SRussell King pmd_clear(pmd_off_k(addr)); 37690072059SRussell King 37790072059SRussell King /* 37890072059SRussell King * Locate which node contains the ramdisk image, if any. 37990072059SRussell King */ 38090072059SRussell King initrd_node = check_initrd(mi); 38190072059SRussell King 38290072059SRussell King /* 38390072059SRussell King * Run through each node initialising the bootmem allocator. 38490072059SRussell King */ 38590072059SRussell King for_each_node(node) { 38690072059SRussell King unsigned long end_pfn; 38790072059SRussell King 38890072059SRussell King end_pfn = bootmem_init_node(node, initrd_node, mi); 38990072059SRussell King 39090072059SRussell King /* 39190072059SRussell King * Remember the highest memory PFN. 39290072059SRussell King */ 39390072059SRussell King if (end_pfn > memend_pfn) 39490072059SRussell King memend_pfn = end_pfn; 39590072059SRussell King } 39690072059SRussell King 39790072059SRussell King high_memory = __va(memend_pfn << PAGE_SHIFT); 39890072059SRussell King 39990072059SRussell King /* 40090072059SRussell King * This doesn't seem to be used by the Linux memory manager any 40190072059SRussell King * more, but is used by ll_rw_block. If we can get rid of it, we 40290072059SRussell King * also get rid of some of the stuff above as well. 40390072059SRussell King * 40490072059SRussell King * Note: max_low_pfn and max_pfn reflect the number of _pages_ in 40590072059SRussell King * the system, not the maximum PFN. 40690072059SRussell King */ 40790072059SRussell King max_pfn = max_low_pfn = memend_pfn - PHYS_PFN_OFFSET; 40890072059SRussell King } 40990072059SRussell King 41090072059SRussell King /* 41190072059SRussell King * Set up device the mappings. Since we clear out the page tables for all 41290072059SRussell King * mappings above VMALLOC_END, we will remove any debug device mappings. 41390072059SRussell King * This means you have to be careful how you debug this function, or any 41402b30839SRussell King * called function. This means you can't use any function or debugging 41502b30839SRussell King * method which may touch any device, otherwise the kernel _will_ crash. 41690072059SRussell King */ 41790072059SRussell King static void __init devicemaps_init(struct machine_desc *mdesc) 41890072059SRussell King { 41990072059SRussell King struct map_desc map; 42090072059SRussell King unsigned long addr; 42190072059SRussell King void *vectors; 42290072059SRussell King 42302b30839SRussell King /* 42402b30839SRussell King * Allocate the vector page early. 42502b30839SRussell King */ 42602b30839SRussell King vectors = alloc_bootmem_low_pages(PAGE_SIZE); 42702b30839SRussell King BUG_ON(!vectors); 42802b30839SRussell King 42990072059SRussell King for (addr = VMALLOC_END; addr; addr += PGDIR_SIZE) 43090072059SRussell King pmd_clear(pmd_off_k(addr)); 43190072059SRussell King 43290072059SRussell King /* 4331a47ebc0SNicolas Pitre * Map the kernel if it is XIP. 4341a47ebc0SNicolas Pitre * It is always first in the modulearea. 4351a47ebc0SNicolas Pitre */ 4361a47ebc0SNicolas Pitre #ifdef CONFIG_XIP_KERNEL 4371a47ebc0SNicolas Pitre map.pfn = __phys_to_pfn(CONFIG_XIP_PHYS_ADDR & PGDIR_MASK); 4381a47ebc0SNicolas Pitre map.virtual = MODULE_START; 4391a47ebc0SNicolas Pitre map.length = ((unsigned long)&_etext - map.virtual + ~PGDIR_MASK) & PGDIR_MASK; 4401a47ebc0SNicolas Pitre map.type = MT_ROM; 4411a47ebc0SNicolas Pitre create_mapping(&map); 4421a47ebc0SNicolas Pitre #endif 4431a47ebc0SNicolas Pitre 4441a47ebc0SNicolas Pitre /* 44590072059SRussell King * Map the cache flushing regions. 44690072059SRussell King */ 44790072059SRussell King #ifdef FLUSH_BASE 4489769c246SDeepak Saxena map.pfn = __phys_to_pfn(FLUSH_BASE_PHYS); 44990072059SRussell King map.virtual = FLUSH_BASE; 45074d02fb9SRussell King map.length = SZ_1M; 45190072059SRussell King map.type = MT_CACHECLEAN; 45290072059SRussell King create_mapping(&map); 45390072059SRussell King #endif 45490072059SRussell King #ifdef FLUSH_BASE_MINICACHE 45574d02fb9SRussell King map.pfn = __phys_to_pfn(FLUSH_BASE_PHYS + SZ_1M); 45690072059SRussell King map.virtual = FLUSH_BASE_MINICACHE; 45774d02fb9SRussell King map.length = SZ_1M; 45890072059SRussell King map.type = MT_MINICLEAN; 45990072059SRussell King create_mapping(&map); 46090072059SRussell King #endif 46190072059SRussell King 46290072059SRussell King /* 46390072059SRussell King * Create a mapping for the machine vectors at the high-vectors 46490072059SRussell King * location (0xffff0000). If we aren't using high-vectors, also 46590072059SRussell King * create a mapping at the low-vectors virtual address. 46690072059SRussell King */ 4679769c246SDeepak Saxena map.pfn = __phys_to_pfn(virt_to_phys(vectors)); 46890072059SRussell King map.virtual = 0xffff0000; 46990072059SRussell King map.length = PAGE_SIZE; 47090072059SRussell King map.type = MT_HIGH_VECTORS; 47190072059SRussell King create_mapping(&map); 47290072059SRussell King 47390072059SRussell King if (!vectors_high()) { 47490072059SRussell King map.virtual = 0; 47590072059SRussell King map.type = MT_LOW_VECTORS; 47690072059SRussell King create_mapping(&map); 47790072059SRussell King } 47890072059SRussell King 47990072059SRussell King /* 48090072059SRussell King * Ask the machine support to map in the statically mapped devices. 48190072059SRussell King */ 48290072059SRussell King if (mdesc->map_io) 48390072059SRussell King mdesc->map_io(); 4846bf7bd69SRussell King 4856bf7bd69SRussell King /* 48602b30839SRussell King * Finally flush the caches and tlb to ensure that we're in a 48702b30839SRussell King * consistent state wrt the writebuffer. This also ensures that 48802b30839SRussell King * any write-allocated cache lines in the vector page are written 48902b30839SRussell King * back. After this point, we can start to touch devices again. 4906bf7bd69SRussell King */ 4916bf7bd69SRussell King local_flush_tlb_all(); 49202b30839SRussell King flush_cache_all(); 4931da177e4SLinus Torvalds } 4941da177e4SLinus Torvalds 4951da177e4SLinus Torvalds /* 4961da177e4SLinus Torvalds * paging_init() sets up the page tables, initialises the zone memory 4971da177e4SLinus Torvalds * maps, and sets up the zero page, bad page and bad page tables. 4981da177e4SLinus Torvalds */ 4991da177e4SLinus Torvalds void __init paging_init(struct meminfo *mi, struct machine_desc *mdesc) 5001da177e4SLinus Torvalds { 5011da177e4SLinus Torvalds void *zero_page; 5021da177e4SLinus Torvalds 50390072059SRussell King build_mem_type_table(); 5041da177e4SLinus Torvalds bootmem_init(mi); 50590072059SRussell King devicemaps_init(mdesc); 5061da177e4SLinus Torvalds 50790072059SRussell King top_pmd = pmd_off_k(0xffff0000); 5081da177e4SLinus Torvalds 5091da177e4SLinus Torvalds /* 5101da177e4SLinus Torvalds * allocate the zero page. Note that we count on this going ok. 5111da177e4SLinus Torvalds */ 5121da177e4SLinus Torvalds zero_page = alloc_bootmem_low_pages(PAGE_SIZE); 5131da177e4SLinus Torvalds memzero(zero_page, PAGE_SIZE); 5141da177e4SLinus Torvalds empty_zero_page = virt_to_page(zero_page); 5151da177e4SLinus Torvalds flush_dcache_page(empty_zero_page); 5161da177e4SLinus Torvalds } 5171da177e4SLinus Torvalds 5181da177e4SLinus Torvalds static inline void free_area(unsigned long addr, unsigned long end, char *s) 5191da177e4SLinus Torvalds { 5201da177e4SLinus Torvalds unsigned int size = (end - addr) >> 10; 5211da177e4SLinus Torvalds 5221da177e4SLinus Torvalds for (; addr < end; addr += PAGE_SIZE) { 5231da177e4SLinus Torvalds struct page *page = virt_to_page(addr); 5241da177e4SLinus Torvalds ClearPageReserved(page); 5257835e98bSNick Piggin init_page_count(page); 5261da177e4SLinus Torvalds free_page(addr); 5271da177e4SLinus Torvalds totalram_pages++; 5281da177e4SLinus Torvalds } 5291da177e4SLinus Torvalds 5301da177e4SLinus Torvalds if (size && s) 5311da177e4SLinus Torvalds printk(KERN_INFO "Freeing %s memory: %dK\n", s, size); 5321da177e4SLinus Torvalds } 5331da177e4SLinus Torvalds 534a013053dSRussell King static inline void 535a013053dSRussell King free_memmap(int node, unsigned long start_pfn, unsigned long end_pfn) 536a013053dSRussell King { 537a013053dSRussell King struct page *start_pg, *end_pg; 538a013053dSRussell King unsigned long pg, pgend; 539a013053dSRussell King 540a013053dSRussell King /* 541a013053dSRussell King * Convert start_pfn/end_pfn to a struct page pointer. 542a013053dSRussell King */ 543a013053dSRussell King start_pg = pfn_to_page(start_pfn); 544a013053dSRussell King end_pg = pfn_to_page(end_pfn); 545a013053dSRussell King 546a013053dSRussell King /* 547a013053dSRussell King * Convert to physical addresses, and 548a013053dSRussell King * round start upwards and end downwards. 549a013053dSRussell King */ 550a013053dSRussell King pg = PAGE_ALIGN(__pa(start_pg)); 551a013053dSRussell King pgend = __pa(end_pg) & PAGE_MASK; 552a013053dSRussell King 553a013053dSRussell King /* 554a013053dSRussell King * If there are free pages between these, 555a013053dSRussell King * free the section of the memmap array. 556a013053dSRussell King */ 557a013053dSRussell King if (pg < pgend) 558a013053dSRussell King free_bootmem_node(NODE_DATA(node), pg, pgend - pg); 559a013053dSRussell King } 560a013053dSRussell King 561a013053dSRussell King /* 562a013053dSRussell King * The mem_map array can get very big. Free the unused area of the memory map. 563a013053dSRussell King */ 564a013053dSRussell King static void __init free_unused_memmap_node(int node, struct meminfo *mi) 565a013053dSRussell King { 566a013053dSRussell King unsigned long bank_start, prev_bank_end = 0; 567a013053dSRussell King unsigned int i; 568a013053dSRussell King 569a013053dSRussell King /* 570a013053dSRussell King * [FIXME] This relies on each bank being in address order. This 571a013053dSRussell King * may not be the case, especially if the user has provided the 572a013053dSRussell King * information on the command line. 573a013053dSRussell King */ 57490072059SRussell King for_each_nodebank(i, mi, node) { 575a013053dSRussell King bank_start = mi->bank[i].start >> PAGE_SHIFT; 576a013053dSRussell King if (bank_start < prev_bank_end) { 577a013053dSRussell King printk(KERN_ERR "MEM: unordered memory banks. " 578a013053dSRussell King "Not freeing memmap.\n"); 579a013053dSRussell King break; 580a013053dSRussell King } 581a013053dSRussell King 582a013053dSRussell King /* 583a013053dSRussell King * If we had a previous bank, and there is a space 584a013053dSRussell King * between the current bank and the previous, free it. 585a013053dSRussell King */ 586a013053dSRussell King if (prev_bank_end && prev_bank_end != bank_start) 587a013053dSRussell King free_memmap(node, prev_bank_end, bank_start); 588a013053dSRussell King 589a013053dSRussell King prev_bank_end = (mi->bank[i].start + 590a013053dSRussell King mi->bank[i].size) >> PAGE_SHIFT; 591a013053dSRussell King } 592a013053dSRussell King } 593a013053dSRussell King 5941da177e4SLinus Torvalds /* 5951da177e4SLinus Torvalds * mem_init() marks the free areas in the mem_map and tells us how much 5961da177e4SLinus Torvalds * memory is free. This is done after various parts of the system have 5971da177e4SLinus Torvalds * claimed their memory after the kernel image. 5981da177e4SLinus Torvalds */ 5991da177e4SLinus Torvalds void __init mem_init(void) 6001da177e4SLinus Torvalds { 6011da177e4SLinus Torvalds unsigned int codepages, datapages, initpages; 6021da177e4SLinus Torvalds int i, node; 6031da177e4SLinus Torvalds 6041da177e4SLinus Torvalds codepages = &_etext - &_text; 6051da177e4SLinus Torvalds datapages = &_end - &__data_start; 6061da177e4SLinus Torvalds initpages = &__init_end - &__init_begin; 6071da177e4SLinus Torvalds 6081da177e4SLinus Torvalds #ifndef CONFIG_DISCONTIGMEM 6091da177e4SLinus Torvalds max_mapnr = virt_to_page(high_memory) - mem_map; 6101da177e4SLinus Torvalds #endif 6111da177e4SLinus Torvalds 6121da177e4SLinus Torvalds /* this will put all unused low memory onto the freelists */ 6131da177e4SLinus Torvalds for_each_online_node(node) { 6141da177e4SLinus Torvalds pg_data_t *pgdat = NODE_DATA(node); 6151da177e4SLinus Torvalds 616a013053dSRussell King free_unused_memmap_node(node, &meminfo); 617a013053dSRussell King 6181da177e4SLinus Torvalds if (pgdat->node_spanned_pages != 0) 6191da177e4SLinus Torvalds totalram_pages += free_all_bootmem_node(pgdat); 6201da177e4SLinus Torvalds } 6211da177e4SLinus Torvalds 6221da177e4SLinus Torvalds #ifdef CONFIG_SA1111 6231da177e4SLinus Torvalds /* now that our DMA memory is actually so designated, we can free it */ 6241da177e4SLinus Torvalds free_area(PAGE_OFFSET, (unsigned long)swapper_pg_dir, NULL); 6251da177e4SLinus Torvalds #endif 6261da177e4SLinus Torvalds 6271da177e4SLinus Torvalds /* 6281da177e4SLinus Torvalds * Since our memory may not be contiguous, calculate the 6291da177e4SLinus Torvalds * real number of pages we have in this system 6301da177e4SLinus Torvalds */ 6311da177e4SLinus Torvalds printk(KERN_INFO "Memory:"); 6321da177e4SLinus Torvalds 6331da177e4SLinus Torvalds num_physpages = 0; 6341da177e4SLinus Torvalds for (i = 0; i < meminfo.nr_banks; i++) { 6351da177e4SLinus Torvalds num_physpages += meminfo.bank[i].size >> PAGE_SHIFT; 6361da177e4SLinus Torvalds printk(" %ldMB", meminfo.bank[i].size >> 20); 6371da177e4SLinus Torvalds } 6381da177e4SLinus Torvalds 6391da177e4SLinus Torvalds printk(" = %luMB total\n", num_physpages >> (20 - PAGE_SHIFT)); 6401da177e4SLinus Torvalds printk(KERN_NOTICE "Memory: %luKB available (%dK code, " 6411da177e4SLinus Torvalds "%dK data, %dK init)\n", 6421da177e4SLinus Torvalds (unsigned long) nr_free_pages() << (PAGE_SHIFT-10), 6431da177e4SLinus Torvalds codepages >> 10, datapages >> 10, initpages >> 10); 6441da177e4SLinus Torvalds 6451da177e4SLinus Torvalds if (PAGE_SIZE >= 16384 && num_physpages <= 128) { 6461da177e4SLinus Torvalds extern int sysctl_overcommit_memory; 6471da177e4SLinus Torvalds /* 6481da177e4SLinus Torvalds * On a machine this small we won't get 6491da177e4SLinus Torvalds * anywhere without overcommit, so turn 6501da177e4SLinus Torvalds * it on by default. 6511da177e4SLinus Torvalds */ 6521da177e4SLinus Torvalds sysctl_overcommit_memory = OVERCOMMIT_ALWAYS; 6531da177e4SLinus Torvalds } 6541da177e4SLinus Torvalds } 6551da177e4SLinus Torvalds 6561da177e4SLinus Torvalds void free_initmem(void) 6571da177e4SLinus Torvalds { 6581da177e4SLinus Torvalds if (!machine_is_integrator() && !machine_is_cintegrator()) { 6591da177e4SLinus Torvalds free_area((unsigned long)(&__init_begin), 6601da177e4SLinus Torvalds (unsigned long)(&__init_end), 6611da177e4SLinus Torvalds "init"); 6621da177e4SLinus Torvalds } 6631da177e4SLinus Torvalds } 6641da177e4SLinus Torvalds 6651da177e4SLinus Torvalds #ifdef CONFIG_BLK_DEV_INITRD 6661da177e4SLinus Torvalds 6671da177e4SLinus Torvalds static int keep_initrd; 6681da177e4SLinus Torvalds 6691da177e4SLinus Torvalds void free_initrd_mem(unsigned long start, unsigned long end) 6701da177e4SLinus Torvalds { 6711da177e4SLinus Torvalds if (!keep_initrd) 6721da177e4SLinus Torvalds free_area(start, end, "initrd"); 6731da177e4SLinus Torvalds } 6741da177e4SLinus Torvalds 6751da177e4SLinus Torvalds static int __init keepinitrd_setup(char *__unused) 6761da177e4SLinus Torvalds { 6771da177e4SLinus Torvalds keep_initrd = 1; 6781da177e4SLinus Torvalds return 1; 6791da177e4SLinus Torvalds } 6801da177e4SLinus Torvalds 6811da177e4SLinus Torvalds __setup("keepinitrd", keepinitrd_setup); 6821da177e4SLinus Torvalds #endif 683