114cf11afSPaul Mackerras /* 214cf11afSPaul Mackerras * PowerPC version 314cf11afSPaul Mackerras * Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org) 414cf11afSPaul Mackerras * 514cf11afSPaul Mackerras * Modifications by Paul Mackerras (PowerMac) (paulus@cs.anu.edu.au) 614cf11afSPaul Mackerras * and Cort Dougan (PReP) (cort@cs.nmt.edu) 714cf11afSPaul Mackerras * Copyright (C) 1996 Paul Mackerras 814cf11afSPaul Mackerras * PPC44x/36-bit changes by Matt Porter (mporter@mvista.com) 914cf11afSPaul Mackerras * 1014cf11afSPaul Mackerras * Derived from "arch/i386/mm/init.c" 1114cf11afSPaul Mackerras * Copyright (C) 1991, 1992, 1993, 1994 Linus Torvalds 1214cf11afSPaul Mackerras * 1314cf11afSPaul Mackerras * This program is free software; you can redistribute it and/or 1414cf11afSPaul Mackerras * modify it under the terms of the GNU General Public License 1514cf11afSPaul Mackerras * as published by the Free Software Foundation; either version 1614cf11afSPaul Mackerras * 2 of the License, or (at your option) any later version. 1714cf11afSPaul Mackerras * 1814cf11afSPaul Mackerras */ 1914cf11afSPaul Mackerras 204b16f8e2SPaul Gortmaker #include <linux/export.h> 2114cf11afSPaul Mackerras #include <linux/sched.h> 2214cf11afSPaul Mackerras #include <linux/kernel.h> 2314cf11afSPaul Mackerras #include <linux/errno.h> 2414cf11afSPaul Mackerras #include <linux/string.h> 255a0e3ad6STejun Heo #include <linux/gfp.h> 2614cf11afSPaul Mackerras #include <linux/types.h> 2714cf11afSPaul Mackerras #include <linux/mm.h> 2814cf11afSPaul Mackerras #include <linux/stddef.h> 2914cf11afSPaul Mackerras #include <linux/init.h> 3014cf11afSPaul Mackerras #include <linux/bootmem.h> 3114cf11afSPaul Mackerras #include <linux/highmem.h> 3214cf11afSPaul Mackerras #include <linux/initrd.h> 3314cf11afSPaul Mackerras #include <linux/pagemap.h> 344e8ad3e8SJohannes Berg #include <linux/suspend.h> 3595f72d1eSYinghai Lu #include <linux/memblock.h> 360895ecdaSDavid Gibson #include <linux/hugetlb.h> 37c40dd2f7SAnton Blanchard #include <linux/slab.h> 3816d0f5c4SAnton Blanchard #include <linux/vmalloc.h> 3914cf11afSPaul Mackerras 4014cf11afSPaul Mackerras #include <asm/pgalloc.h> 4114cf11afSPaul Mackerras #include <asm/prom.h> 4214cf11afSPaul Mackerras #include <asm/io.h> 4314cf11afSPaul Mackerras #include <asm/mmu_context.h> 4414cf11afSPaul Mackerras #include <asm/pgtable.h> 4514cf11afSPaul Mackerras #include <asm/mmu.h> 4614cf11afSPaul Mackerras #include <asm/smp.h> 4714cf11afSPaul Mackerras #include <asm/machdep.h> 4814cf11afSPaul Mackerras #include <asm/btext.h> 4914cf11afSPaul Mackerras #include <asm/tlb.h> 507c8c6b97SPaul Mackerras #include <asm/sections.h> 51db7f37deSTony Breeds #include <asm/sparsemem.h> 52ab1f9dacSPaul Mackerras #include <asm/vdso.h> 532c419bdeSKumar Gala #include <asm/fixmap.h> 54a9327296SFUJITA Tomonori #include <asm/swiotlb.h> 558a3e3d31Ssukadev@linux.vnet.ibm.com #include <asm/rtas.h> 5614cf11afSPaul Mackerras 5714cf11afSPaul Mackerras #include "mmu_decl.h" 5814cf11afSPaul Mackerras 5914cf11afSPaul Mackerras #ifndef CPU_FTR_COHERENT_ICACHE 6014cf11afSPaul Mackerras #define CPU_FTR_COHERENT_ICACHE 0 /* XXX for now */ 6114cf11afSPaul Mackerras #define CPU_FTR_NOEXECUTE 0 6214cf11afSPaul Mackerras #endif 6314cf11afSPaul Mackerras 647c8c6b97SPaul Mackerras int mem_init_done; 65a84fcd46SSuzuki Poulose unsigned long long memory_limit; 667c8c6b97SPaul Mackerras 672c419bdeSKumar Gala #ifdef CONFIG_HIGHMEM 682c419bdeSKumar Gala pte_t *kmap_pte; 692c419bdeSKumar Gala EXPORT_SYMBOL(kmap_pte); 708040bda3SValentina Manea pgprot_t kmap_prot; 718040bda3SValentina Manea EXPORT_SYMBOL(kmap_prot); 722c419bdeSKumar Gala 732c419bdeSKumar Gala static inline pte_t *virt_to_kpte(unsigned long vaddr) 742c419bdeSKumar Gala { 752c419bdeSKumar Gala return pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), 762c419bdeSKumar Gala vaddr), vaddr), vaddr); 772c419bdeSKumar Gala } 782c419bdeSKumar Gala #endif 792c419bdeSKumar Gala 8014cf11afSPaul Mackerras int page_is_ram(unsigned long pfn) 8114cf11afSPaul Mackerras { 8214cf11afSPaul Mackerras #ifndef CONFIG_PPC64 /* XXX for now */ 83a880e762SRoland Dreier return pfn < max_pfn; 8414cf11afSPaul Mackerras #else 85a880e762SRoland Dreier unsigned long paddr = (pfn << PAGE_SHIFT); 8628be7072SBenjamin Herrenschmidt struct memblock_region *reg; 8714cf11afSPaul Mackerras 8828be7072SBenjamin Herrenschmidt for_each_memblock(memory, reg) 8928be7072SBenjamin Herrenschmidt if (paddr >= reg->base && paddr < (reg->base + reg->size)) 9014cf11afSPaul Mackerras return 1; 9114cf11afSPaul Mackerras return 0; 9214cf11afSPaul Mackerras #endif 9314cf11afSPaul Mackerras } 9414cf11afSPaul Mackerras 958b150478SRoland Dreier pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn, 9614cf11afSPaul Mackerras unsigned long size, pgprot_t vma_prot) 9714cf11afSPaul Mackerras { 9814cf11afSPaul Mackerras if (ppc_md.phys_mem_access_prot) 998b150478SRoland Dreier return ppc_md.phys_mem_access_prot(file, pfn, size, vma_prot); 10014cf11afSPaul Mackerras 1018b150478SRoland Dreier if (!page_is_ram(pfn)) 10264b3d0e8SBenjamin Herrenschmidt vma_prot = pgprot_noncached(vma_prot); 10364b3d0e8SBenjamin Herrenschmidt 10414cf11afSPaul Mackerras return vma_prot; 10514cf11afSPaul Mackerras } 10614cf11afSPaul Mackerras EXPORT_SYMBOL(phys_mem_access_prot); 10714cf11afSPaul Mackerras 10823fd0775SPaul Mackerras #ifdef CONFIG_MEMORY_HOTPLUG 10923fd0775SPaul Mackerras 110bc02af93SYasunori Goto #ifdef CONFIG_NUMA 111bc02af93SYasunori Goto int memory_add_physaddr_to_nid(u64 start) 112bc02af93SYasunori Goto { 113bc02af93SYasunori Goto return hot_add_scn_to_nid(start); 114bc02af93SYasunori Goto } 115bc02af93SYasunori Goto #endif 116bc02af93SYasunori Goto 117fa90f70aSGeert Uytterhoeven int arch_add_memory(int nid, u64 start, u64 size) 11823fd0775SPaul Mackerras { 119237a0989SMike Kravetz struct pglist_data *pgdata; 12023fd0775SPaul Mackerras struct zone *zone; 12123fd0775SPaul Mackerras unsigned long start_pfn = start >> PAGE_SHIFT; 12223fd0775SPaul Mackerras unsigned long nr_pages = size >> PAGE_SHIFT; 12323fd0775SPaul Mackerras 124237a0989SMike Kravetz pgdata = NODE_DATA(nid); 125237a0989SMike Kravetz 1262d0eee14SAndrew Morton start = (unsigned long)__va(start); 127a1194097SAnton Blanchard if (create_section_mapping(start, start + size)) 128a1194097SAnton Blanchard return -EINVAL; 12954b79248SMike Kravetz 13023fd0775SPaul Mackerras /* this should work for most non-highmem platforms */ 131f51202deSWang Nan zone = pgdata->node_zones + 132f51202deSWang Nan zone_for_memory(nid, start, size, 0); 13323fd0775SPaul Mackerras 134c04fc586SGary Hade return __add_pages(nid, zone, start_pfn, nr_pages); 13523fd0775SPaul Mackerras } 13624d335caSWen Congyang 13724d335caSWen Congyang #ifdef CONFIG_MEMORY_HOTREMOVE 13824d335caSWen Congyang int arch_remove_memory(u64 start, u64 size) 13924d335caSWen Congyang { 14024d335caSWen Congyang unsigned long start_pfn = start >> PAGE_SHIFT; 14124d335caSWen Congyang unsigned long nr_pages = size >> PAGE_SHIFT; 14224d335caSWen Congyang struct zone *zone; 1439ac8cde9SNathan Fontenot int ret; 14424d335caSWen Congyang 14524d335caSWen Congyang zone = page_zone(pfn_to_page(start_pfn)); 1469ac8cde9SNathan Fontenot ret = __remove_pages(zone, start_pfn, nr_pages); 14716d0f5c4SAnton Blanchard if (ret) 14816d0f5c4SAnton Blanchard return ret; 14916d0f5c4SAnton Blanchard 15016d0f5c4SAnton Blanchard /* Remove htab bolted mappings for this section of memory */ 15116d0f5c4SAnton Blanchard start = (unsigned long)__va(start); 15216d0f5c4SAnton Blanchard ret = remove_section_mapping(start, start + size); 15316d0f5c4SAnton Blanchard 15416d0f5c4SAnton Blanchard /* Ensure all vmalloc mappings are flushed in case they also 15516d0f5c4SAnton Blanchard * hit that section of memory 15616d0f5c4SAnton Blanchard */ 15716d0f5c4SAnton Blanchard vm_unmap_aliases(); 1589ac8cde9SNathan Fontenot 1599ac8cde9SNathan Fontenot return ret; 16024d335caSWen Congyang } 16124d335caSWen Congyang #endif 1620d579944SNathan Lynch #endif /* CONFIG_MEMORY_HOTPLUG */ 163a99824f3SBadari Pulavarty 164a99824f3SBadari Pulavarty /* 165a99824f3SBadari Pulavarty * walk_memory_resource() needs to make sure there is no holes in a given 1669d88a2ebSBadari Pulavarty * memory range. PPC64 does not maintain the memory layout in /proc/iomem. 16795f72d1eSYinghai Lu * Instead it maintains it in memblock.memory structures. Walk through the 1689d88a2ebSBadari Pulavarty * memory regions, find holes and callback for contiguous regions. 169a99824f3SBadari Pulavarty */ 170a99824f3SBadari Pulavarty int 171908eedc6SKAMEZAWA Hiroyuki walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, 172908eedc6SKAMEZAWA Hiroyuki void *arg, int (*func)(unsigned long, unsigned long, void *)) 173a99824f3SBadari Pulavarty { 17428be7072SBenjamin Herrenschmidt struct memblock_region *reg; 17528be7072SBenjamin Herrenschmidt unsigned long end_pfn = start_pfn + nr_pages; 17628be7072SBenjamin Herrenschmidt unsigned long tstart, tend; 1779d88a2ebSBadari Pulavarty int ret = -1; 1789d88a2ebSBadari Pulavarty 17928be7072SBenjamin Herrenschmidt for_each_memblock(memory, reg) { 180c7fc2de0SYinghai Lu tstart = max(start_pfn, memblock_region_memory_base_pfn(reg)); 181c7fc2de0SYinghai Lu tend = min(end_pfn, memblock_region_memory_end_pfn(reg)); 18228be7072SBenjamin Herrenschmidt if (tstart >= tend) 18328be7072SBenjamin Herrenschmidt continue; 18428be7072SBenjamin Herrenschmidt ret = (*func)(tstart, tend - tstart, arg); 1859d88a2ebSBadari Pulavarty if (ret) 1869d88a2ebSBadari Pulavarty break; 187a99824f3SBadari Pulavarty } 1889d88a2ebSBadari Pulavarty return ret; 1899d88a2ebSBadari Pulavarty } 190908eedc6SKAMEZAWA Hiroyuki EXPORT_SYMBOL_GPL(walk_system_ram_range); 191a99824f3SBadari Pulavarty 1927c8c6b97SPaul Mackerras #ifndef CONFIG_NEED_MULTIPLE_NODES 19310239733SAnton Blanchard void __init initmem_init(void) 1947c8c6b97SPaul Mackerras { 19595f72d1eSYinghai Lu max_low_pfn = max_pfn = memblock_end_of_DRAM() >> PAGE_SHIFT; 19610239733SAnton Blanchard min_low_pfn = MEMORY_START >> PAGE_SHIFT; 1977c8c6b97SPaul Mackerras #ifdef CONFIG_HIGHMEM 198d7917ba7SKumar Gala max_low_pfn = lowmem_end_addr >> PAGE_SHIFT; 1997c8c6b97SPaul Mackerras #endif 2007c8c6b97SPaul Mackerras 2014e8309baSCody P Schafer /* Place all memblock_regions in the same node and merge contiguous 2024e8309baSCody P Schafer * memblock_regions 2034e8309baSCody P Schafer */ 204e7e8de59STang Chen memblock_set_node(0, (phys_addr_t)ULLONG_MAX, &memblock.memory, 0); 205c67c3cb4SMel Gorman 2067c8c6b97SPaul Mackerras /* XXX need to clip this if using highmem? */ 207c67c3cb4SMel Gorman sparse_memory_present_with_active_regions(0); 208*21098b9eSAnton Blanchard sparse_init(); 2097c8c6b97SPaul Mackerras } 2107c8c6b97SPaul Mackerras 2114e8ad3e8SJohannes Berg /* mark pages that don't exist as nosave */ 2124e8ad3e8SJohannes Berg static int __init mark_nonram_nosave(void) 2134e8ad3e8SJohannes Berg { 21428be7072SBenjamin Herrenschmidt struct memblock_region *reg, *prev = NULL; 2154e8ad3e8SJohannes Berg 21628be7072SBenjamin Herrenschmidt for_each_memblock(memory, reg) { 21728be7072SBenjamin Herrenschmidt if (prev && 218c7fc2de0SYinghai Lu memblock_region_memory_end_pfn(prev) < memblock_region_memory_base_pfn(reg)) 219c7fc2de0SYinghai Lu register_nosave_region(memblock_region_memory_end_pfn(prev), 220c7fc2de0SYinghai Lu memblock_region_memory_base_pfn(reg)); 22128be7072SBenjamin Herrenschmidt prev = reg; 2224e8ad3e8SJohannes Berg } 2234e8ad3e8SJohannes Berg return 0; 2244e8ad3e8SJohannes Berg } 2256db35ad2SScott Wood #else /* CONFIG_NEED_MULTIPLE_NODES */ 2266db35ad2SScott Wood static int __init mark_nonram_nosave(void) 2276db35ad2SScott Wood { 2286db35ad2SScott Wood return 0; 2296db35ad2SScott Wood } 2306db35ad2SScott Wood #endif 2314e8ad3e8SJohannes Berg 2321c98025cSScott Wood static bool zone_limits_final; 2331c98025cSScott Wood 2341c98025cSScott Wood static unsigned long max_zone_pfns[MAX_NR_ZONES] = { 2351c98025cSScott Wood [0 ... MAX_NR_ZONES - 1] = ~0UL 2361c98025cSScott Wood }; 2371c98025cSScott Wood 2381c98025cSScott Wood /* 2391c98025cSScott Wood * Restrict the specified zone and all more restrictive zones 2401c98025cSScott Wood * to be below the specified pfn. May not be called after 2411c98025cSScott Wood * paging_init(). 2421c98025cSScott Wood */ 2431c98025cSScott Wood void __init limit_zone_pfn(enum zone_type zone, unsigned long pfn_limit) 2441c98025cSScott Wood { 2451c98025cSScott Wood int i; 2461c98025cSScott Wood 2471c98025cSScott Wood if (WARN_ON(zone_limits_final)) 2481c98025cSScott Wood return; 2491c98025cSScott Wood 2501c98025cSScott Wood for (i = zone; i >= 0; i--) { 2511c98025cSScott Wood if (max_zone_pfns[i] > pfn_limit) 2521c98025cSScott Wood max_zone_pfns[i] = pfn_limit; 2531c98025cSScott Wood } 2541c98025cSScott Wood } 2551c98025cSScott Wood 2561c98025cSScott Wood /* 2571c98025cSScott Wood * Find the least restrictive zone that is entirely below the 2581c98025cSScott Wood * specified pfn limit. Returns < 0 if no suitable zone is found. 2591c98025cSScott Wood * 2601c98025cSScott Wood * pfn_limit must be u64 because it can exceed 32 bits even on 32-bit 2611c98025cSScott Wood * systems -- the DMA limit can be higher than any possible real pfn. 2621c98025cSScott Wood */ 2631c98025cSScott Wood int dma_pfn_limit_to_zone(u64 pfn_limit) 2641c98025cSScott Wood { 2651c98025cSScott Wood enum zone_type top_zone = ZONE_NORMAL; 2661c98025cSScott Wood int i; 2671c98025cSScott Wood 2681c98025cSScott Wood #ifdef CONFIG_HIGHMEM 2691c98025cSScott Wood top_zone = ZONE_HIGHMEM; 2701c98025cSScott Wood #endif 2711c98025cSScott Wood 2721c98025cSScott Wood for (i = top_zone; i >= 0; i--) { 2731c98025cSScott Wood if (max_zone_pfns[i] <= pfn_limit) 2741c98025cSScott Wood return i; 2751c98025cSScott Wood } 2761c98025cSScott Wood 2771c98025cSScott Wood return -EPERM; 2781c98025cSScott Wood } 2797c8c6b97SPaul Mackerras 2807c8c6b97SPaul Mackerras /* 2817c8c6b97SPaul Mackerras * paging_init() sets up the page tables - in fact we've already done this. 2827c8c6b97SPaul Mackerras */ 2837c8c6b97SPaul Mackerras void __init paging_init(void) 2847c8c6b97SPaul Mackerras { 285f7ba2991STony Breeds unsigned long long total_ram = memblock_phys_mem_size(); 28695f72d1eSYinghai Lu phys_addr_t top_of_ram = memblock_end_of_DRAM(); 2871c98025cSScott Wood enum zone_type top_zone; 2887c8c6b97SPaul Mackerras 2892c419bdeSKumar Gala #ifdef CONFIG_PPC32 2902c419bdeSKumar Gala unsigned long v = __fix_to_virt(__end_of_fixed_addresses - 1); 2912c419bdeSKumar Gala unsigned long end = __fix_to_virt(FIX_HOLE); 2922c419bdeSKumar Gala 2932c419bdeSKumar Gala for (; v < end; v += PAGE_SIZE) 2942c419bdeSKumar Gala map_page(v, 0, 0); /* XXX gross */ 2952c419bdeSKumar Gala #endif 2962c419bdeSKumar Gala 2977c8c6b97SPaul Mackerras #ifdef CONFIG_HIGHMEM 2987c8c6b97SPaul Mackerras map_page(PKMAP_BASE, 0, 0); /* XXX gross */ 2992c419bdeSKumar Gala pkmap_page_table = virt_to_kpte(PKMAP_BASE); 3002c419bdeSKumar Gala 3012c419bdeSKumar Gala kmap_pte = virt_to_kpte(__fix_to_virt(FIX_KMAP_BEGIN)); 3027c8c6b97SPaul Mackerras kmap_prot = PAGE_KERNEL; 3037c8c6b97SPaul Mackerras #endif /* CONFIG_HIGHMEM */ 3047c8c6b97SPaul Mackerras 305f7ba2991STony Breeds printk(KERN_DEBUG "Top of RAM: 0x%llx, Total RAM: 0x%llx\n", 306fb610635STony Breeds (unsigned long long)top_of_ram, total_ram); 307e110b281SOlof Johansson printk(KERN_DEBUG "Memory hole size: %ldMB\n", 3082bf3016fSStefan Roese (long int)((top_of_ram - total_ram) >> 20)); 3091c98025cSScott Wood 3107c8c6b97SPaul Mackerras #ifdef CONFIG_HIGHMEM 3111c98025cSScott Wood top_zone = ZONE_HIGHMEM; 3121c98025cSScott Wood limit_zone_pfn(ZONE_NORMAL, lowmem_end_addr >> PAGE_SHIFT); 3137c8c6b97SPaul Mackerras #else 3141c98025cSScott Wood top_zone = ZONE_NORMAL; 315c67c3cb4SMel Gorman #endif 3161c98025cSScott Wood 3171c98025cSScott Wood limit_zone_pfn(top_zone, top_of_ram >> PAGE_SHIFT); 3181c98025cSScott Wood zone_limits_final = true; 319c67c3cb4SMel Gorman free_area_init_nodes(max_zone_pfns); 3204e8ad3e8SJohannes Berg 3214e8ad3e8SJohannes Berg mark_nonram_nosave(); 3227c8c6b97SPaul Mackerras } 3237c8c6b97SPaul Mackerras 3247c8c6b97SPaul Mackerras void __init mem_init(void) 3257c8c6b97SPaul Mackerras { 32628efc35fSScott Wood /* 32728efc35fSScott Wood * book3s is limited to 16 page sizes due to encoding this in 32828efc35fSScott Wood * a 4-bit field for slices. 32928efc35fSScott Wood */ 33028efc35fSScott Wood BUILD_BUG_ON(MMU_PAGE_COUNT > 16); 33128efc35fSScott Wood 332a9327296SFUJITA Tomonori #ifdef CONFIG_SWIOTLB 333688ba1dbSJia Hongtao swiotlb_init(0); 334a9327296SFUJITA Tomonori #endif 335a9327296SFUJITA Tomonori 3367c8c6b97SPaul Mackerras high_memory = (void *) __va(max_low_pfn * PAGE_SIZE); 337602ddc70SJiang Liu set_max_mapnr(max_pfn); 3380c988534SJiang Liu free_all_bootmem(); 3397c8c6b97SPaul Mackerras 3407c8c6b97SPaul Mackerras #ifdef CONFIG_HIGHMEM 3417c8c6b97SPaul Mackerras { 3427c8c6b97SPaul Mackerras unsigned long pfn, highmem_mapnr; 3437c8c6b97SPaul Mackerras 344d7917ba7SKumar Gala highmem_mapnr = lowmem_end_addr >> PAGE_SHIFT; 3457c8c6b97SPaul Mackerras for (pfn = highmem_mapnr; pfn < max_mapnr; ++pfn) { 3463d41e0f6SBecky Bruce phys_addr_t paddr = (phys_addr_t)pfn << PAGE_SHIFT; 3477c8c6b97SPaul Mackerras struct page *page = pfn_to_page(pfn); 348369a9d85SJiang Liu if (!memblock_is_reserved(paddr)) 3497db6f78cSJiang Liu free_highmem_page(page); 3507c8c6b97SPaul Mackerras } 3517c8c6b97SPaul Mackerras } 3527c8c6b97SPaul Mackerras #endif /* CONFIG_HIGHMEM */ 3537c8c6b97SPaul Mackerras 3543160b097SBecky Bruce #if defined(CONFIG_PPC_FSL_BOOK3E) && !defined(CONFIG_SMP) 3553160b097SBecky Bruce /* 3563160b097SBecky Bruce * If smp is enabled, next_tlbcam_idx is initialized in the cpu up 3573160b097SBecky Bruce * functions.... do it here for the non-smp case. 3583160b097SBecky Bruce */ 3593160b097SBecky Bruce per_cpu(next_tlbcam_idx, smp_processor_id()) = 3603160b097SBecky Bruce (mfspr(SPRN_TLB1CFG) & TLBnCFG_N_ENTRY) - 1; 3613160b097SBecky Bruce #endif 3623160b097SBecky Bruce 363369a9d85SJiang Liu mem_init_print_info(NULL); 364f637a49eSBenjamin Herrenschmidt #ifdef CONFIG_PPC32 365f637a49eSBenjamin Herrenschmidt pr_info("Kernel virtual memory layout:\n"); 366f637a49eSBenjamin Herrenschmidt pr_info(" * 0x%08lx..0x%08lx : fixmap\n", FIXADDR_START, FIXADDR_TOP); 367f637a49eSBenjamin Herrenschmidt #ifdef CONFIG_HIGHMEM 368f637a49eSBenjamin Herrenschmidt pr_info(" * 0x%08lx..0x%08lx : highmem PTEs\n", 369f637a49eSBenjamin Herrenschmidt PKMAP_BASE, PKMAP_ADDR(LAST_PKMAP)); 370f637a49eSBenjamin Herrenschmidt #endif /* CONFIG_HIGHMEM */ 3718b31e49dSBenjamin Herrenschmidt #ifdef CONFIG_NOT_COHERENT_CACHE 3728b31e49dSBenjamin Herrenschmidt pr_info(" * 0x%08lx..0x%08lx : consistent mem\n", 3738b31e49dSBenjamin Herrenschmidt IOREMAP_TOP, IOREMAP_TOP + CONFIG_CONSISTENT_SIZE); 3748b31e49dSBenjamin Herrenschmidt #endif /* CONFIG_NOT_COHERENT_CACHE */ 375f637a49eSBenjamin Herrenschmidt pr_info(" * 0x%08lx..0x%08lx : early ioremap\n", 376f637a49eSBenjamin Herrenschmidt ioremap_bot, IOREMAP_TOP); 377f637a49eSBenjamin Herrenschmidt pr_info(" * 0x%08lx..0x%08lx : vmalloc & ioremap\n", 378f637a49eSBenjamin Herrenschmidt VMALLOC_START, VMALLOC_END); 379f637a49eSBenjamin Herrenschmidt #endif /* CONFIG_PPC32 */ 380f637a49eSBenjamin Herrenschmidt 3817c8c6b97SPaul Mackerras mem_init_done = 1; 3827c8c6b97SPaul Mackerras } 3837c8c6b97SPaul Mackerras 3842773fcc8SDave Carroll void free_initmem(void) 3852773fcc8SDave Carroll { 386a9c0f41bSDave Carroll ppc_md.progress = ppc_printk_progress; 3875d585e5cSJiang Liu free_initmem_default(POISON_FREE_INITMEM); 3882773fcc8SDave Carroll } 3892773fcc8SDave Carroll 390307cfe71SBenjamin Herrenschmidt #ifdef CONFIG_BLK_DEV_INITRD 391307cfe71SBenjamin Herrenschmidt void __init free_initrd_mem(unsigned long start, unsigned long end) 392307cfe71SBenjamin Herrenschmidt { 393dbe67df4SJiang Liu free_reserved_area((void *)start, (void *)end, -1, "initrd"); 394307cfe71SBenjamin Herrenschmidt } 395307cfe71SBenjamin Herrenschmidt #endif 396307cfe71SBenjamin Herrenschmidt 3977c8c6b97SPaul Mackerras /* 39814cf11afSPaul Mackerras * This is called when a page has been modified by the kernel. 39914cf11afSPaul Mackerras * It just marks the page as not i-cache clean. We do the i-cache 40014cf11afSPaul Mackerras * flush later when the page is given to a user process, if necessary. 40114cf11afSPaul Mackerras */ 40214cf11afSPaul Mackerras void flush_dcache_page(struct page *page) 40314cf11afSPaul Mackerras { 40414cf11afSPaul Mackerras if (cpu_has_feature(CPU_FTR_COHERENT_ICACHE)) 40514cf11afSPaul Mackerras return; 40614cf11afSPaul Mackerras /* avoid an atomic op if possible */ 40714cf11afSPaul Mackerras if (test_bit(PG_arch_1, &page->flags)) 40814cf11afSPaul Mackerras clear_bit(PG_arch_1, &page->flags); 40914cf11afSPaul Mackerras } 41014cf11afSPaul Mackerras EXPORT_SYMBOL(flush_dcache_page); 41114cf11afSPaul Mackerras 41214cf11afSPaul Mackerras void flush_dcache_icache_page(struct page *page) 41314cf11afSPaul Mackerras { 4140895ecdaSDavid Gibson #ifdef CONFIG_HUGETLB_PAGE 4150895ecdaSDavid Gibson if (PageCompound(page)) { 4160895ecdaSDavid Gibson flush_dcache_icache_hugepage(page); 4170895ecdaSDavid Gibson return; 4180895ecdaSDavid Gibson } 4190895ecdaSDavid Gibson #endif 42014cf11afSPaul Mackerras #ifdef CONFIG_BOOKE 4210895ecdaSDavid Gibson { 4222480b208SCong Wang void *start = kmap_atomic(page); 42314cf11afSPaul Mackerras __flush_dcache_icache(start); 4242480b208SCong Wang kunmap_atomic(start); 4250895ecdaSDavid Gibson } 426ab1f9dacSPaul Mackerras #elif defined(CONFIG_8xx) || defined(CONFIG_PPC64) 42714cf11afSPaul Mackerras /* On 8xx there is no need to kmap since highmem is not supported */ 42814cf11afSPaul Mackerras __flush_dcache_icache(page_address(page)); 42914cf11afSPaul Mackerras #else 43014cf11afSPaul Mackerras __flush_dcache_icache_phys(page_to_pfn(page) << PAGE_SHIFT); 43114cf11afSPaul Mackerras #endif 43214cf11afSPaul Mackerras } 433249ba1eeSAlexander Graf EXPORT_SYMBOL(flush_dcache_icache_page); 4340895ecdaSDavid Gibson 43514cf11afSPaul Mackerras void clear_user_page(void *page, unsigned long vaddr, struct page *pg) 43614cf11afSPaul Mackerras { 43714cf11afSPaul Mackerras clear_page(page); 43814cf11afSPaul Mackerras 43914cf11afSPaul Mackerras /* 44025985edcSLucas De Marchi * We shouldn't have to do this, but some versions of glibc 44114cf11afSPaul Mackerras * require it (ld.so assumes zero filled pages are icache clean) 44214cf11afSPaul Mackerras * - Anton 44314cf11afSPaul Mackerras */ 44409f5dc44SDavid Gibson flush_dcache_page(pg); 44514cf11afSPaul Mackerras } 44614cf11afSPaul Mackerras EXPORT_SYMBOL(clear_user_page); 44714cf11afSPaul Mackerras 44814cf11afSPaul Mackerras void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, 44914cf11afSPaul Mackerras struct page *pg) 45014cf11afSPaul Mackerras { 45114cf11afSPaul Mackerras copy_page(vto, vfrom); 45214cf11afSPaul Mackerras 45314cf11afSPaul Mackerras /* 45414cf11afSPaul Mackerras * We should be able to use the following optimisation, however 45514cf11afSPaul Mackerras * there are two problems. 45614cf11afSPaul Mackerras * Firstly a bug in some versions of binutils meant PLT sections 45714cf11afSPaul Mackerras * were not marked executable. 45814cf11afSPaul Mackerras * Secondly the first word in the GOT section is blrl, used 45914cf11afSPaul Mackerras * to establish the GOT address. Until recently the GOT was 46014cf11afSPaul Mackerras * not marked executable. 46114cf11afSPaul Mackerras * - Anton 46214cf11afSPaul Mackerras */ 46314cf11afSPaul Mackerras #if 0 46414cf11afSPaul Mackerras if (!vma->vm_file && ((vma->vm_flags & VM_EXEC) == 0)) 46514cf11afSPaul Mackerras return; 46614cf11afSPaul Mackerras #endif 46714cf11afSPaul Mackerras 46809f5dc44SDavid Gibson flush_dcache_page(pg); 46914cf11afSPaul Mackerras } 47014cf11afSPaul Mackerras 47114cf11afSPaul Mackerras void flush_icache_user_range(struct vm_area_struct *vma, struct page *page, 47214cf11afSPaul Mackerras unsigned long addr, int len) 47314cf11afSPaul Mackerras { 47414cf11afSPaul Mackerras unsigned long maddr; 47514cf11afSPaul Mackerras 47614cf11afSPaul Mackerras maddr = (unsigned long) kmap(page) + (addr & ~PAGE_MASK); 47714cf11afSPaul Mackerras flush_icache_range(maddr, maddr + len); 47814cf11afSPaul Mackerras kunmap(page); 47914cf11afSPaul Mackerras } 48014cf11afSPaul Mackerras EXPORT_SYMBOL(flush_icache_user_range); 48114cf11afSPaul Mackerras 48214cf11afSPaul Mackerras /* 48314cf11afSPaul Mackerras * This is called at the end of handling a user page fault, when the 48414cf11afSPaul Mackerras * fault has been handled by updating a PTE in the linux page tables. 48514cf11afSPaul Mackerras * We use it to preload an HPTE into the hash table corresponding to 48614cf11afSPaul Mackerras * the updated linux PTE. 48714cf11afSPaul Mackerras * 48801edcd89SHugh Dickins * This must always be called with the pte lock held. 48914cf11afSPaul Mackerras */ 49014cf11afSPaul Mackerras void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, 4914b3073e1SRussell King pte_t *ptep) 49214cf11afSPaul Mackerras { 4933c726f8dSBenjamin Herrenschmidt #ifdef CONFIG_PPC_STD_MMU 4940ac52dd7SAneesh Kumar K.V /* 4950ac52dd7SAneesh Kumar K.V * We don't need to worry about _PAGE_PRESENT here because we are 4960ac52dd7SAneesh Kumar K.V * called with either mm->page_table_lock held or ptl lock held 4970ac52dd7SAneesh Kumar K.V */ 4983c726f8dSBenjamin Herrenschmidt unsigned long access = 0, trap; 49914cf11afSPaul Mackerras 50014cf11afSPaul Mackerras /* We only want HPTEs for linux PTEs that have _PAGE_ACCESSED set */ 5014b3073e1SRussell King if (!pte_young(*ptep) || address >= TASK_SIZE) 50214cf11afSPaul Mackerras return; 5033c726f8dSBenjamin Herrenschmidt 5043c726f8dSBenjamin Herrenschmidt /* We try to figure out if we are coming from an instruction 5053c726f8dSBenjamin Herrenschmidt * access fault and pass that down to __hash_page so we avoid 5063c726f8dSBenjamin Herrenschmidt * double-faulting on execution of fresh text. We have to test 5073c726f8dSBenjamin Herrenschmidt * for regs NULL since init will get here first thing at boot 5083c726f8dSBenjamin Herrenschmidt * 5093c726f8dSBenjamin Herrenschmidt * We also avoid filling the hash if not coming from a fault 5103c726f8dSBenjamin Herrenschmidt */ 5113c726f8dSBenjamin Herrenschmidt if (current->thread.regs == NULL) 51214cf11afSPaul Mackerras return; 5133c726f8dSBenjamin Herrenschmidt trap = TRAP(current->thread.regs); 5143c726f8dSBenjamin Herrenschmidt if (trap == 0x400) 5153c726f8dSBenjamin Herrenschmidt access |= _PAGE_EXEC; 5163c726f8dSBenjamin Herrenschmidt else if (trap != 0x300) 51714cf11afSPaul Mackerras return; 5183c726f8dSBenjamin Herrenschmidt hash_preload(vma->vm_mm, address, access, trap); 5193c726f8dSBenjamin Herrenschmidt #endif /* CONFIG_PPC_STD_MMU */ 52041151e77SBecky Bruce #if (defined(CONFIG_PPC_BOOK3E_64) || defined(CONFIG_PPC_FSL_BOOK3E)) \ 52141151e77SBecky Bruce && defined(CONFIG_HUGETLB_PAGE) 52241151e77SBecky Bruce if (is_vm_hugetlb_page(vma)) 523d93e4d7dSBecky Bruce book3e_hugetlb_preload(vma, address, *ptep); 52441151e77SBecky Bruce #endif 52514cf11afSPaul Mackerras } 526c40dd2f7SAnton Blanchard 527c40dd2f7SAnton Blanchard /* 528c40dd2f7SAnton Blanchard * System memory should not be in /proc/iomem but various tools expect it 529c40dd2f7SAnton Blanchard * (eg kdump). 530c40dd2f7SAnton Blanchard */ 5314f770924SGeert Uytterhoeven static int __init add_system_ram_resources(void) 532c40dd2f7SAnton Blanchard { 533c40dd2f7SAnton Blanchard struct memblock_region *reg; 534c40dd2f7SAnton Blanchard 535c40dd2f7SAnton Blanchard for_each_memblock(memory, reg) { 536c40dd2f7SAnton Blanchard struct resource *res; 537c40dd2f7SAnton Blanchard unsigned long base = reg->base; 538c40dd2f7SAnton Blanchard unsigned long size = reg->size; 539c40dd2f7SAnton Blanchard 540c40dd2f7SAnton Blanchard res = kzalloc(sizeof(struct resource), GFP_KERNEL); 541c40dd2f7SAnton Blanchard WARN_ON(!res); 542c40dd2f7SAnton Blanchard 543c40dd2f7SAnton Blanchard if (res) { 544c40dd2f7SAnton Blanchard res->name = "System RAM"; 545c40dd2f7SAnton Blanchard res->start = base; 546c40dd2f7SAnton Blanchard res->end = base + size - 1; 547f748edafSNathan Fontenot res->flags = IORESOURCE_MEM | IORESOURCE_BUSY; 548c40dd2f7SAnton Blanchard WARN_ON(request_resource(&iomem_resource, res) < 0); 549c40dd2f7SAnton Blanchard } 550c40dd2f7SAnton Blanchard } 551c40dd2f7SAnton Blanchard 552c40dd2f7SAnton Blanchard return 0; 553c40dd2f7SAnton Blanchard } 554c40dd2f7SAnton Blanchard subsys_initcall(add_system_ram_resources); 5551d54cf2bSsukadev@linux.vnet.ibm.com 5561d54cf2bSsukadev@linux.vnet.ibm.com #ifdef CONFIG_STRICT_DEVMEM 5571d54cf2bSsukadev@linux.vnet.ibm.com /* 5581d54cf2bSsukadev@linux.vnet.ibm.com * devmem_is_allowed(): check to see if /dev/mem access to a certain address 5591d54cf2bSsukadev@linux.vnet.ibm.com * is valid. The argument is a physical page number. 5601d54cf2bSsukadev@linux.vnet.ibm.com * 5611d54cf2bSsukadev@linux.vnet.ibm.com * Access has to be given to non-kernel-ram areas as well, these contain the 5621d54cf2bSsukadev@linux.vnet.ibm.com * PCI mmio resources as well as potential bios/acpi data regions. 5631d54cf2bSsukadev@linux.vnet.ibm.com */ 5641d54cf2bSsukadev@linux.vnet.ibm.com int devmem_is_allowed(unsigned long pfn) 5651d54cf2bSsukadev@linux.vnet.ibm.com { 5661d54cf2bSsukadev@linux.vnet.ibm.com if (iomem_is_exclusive(pfn << PAGE_SHIFT)) 5671d54cf2bSsukadev@linux.vnet.ibm.com return 0; 5681d54cf2bSsukadev@linux.vnet.ibm.com if (!page_is_ram(pfn)) 5691d54cf2bSsukadev@linux.vnet.ibm.com return 1; 5708a3e3d31Ssukadev@linux.vnet.ibm.com if (page_is_rtas_user_buf(pfn)) 5718a3e3d31Ssukadev@linux.vnet.ibm.com return 1; 5721d54cf2bSsukadev@linux.vnet.ibm.com return 0; 5731d54cf2bSsukadev@linux.vnet.ibm.com } 5741d54cf2bSsukadev@linux.vnet.ibm.com #endif /* CONFIG_STRICT_DEVMEM */ 575