1b2441318SGreg Kroah-Hartman // SPDX-License-Identifier: GPL-2.0 288df125fSHeiko Carstens /* 388df125fSHeiko Carstens * Access kernel memory without faulting -- s390 specific implementation. 488df125fSHeiko Carstens * 53c1a3bceSHeiko Carstens * Copyright IBM Corp. 2009, 2015 688df125fSHeiko Carstens * 788df125fSHeiko Carstens * Author(s): Heiko Carstens <heiko.carstens@de.ibm.com>, 888df125fSHeiko Carstens * 988df125fSHeiko Carstens */ 1088df125fSHeiko Carstens 1188df125fSHeiko Carstens #include <linux/uaccess.h> 1288df125fSHeiko Carstens #include <linux/kernel.h> 1388df125fSHeiko Carstens #include <linux/types.h> 1488df125fSHeiko Carstens #include <linux/errno.h> 157f0bf656SMichael Holzheu #include <linux/gfp.h> 16b2a68c23SMichael Holzheu #include <linux/cpu.h> 17a0616cdeSDavid Howells #include <asm/ctl_reg.h> 1863df41d6SHeiko Carstens #include <asm/io.h> 1978c98f90SMartin Schwidefsky #include <asm/stacktrace.h> 2088df125fSHeiko Carstens 218a5d8473SHeiko Carstens static notrace long s390_kernel_write_odd(void *dst, const void *src, size_t size) 2288df125fSHeiko Carstens { 233c1a3bceSHeiko Carstens unsigned long aligned, offset, count; 243c1a3bceSHeiko Carstens char tmp[8]; 2588df125fSHeiko Carstens 263c1a3bceSHeiko Carstens aligned = (unsigned long) dst & ~7UL; 273c1a3bceSHeiko Carstens offset = (unsigned long) dst & 7UL; 283c1a3bceSHeiko Carstens size = min(8UL - offset, size); 293c1a3bceSHeiko Carstens count = size - 1; 3088df125fSHeiko Carstens asm volatile( 3188df125fSHeiko Carstens " bras 1,0f\n" 323c1a3bceSHeiko Carstens " mvc 0(1,%4),0(%5)\n" 333c1a3bceSHeiko Carstens "0: mvc 0(8,%3),0(%0)\n" 343c1a3bceSHeiko Carstens " ex %1,0(1)\n" 353c1a3bceSHeiko Carstens " lg %1,0(%3)\n" 363c1a3bceSHeiko Carstens " lra %0,0(%0)\n" 373c1a3bceSHeiko Carstens " sturg %1,%0\n" 383c1a3bceSHeiko Carstens : "+&a" (aligned), "+&a" (count), "=m" (tmp) 393c1a3bceSHeiko Carstens : "a" (&tmp), "a" (&tmp[offset]), "a" (src) 403c1a3bceSHeiko Carstens : "cc", "memory", "1"); 413c1a3bceSHeiko Carstens return size; 4288df125fSHeiko Carstens } 4388df125fSHeiko Carstens 448a5d8473SHeiko Carstens /* 458a5d8473SHeiko Carstens * s390_kernel_write - write to kernel memory bypassing DAT 468a5d8473SHeiko Carstens * @dst: destination address 478a5d8473SHeiko Carstens * @src: source address 488a5d8473SHeiko Carstens * @size: number of bytes to copy 498a5d8473SHeiko Carstens * 508a5d8473SHeiko Carstens * This function writes to kernel memory bypassing DAT and possible page table 518a5d8473SHeiko Carstens * write protection. It writes to the destination using the sturg instruction. 523c1a3bceSHeiko Carstens * Therefore we have a read-modify-write sequence: the function reads eight 533c1a3bceSHeiko Carstens * bytes from destination at an eight byte boundary, modifies the bytes 548a5d8473SHeiko Carstens * requested and writes the result back in a loop. 558a5d8473SHeiko Carstens */ 56*a646ef39SMartin Schwidefsky static DEFINE_SPINLOCK(s390_kernel_write_lock); 57*a646ef39SMartin Schwidefsky 588a5d8473SHeiko Carstens void notrace s390_kernel_write(void *dst, const void *src, size_t size) 5988df125fSHeiko Carstens { 60*a646ef39SMartin Schwidefsky unsigned long flags; 613c1a3bceSHeiko Carstens long copied; 6288df125fSHeiko Carstens 63*a646ef39SMartin Schwidefsky spin_lock_irqsave(&s390_kernel_write_lock, flags); 6488df125fSHeiko Carstens while (size) { 658a5d8473SHeiko Carstens copied = s390_kernel_write_odd(dst, src, size); 6688df125fSHeiko Carstens dst += copied; 6788df125fSHeiko Carstens src += copied; 6888df125fSHeiko Carstens size -= copied; 6988df125fSHeiko Carstens } 70*a646ef39SMartin Schwidefsky spin_unlock_irqrestore(&s390_kernel_write_lock, flags); 7188df125fSHeiko Carstens } 7292fe3132SMichael Holzheu 73b785e0d0SMichael Holzheu static int __memcpy_real(void *dest, void *src, size_t count) 7492fe3132SMichael Holzheu { 7592fe3132SMichael Holzheu register unsigned long _dest asm("2") = (unsigned long) dest; 7692fe3132SMichael Holzheu register unsigned long _len1 asm("3") = (unsigned long) count; 7792fe3132SMichael Holzheu register unsigned long _src asm("4") = (unsigned long) src; 7892fe3132SMichael Holzheu register unsigned long _len2 asm("5") = (unsigned long) count; 7992fe3132SMichael Holzheu int rc = -EFAULT; 8092fe3132SMichael Holzheu 8192fe3132SMichael Holzheu asm volatile ( 8292fe3132SMichael Holzheu "0: mvcle %1,%2,0x0\n" 8392fe3132SMichael Holzheu "1: jo 0b\n" 8492fe3132SMichael Holzheu " lhi %0,0x0\n" 8592fe3132SMichael Holzheu "2:\n" 8692fe3132SMichael Holzheu EX_TABLE(1b,2b) 8792fe3132SMichael Holzheu : "+d" (rc), "+d" (_dest), "+d" (_src), "+d" (_len1), 8892fe3132SMichael Holzheu "+d" (_len2), "=m" (*((long *) dest)) 8992fe3132SMichael Holzheu : "m" (*((long *) src)) 9092fe3132SMichael Holzheu : "cc", "memory"); 91b785e0d0SMichael Holzheu return rc; 92b785e0d0SMichael Holzheu } 93b785e0d0SMichael Holzheu 94ce3dc447SMartin Schwidefsky static unsigned long _memcpy_real(unsigned long dest, unsigned long src, 95ce3dc447SMartin Schwidefsky unsigned long count) 96b785e0d0SMichael Holzheu { 9752499d93SHeiko Carstens int irqs_disabled, rc; 98b785e0d0SMichael Holzheu unsigned long flags; 99b785e0d0SMichael Holzheu 100b785e0d0SMichael Holzheu if (!count) 101b785e0d0SMichael Holzheu return 0; 10252499d93SHeiko Carstens flags = __arch_local_irq_stnsm(0xf8UL); 10352499d93SHeiko Carstens irqs_disabled = arch_irqs_disabled_flags(flags); 10452499d93SHeiko Carstens if (!irqs_disabled) 10552499d93SHeiko Carstens trace_hardirqs_off(); 106ce3dc447SMartin Schwidefsky rc = __memcpy_real((void *) dest, (void *) src, (size_t) count); 10752499d93SHeiko Carstens if (!irqs_disabled) 1080986d977SChristian Borntraeger trace_hardirqs_on(); 1090986d977SChristian Borntraeger __arch_local_irq_ssm(flags); 11092fe3132SMichael Holzheu return rc; 11192fe3132SMichael Holzheu } 1127dd6b334SMichael Holzheu 1137dd6b334SMichael Holzheu /* 114ce3dc447SMartin Schwidefsky * Copy memory in real mode (kernel to kernel) 115ce3dc447SMartin Schwidefsky */ 116ce3dc447SMartin Schwidefsky int memcpy_real(void *dest, void *src, size_t count) 117ce3dc447SMartin Schwidefsky { 118ce3dc447SMartin Schwidefsky if (S390_lowcore.nodat_stack != 0) 119ce3dc447SMartin Schwidefsky return CALL_ON_STACK(_memcpy_real, S390_lowcore.nodat_stack, 120ce3dc447SMartin Schwidefsky 3, dest, src, count); 121ce3dc447SMartin Schwidefsky /* 122ce3dc447SMartin Schwidefsky * This is a really early memcpy_real call, the stacks are 123ce3dc447SMartin Schwidefsky * not set up yet. Just call _memcpy_real on the early boot 124ce3dc447SMartin Schwidefsky * stack 125ce3dc447SMartin Schwidefsky */ 126ce3dc447SMartin Schwidefsky return _memcpy_real((unsigned long) dest,(unsigned long) src, 127ce3dc447SMartin Schwidefsky (unsigned long) count); 128ce3dc447SMartin Schwidefsky } 129ce3dc447SMartin Schwidefsky 130ce3dc447SMartin Schwidefsky /* 13173bf463eSMichael Holzheu * Copy memory in absolute mode (kernel to kernel) 1327dd6b334SMichael Holzheu */ 13373bf463eSMichael Holzheu void memcpy_absolute(void *dest, void *src, size_t count) 1347dd6b334SMichael Holzheu { 13573bf463eSMichael Holzheu unsigned long cr0, flags, prefix; 1367dd6b334SMichael Holzheu 13773bf463eSMichael Holzheu flags = arch_local_irq_save(); 1387dd6b334SMichael Holzheu __ctl_store(cr0, 0, 0); 1397dd6b334SMichael Holzheu __ctl_clear_bit(0, 28); /* disable lowcore protection */ 14073bf463eSMichael Holzheu prefix = store_prefix(); 14173bf463eSMichael Holzheu if (prefix) { 14273bf463eSMichael Holzheu local_mcck_disable(); 14373bf463eSMichael Holzheu set_prefix(0); 14473bf463eSMichael Holzheu memcpy(dest, src, count); 14573bf463eSMichael Holzheu set_prefix(prefix); 14673bf463eSMichael Holzheu local_mcck_enable(); 14773bf463eSMichael Holzheu } else { 14873bf463eSMichael Holzheu memcpy(dest, src, count); 14973bf463eSMichael Holzheu } 1507dd6b334SMichael Holzheu __ctl_load(cr0, 0, 0); 15173bf463eSMichael Holzheu arch_local_irq_restore(flags); 1527dd6b334SMichael Holzheu } 1537f0bf656SMichael Holzheu 1547f0bf656SMichael Holzheu /* 1557f0bf656SMichael Holzheu * Copy memory from kernel (real) to user (virtual) 1567f0bf656SMichael Holzheu */ 157211deca6SHeiko Carstens int copy_to_user_real(void __user *dest, void *src, unsigned long count) 1587f0bf656SMichael Holzheu { 1597f0bf656SMichael Holzheu int offs = 0, size, rc; 1607f0bf656SMichael Holzheu char *buf; 1617f0bf656SMichael Holzheu 1627f0bf656SMichael Holzheu buf = (char *) __get_free_page(GFP_KERNEL); 1637f0bf656SMichael Holzheu if (!buf) 1647f0bf656SMichael Holzheu return -ENOMEM; 1657f0bf656SMichael Holzheu rc = -EFAULT; 1667f0bf656SMichael Holzheu while (offs < count) { 1677f0bf656SMichael Holzheu size = min(PAGE_SIZE, count - offs); 1687f0bf656SMichael Holzheu if (memcpy_real(buf, src + offs, size)) 1697f0bf656SMichael Holzheu goto out; 1707f0bf656SMichael Holzheu if (copy_to_user(dest + offs, buf, size)) 1717f0bf656SMichael Holzheu goto out; 1727f0bf656SMichael Holzheu offs += size; 1737f0bf656SMichael Holzheu } 1747f0bf656SMichael Holzheu rc = 0; 1757f0bf656SMichael Holzheu out: 1767f0bf656SMichael Holzheu free_page((unsigned long) buf); 1777f0bf656SMichael Holzheu return rc; 1787f0bf656SMichael Holzheu } 1797f0bf656SMichael Holzheu 1807f0bf656SMichael Holzheu /* 181b2a68c23SMichael Holzheu * Check if physical address is within prefix or zero page 182b2a68c23SMichael Holzheu */ 183b2a68c23SMichael Holzheu static int is_swapped(unsigned long addr) 184b2a68c23SMichael Holzheu { 185b2a68c23SMichael Holzheu unsigned long lc; 186b2a68c23SMichael Holzheu int cpu; 187b2a68c23SMichael Holzheu 188c667aeacSHeiko Carstens if (addr < sizeof(struct lowcore)) 189b2a68c23SMichael Holzheu return 1; 190b2a68c23SMichael Holzheu for_each_online_cpu(cpu) { 191b2a68c23SMichael Holzheu lc = (unsigned long) lowcore_ptr[cpu]; 192c667aeacSHeiko Carstens if (addr > lc + sizeof(struct lowcore) - 1 || addr < lc) 193b2a68c23SMichael Holzheu continue; 194b2a68c23SMichael Holzheu return 1; 195b2a68c23SMichael Holzheu } 196b2a68c23SMichael Holzheu return 0; 197b2a68c23SMichael Holzheu } 198b2a68c23SMichael Holzheu 199b2a68c23SMichael Holzheu /* 200b2a68c23SMichael Holzheu * Convert a physical pointer for /dev/mem access 201b2a68c23SMichael Holzheu * 202b2a68c23SMichael Holzheu * For swapped prefix pages a new buffer is returned that contains a copy of 203b2a68c23SMichael Holzheu * the absolute memory. The buffer size is maximum one page large. 204b2a68c23SMichael Holzheu */ 2054707a341SThierry Reding void *xlate_dev_mem_ptr(phys_addr_t addr) 206b2a68c23SMichael Holzheu { 207b2a68c23SMichael Holzheu void *bounce = (void *) addr; 208b2a68c23SMichael Holzheu unsigned long size; 209b2a68c23SMichael Holzheu 210b2a68c23SMichael Holzheu get_online_cpus(); 211b2a68c23SMichael Holzheu preempt_disable(); 212b2a68c23SMichael Holzheu if (is_swapped(addr)) { 213b2a68c23SMichael Holzheu size = PAGE_SIZE - (addr & ~PAGE_MASK); 214b2a68c23SMichael Holzheu bounce = (void *) __get_free_page(GFP_ATOMIC); 215b2a68c23SMichael Holzheu if (bounce) 21673bf463eSMichael Holzheu memcpy_absolute(bounce, (void *) addr, size); 217b2a68c23SMichael Holzheu } 218b2a68c23SMichael Holzheu preempt_enable(); 219b2a68c23SMichael Holzheu put_online_cpus(); 220b2a68c23SMichael Holzheu return bounce; 221b2a68c23SMichael Holzheu } 222b2a68c23SMichael Holzheu 223b2a68c23SMichael Holzheu /* 224b2a68c23SMichael Holzheu * Free converted buffer for /dev/mem access (if necessary) 225b2a68c23SMichael Holzheu */ 2264707a341SThierry Reding void unxlate_dev_mem_ptr(phys_addr_t addr, void *buf) 227b2a68c23SMichael Holzheu { 228b2a68c23SMichael Holzheu if ((void *) addr != buf) 229b2a68c23SMichael Holzheu free_page((unsigned long) buf); 230b2a68c23SMichael Holzheu } 231