1deae26bfSKyle McMartin #ifndef __PARISC_UACCESS_H 2deae26bfSKyle McMartin #define __PARISC_UACCESS_H 3deae26bfSKyle McMartin 4deae26bfSKyle McMartin /* 5deae26bfSKyle McMartin * User space memory access functions 6deae26bfSKyle McMartin */ 7deae26bfSKyle McMartin #include <asm/page.h> 8deae26bfSKyle McMartin #include <asm/cache.h> 9888c31fcSHelge Deller #include <asm/errno.h> 105b17e1cdSArnd Bergmann #include <asm-generic/uaccess-unaligned.h> 11deae26bfSKyle McMartin 128dd95c68SHelge Deller #include <linux/bug.h> 13aace880fSAl Viro #include <linux/string.h> 14*9e91db6bSHelge Deller #include <linux/thread_info.h> 158dd95c68SHelge Deller 16deae26bfSKyle McMartin #define VERIFY_READ 0 17deae26bfSKyle McMartin #define VERIFY_WRITE 1 18deae26bfSKyle McMartin 19deae26bfSKyle McMartin #define KERNEL_DS ((mm_segment_t){0}) 20deae26bfSKyle McMartin #define USER_DS ((mm_segment_t){1}) 21deae26bfSKyle McMartin 22deae26bfSKyle McMartin #define segment_eq(a, b) ((a).seg == (b).seg) 23deae26bfSKyle McMartin 24deae26bfSKyle McMartin #define get_ds() (KERNEL_DS) 25deae26bfSKyle McMartin #define get_fs() (current_thread_info()->addr_limit) 26deae26bfSKyle McMartin #define set_fs(x) (current_thread_info()->addr_limit = (x)) 27deae26bfSKyle McMartin 28deae26bfSKyle McMartin /* 29deae26bfSKyle McMartin * Note that since kernel addresses are in a separate address space on 30deae26bfSKyle McMartin * parisc, we don't need to do anything for access_ok(). 31deae26bfSKyle McMartin * We just let the page fault handler do the right thing. This also means 32deae26bfSKyle McMartin * that put_user is the same as __put_user, etc. 33deae26bfSKyle McMartin */ 34deae26bfSKyle McMartin 35a0ffa8f0SHelge Deller static inline long access_ok(int type, const void __user * addr, 36a0ffa8f0SHelge Deller unsigned long size) 37deae26bfSKyle McMartin { 38a0ffa8f0SHelge Deller return 1; 39deae26bfSKyle McMartin } 40deae26bfSKyle McMartin 41deae26bfSKyle McMartin #define put_user __put_user 42deae26bfSKyle McMartin #define get_user __get_user 43deae26bfSKyle McMartin 44deae26bfSKyle McMartin #if !defined(CONFIG_64BIT) 45d2ad824fSHelge Deller #define LDD_USER(ptr) __get_user_asm64(ptr) 46deae26bfSKyle McMartin #define STD_USER(x, ptr) __put_user_asm64(x, ptr) 47deae26bfSKyle McMartin #else 48deae26bfSKyle McMartin #define LDD_USER(ptr) __get_user_asm("ldd", ptr) 49deae26bfSKyle McMartin #define STD_USER(x, ptr) __put_user_asm("std", x, ptr) 50deae26bfSKyle McMartin #endif 51deae26bfSKyle McMartin 52deae26bfSKyle McMartin /* 53cb910c17SHelge Deller * The exception table contains two values: the first is the relative offset to 54cb910c17SHelge Deller * the address of the instruction that is allowed to fault, and the second is 55cb910c17SHelge Deller * the relative offset to the address of the fixup routine. Since relative 56cb910c17SHelge Deller * addresses are used, 32bit values are sufficient even on 64bit kernel. 57deae26bfSKyle McMartin */ 58deae26bfSKyle McMartin 590de79858SHelge Deller #define ARCH_HAS_RELATIVE_EXTABLE 60deae26bfSKyle McMartin struct exception_table_entry { 610de79858SHelge Deller int insn; /* relative address of insn that is allowed to fault. */ 620de79858SHelge Deller int fixup; /* relative address of fixup routine */ 63deae26bfSKyle McMartin }; 64deae26bfSKyle McMartin 65deae26bfSKyle McMartin #define ASM_EXCEPTIONTABLE_ENTRY( fault_addr, except_addr )\ 66deae26bfSKyle McMartin ".section __ex_table,\"aw\"\n" \ 670de79858SHelge Deller ".word (" #fault_addr " - .), (" #except_addr " - .)\n\t" \ 68deae26bfSKyle McMartin ".previous\n" 69deae26bfSKyle McMartin 70deae26bfSKyle McMartin /* 71deae26bfSKyle McMartin * The page fault handler stores, in a per-cpu area, the following information 72deae26bfSKyle McMartin * if a fixup routine is available. 73deae26bfSKyle McMartin */ 74deae26bfSKyle McMartin struct exception_data { 75deae26bfSKyle McMartin unsigned long fault_ip; 762ef4dfd9SHelge Deller unsigned long fault_gp; 77deae26bfSKyle McMartin unsigned long fault_space; 78deae26bfSKyle McMartin unsigned long fault_addr; 79deae26bfSKyle McMartin }; 80deae26bfSKyle McMartin 8106bff6b9SHelge Deller /* 8206bff6b9SHelge Deller * load_sr2() preloads the space register %%sr2 - based on the value of 8306bff6b9SHelge Deller * get_fs() - with either a value of 0 to access kernel space (KERNEL_DS which 8406bff6b9SHelge Deller * is 0), or with the current value of %%sr3 to access user space (USER_DS) 8506bff6b9SHelge Deller * memory. The following __get_user_asm() and __put_user_asm() functions have 8606bff6b9SHelge Deller * %%sr2 hard-coded to access the requested memory. 8706bff6b9SHelge Deller */ 8806bff6b9SHelge Deller #define load_sr2() \ 8906bff6b9SHelge Deller __asm__(" or,= %0,%%r0,%%r0\n\t" \ 9006bff6b9SHelge Deller " mfsp %%sr3,%0\n\t" \ 9106bff6b9SHelge Deller " mtsp %0,%%sr2\n\t" \ 9206bff6b9SHelge Deller : : "r"(get_fs()) : ) 9306bff6b9SHelge Deller 94deae26bfSKyle McMartin #define __get_user(x, ptr) \ 95deae26bfSKyle McMartin ({ \ 96deae26bfSKyle McMartin register long __gu_err __asm__ ("r8") = 0; \ 97deae26bfSKyle McMartin register long __gu_val __asm__ ("r9") = 0; \ 98deae26bfSKyle McMartin \ 9906bff6b9SHelge Deller load_sr2(); \ 100deae26bfSKyle McMartin switch (sizeof(*(ptr))) { \ 101deae26bfSKyle McMartin case 1: __get_user_asm("ldb", ptr); break; \ 102deae26bfSKyle McMartin case 2: __get_user_asm("ldh", ptr); break; \ 103deae26bfSKyle McMartin case 4: __get_user_asm("ldw", ptr); break; \ 104deae26bfSKyle McMartin case 8: LDD_USER(ptr); break; \ 1058dd95c68SHelge Deller default: BUILD_BUG(); break; \ 106deae26bfSKyle McMartin } \ 107deae26bfSKyle McMartin \ 108876b2a00SMichael S. Tsirkin (x) = (__force __typeof__(*(ptr))) __gu_val; \ 109deae26bfSKyle McMartin __gu_err; \ 110deae26bfSKyle McMartin }) 111deae26bfSKyle McMartin 112deae26bfSKyle McMartin #define __get_user_asm(ldx, ptr) \ 11306bff6b9SHelge Deller __asm__("\n1:\t" ldx "\t0(%%sr2,%2),%0\n\t" \ 114deae26bfSKyle McMartin ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_get_user_skip_1)\ 115deae26bfSKyle McMartin : "=r"(__gu_val), "=r"(__gu_err) \ 116deae26bfSKyle McMartin : "r"(ptr), "1"(__gu_err) \ 117deae26bfSKyle McMartin : "r1"); 118deae26bfSKyle McMartin 119d2ad824fSHelge Deller #if !defined(CONFIG_64BIT) 120d2ad824fSHelge Deller 121d2ad824fSHelge Deller #define __get_user_asm64(ptr) \ 122d2ad824fSHelge Deller __asm__("\n1:\tldw 0(%%sr2,%2),%0" \ 123d2ad824fSHelge Deller "\n2:\tldw 4(%%sr2,%2),%R0\n\t" \ 124d2ad824fSHelge Deller ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_get_user_skip_2)\ 125d2ad824fSHelge Deller ASM_EXCEPTIONTABLE_ENTRY(2b, fixup_get_user_skip_1)\ 126d2ad824fSHelge Deller : "=r"(__gu_val), "=r"(__gu_err) \ 127d2ad824fSHelge Deller : "r"(ptr), "1"(__gu_err) \ 128d2ad824fSHelge Deller : "r1"); 129d2ad824fSHelge Deller 130d2ad824fSHelge Deller #endif /* !defined(CONFIG_64BIT) */ 131d2ad824fSHelge Deller 132d2ad824fSHelge Deller 133deae26bfSKyle McMartin #define __put_user(x, ptr) \ 134deae26bfSKyle McMartin ({ \ 135deae26bfSKyle McMartin register long __pu_err __asm__ ("r8") = 0; \ 136deae26bfSKyle McMartin __typeof__(*(ptr)) __x = (__typeof__(*(ptr)))(x); \ 137deae26bfSKyle McMartin \ 13806bff6b9SHelge Deller load_sr2(); \ 139deae26bfSKyle McMartin switch (sizeof(*(ptr))) { \ 140deae26bfSKyle McMartin case 1: __put_user_asm("stb", __x, ptr); break; \ 141deae26bfSKyle McMartin case 2: __put_user_asm("sth", __x, ptr); break; \ 142deae26bfSKyle McMartin case 4: __put_user_asm("stw", __x, ptr); break; \ 143deae26bfSKyle McMartin case 8: STD_USER(__x, ptr); break; \ 1448dd95c68SHelge Deller default: BUILD_BUG(); break; \ 145deae26bfSKyle McMartin } \ 146deae26bfSKyle McMartin \ 147deae26bfSKyle McMartin __pu_err; \ 148deae26bfSKyle McMartin }) 149deae26bfSKyle McMartin 150deae26bfSKyle McMartin /* 151deae26bfSKyle McMartin * The "__put_user/kernel_asm()" macros tell gcc they read from memory 152deae26bfSKyle McMartin * instead of writing. This is because they do not write to any memory 153deae26bfSKyle McMartin * gcc knows about, so there are no aliasing issues. These macros must 154deae26bfSKyle McMartin * also be aware that "fixup_put_user_skip_[12]" are executed in the 155deae26bfSKyle McMartin * context of the fault, and any registers used there must be listed 156deae26bfSKyle McMartin * as clobbers. In this case only "r1" is used by the current routines. 157deae26bfSKyle McMartin * r8/r9 are already listed as err/val. 158deae26bfSKyle McMartin */ 159deae26bfSKyle McMartin 160deae26bfSKyle McMartin #define __put_user_asm(stx, x, ptr) \ 161deae26bfSKyle McMartin __asm__ __volatile__ ( \ 16206bff6b9SHelge Deller "\n1:\t" stx "\t%2,0(%%sr2,%1)\n\t" \ 163deae26bfSKyle McMartin ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_put_user_skip_1)\ 164deae26bfSKyle McMartin : "=r"(__pu_err) \ 165deae26bfSKyle McMartin : "r"(ptr), "r"(x), "0"(__pu_err) \ 166deae26bfSKyle McMartin : "r1") 167deae26bfSKyle McMartin 168deae26bfSKyle McMartin 169deae26bfSKyle McMartin #if !defined(CONFIG_64BIT) 170deae26bfSKyle McMartin 171deae26bfSKyle McMartin #define __put_user_asm64(__val, ptr) do { \ 172deae26bfSKyle McMartin __asm__ __volatile__ ( \ 17306bff6b9SHelge Deller "\n1:\tstw %2,0(%%sr2,%1)" \ 17406bff6b9SHelge Deller "\n2:\tstw %R2,4(%%sr2,%1)\n\t" \ 175deae26bfSKyle McMartin ASM_EXCEPTIONTABLE_ENTRY(1b, fixup_put_user_skip_2)\ 176deae26bfSKyle McMartin ASM_EXCEPTIONTABLE_ENTRY(2b, fixup_put_user_skip_1)\ 177deae26bfSKyle McMartin : "=r"(__pu_err) \ 1780f28b628SWill Deacon : "r"(ptr), "r"(__val), "0"(__pu_err) \ 179deae26bfSKyle McMartin : "r1"); \ 180deae26bfSKyle McMartin } while (0) 181deae26bfSKyle McMartin 182deae26bfSKyle McMartin #endif /* !defined(CONFIG_64BIT) */ 183deae26bfSKyle McMartin 184deae26bfSKyle McMartin 185deae26bfSKyle McMartin /* 186deae26bfSKyle McMartin * Complex access routines -- external declarations 187deae26bfSKyle McMartin */ 188deae26bfSKyle McMartin 189deae26bfSKyle McMartin extern unsigned long lcopy_to_user(void __user *, const void *, unsigned long); 190deae26bfSKyle McMartin extern unsigned long lcopy_from_user(void *, const void __user *, unsigned long); 191deae26bfSKyle McMartin extern unsigned long lcopy_in_user(void __user *, const void __user *, unsigned long); 192b1195c0eSJames Bottomley extern long strncpy_from_user(char *, const char __user *, long); 193deae26bfSKyle McMartin extern unsigned lclear_user(void __user *, unsigned long); 194deae26bfSKyle McMartin extern long lstrnlen_user(const char __user *, long); 195deae26bfSKyle McMartin /* 196deae26bfSKyle McMartin * Complex access routines -- macros 197deae26bfSKyle McMartin */ 198a0ffa8f0SHelge Deller #define user_addr_max() (~0UL) 199deae26bfSKyle McMartin 200deae26bfSKyle McMartin #define strnlen_user lstrnlen_user 201deae26bfSKyle McMartin #define strlen_user(str) lstrnlen_user(str, 0x7fffffffL) 202deae26bfSKyle McMartin #define clear_user lclear_user 203deae26bfSKyle McMartin #define __clear_user lclear_user 204deae26bfSKyle McMartin 205*9e91db6bSHelge Deller unsigned long __must_check __copy_to_user(void __user *dst, const void *src, 206*9e91db6bSHelge Deller unsigned long len); 207*9e91db6bSHelge Deller unsigned long __must_check __copy_from_user(void *dst, const void __user *src, 208*9e91db6bSHelge Deller unsigned long len); 209*9e91db6bSHelge Deller unsigned long copy_in_user(void __user *dst, const void __user *src, 210*9e91db6bSHelge Deller unsigned long len); 211deae26bfSKyle McMartin #define __copy_in_user copy_in_user 212deae26bfSKyle McMartin #define __copy_to_user_inatomic __copy_to_user 213deae26bfSKyle McMartin #define __copy_from_user_inatomic __copy_from_user 214deae26bfSKyle McMartin 2150d025d27SJosh Poimboeuf extern void __compiletime_error("usercopy buffer size is too small") 2160d025d27SJosh Poimboeuf __bad_copy_user(void); 2170d025d27SJosh Poimboeuf 2180d025d27SJosh Poimboeuf static inline void copy_user_overflow(int size, unsigned long count) 2190d025d27SJosh Poimboeuf { 2200d025d27SJosh Poimboeuf WARN(1, "Buffer overflow detected (%d < %lu)!\n", size, count); 2210d025d27SJosh Poimboeuf } 222888c31fcSHelge Deller 223*9e91db6bSHelge Deller static __always_inline unsigned long __must_check 224*9e91db6bSHelge Deller copy_from_user(void *to, const void __user *from, unsigned long n) 225888c31fcSHelge Deller { 226888c31fcSHelge Deller int sz = __compiletime_object_size(to); 227aace880fSAl Viro unsigned long ret = n; 228888c31fcSHelge Deller 229*9e91db6bSHelge Deller if (likely(sz < 0 || sz >= n)) { 230*9e91db6bSHelge Deller check_object_size(to, n, false); 231888c31fcSHelge Deller ret = __copy_from_user(to, from, n); 232*9e91db6bSHelge Deller } else if (!__builtin_constant_p(n)) 2330d025d27SJosh Poimboeuf copy_user_overflow(sz, n); 234888c31fcSHelge Deller else 2350d025d27SJosh Poimboeuf __bad_copy_user(); 236888c31fcSHelge Deller 237aace880fSAl Viro if (unlikely(ret)) 238aace880fSAl Viro memset(to + (n - ret), 0, ret); 239*9e91db6bSHelge Deller 240888c31fcSHelge Deller return ret; 241888c31fcSHelge Deller } 242888c31fcSHelge Deller 243*9e91db6bSHelge Deller static __always_inline unsigned long __must_check 244*9e91db6bSHelge Deller copy_to_user(void __user *to, const void *from, unsigned long n) 245*9e91db6bSHelge Deller { 246*9e91db6bSHelge Deller int sz = __compiletime_object_size(from); 247*9e91db6bSHelge Deller 248*9e91db6bSHelge Deller if (likely(sz < 0 || sz >= n)) { 249*9e91db6bSHelge Deller check_object_size(from, n, true); 250*9e91db6bSHelge Deller n = __copy_to_user(to, from, n); 251*9e91db6bSHelge Deller } else if (!__builtin_constant_p(n)) 252*9e91db6bSHelge Deller copy_user_overflow(sz, n); 253*9e91db6bSHelge Deller else 254*9e91db6bSHelge Deller __bad_copy_user(); 255*9e91db6bSHelge Deller 256*9e91db6bSHelge Deller return n; 257*9e91db6bSHelge Deller } 258*9e91db6bSHelge Deller 259e448372cSHelge Deller struct pt_regs; 260c61c25ebSKyle McMartin int fixup_exception(struct pt_regs *regs); 261c61c25ebSKyle McMartin 262deae26bfSKyle McMartin #endif /* __PARISC_UACCESS_H */ 263