xref: /openbmc/linux/arch/mips/include/asm/uaccess.h (revision 7587eb18)
1 /*
2  * This file is subject to the terms and conditions of the GNU General Public
3  * License.  See the file "COPYING" in the main directory of this archive
4  * for more details.
5  *
6  * Copyright (C) 1996, 1997, 1998, 1999, 2000, 03, 04 by Ralf Baechle
7  * Copyright (C) 1999, 2000 Silicon Graphics, Inc.
8  * Copyright (C) 2007  Maciej W. Rozycki
9  * Copyright (C) 2014, Imagination Technologies Ltd.
10  */
11 #ifndef _ASM_UACCESS_H
12 #define _ASM_UACCESS_H
13 
14 #include <linux/kernel.h>
15 #include <linux/errno.h>
16 #include <linux/thread_info.h>
17 #include <asm/asm-eva.h>
18 
19 /*
20  * The fs value determines whether argument validity checking should be
21  * performed or not.  If get_fs() == USER_DS, checking is performed, with
22  * get_fs() == KERNEL_DS, checking is bypassed.
23  *
24  * For historical reasons, these macros are grossly misnamed.
25  */
26 #ifdef CONFIG_32BIT
27 
28 #ifdef CONFIG_KVM_GUEST
29 #define __UA_LIMIT 0x40000000UL
30 #else
31 #define __UA_LIMIT 0x80000000UL
32 #endif
33 
34 #define __UA_ADDR	".word"
35 #define __UA_LA		"la"
36 #define __UA_ADDU	"addu"
37 #define __UA_t0		"$8"
38 #define __UA_t1		"$9"
39 
40 #endif /* CONFIG_32BIT */
41 
42 #ifdef CONFIG_64BIT
43 
44 extern u64 __ua_limit;
45 
46 #define __UA_LIMIT	__ua_limit
47 
48 #define __UA_ADDR	".dword"
49 #define __UA_LA		"dla"
50 #define __UA_ADDU	"daddu"
51 #define __UA_t0		"$12"
52 #define __UA_t1		"$13"
53 
54 #endif /* CONFIG_64BIT */
55 
56 /*
57  * USER_DS is a bitmask that has the bits set that may not be set in a valid
58  * userspace address.  Note that we limit 32-bit userspace to 0x7fff8000 but
59  * the arithmetic we're doing only works if the limit is a power of two, so
60  * we use 0x80000000 here on 32-bit kernels.  If a process passes an invalid
61  * address in this range it's the process's problem, not ours :-)
62  */
63 
64 #ifdef CONFIG_KVM_GUEST
65 #define KERNEL_DS	((mm_segment_t) { 0x80000000UL })
66 #define USER_DS		((mm_segment_t) { 0xC0000000UL })
67 #else
68 #define KERNEL_DS	((mm_segment_t) { 0UL })
69 #define USER_DS		((mm_segment_t) { __UA_LIMIT })
70 #endif
71 
72 #define VERIFY_READ    0
73 #define VERIFY_WRITE   1
74 
75 #define get_ds()	(KERNEL_DS)
76 #define get_fs()	(current_thread_info()->addr_limit)
77 #define set_fs(x)	(current_thread_info()->addr_limit = (x))
78 
79 #define segment_eq(a, b)	((a).seg == (b).seg)
80 
81 /*
82  * eva_kernel_access() - determine whether kernel memory access on an EVA system
83  *
84  * Determines whether memory accesses should be performed to kernel memory
85  * on a system using Extended Virtual Addressing (EVA).
86  *
87  * Return: true if a kernel memory access on an EVA system, else false.
88  */
89 static inline bool eva_kernel_access(void)
90 {
91 	if (!config_enabled(CONFIG_EVA))
92 		return false;
93 
94 	return segment_eq(get_fs(), get_ds());
95 }
96 
97 /*
98  * Is a address valid? This does a straightforward calculation rather
99  * than tests.
100  *
101  * Address valid if:
102  *  - "addr" doesn't have any high-bits set
103  *  - AND "size" doesn't have any high-bits set
104  *  - AND "addr+size" doesn't have any high-bits set
105  *  - OR we are in kernel mode.
106  *
107  * __ua_size() is a trick to avoid runtime checking of positive constant
108  * sizes; for those we already know at compile time that the size is ok.
109  */
110 #define __ua_size(size)							\
111 	((__builtin_constant_p(size) && (signed long) (size) > 0) ? 0 : (size))
112 
113 /*
114  * access_ok: - Checks if a user space pointer is valid
115  * @type: Type of access: %VERIFY_READ or %VERIFY_WRITE.  Note that
116  *	  %VERIFY_WRITE is a superset of %VERIFY_READ - if it is safe
117  *	  to write to a block, it is always safe to read from it.
118  * @addr: User space pointer to start of block to check
119  * @size: Size of block to check
120  *
121  * Context: User context only. This function may sleep if pagefaults are
122  *          enabled.
123  *
124  * Checks if a pointer to a block of memory in user space is valid.
125  *
126  * Returns true (nonzero) if the memory block may be valid, false (zero)
127  * if it is definitely invalid.
128  *
129  * Note that, depending on architecture, this function probably just
130  * checks that the pointer is in the user space range - after calling
131  * this function, memory access functions may still return -EFAULT.
132  */
133 
134 #define __access_mask get_fs().seg
135 
136 #define __access_ok(addr, size, mask)					\
137 ({									\
138 	unsigned long __addr = (unsigned long) (addr);			\
139 	unsigned long __size = size;					\
140 	unsigned long __mask = mask;					\
141 	unsigned long __ok;						\
142 									\
143 	__chk_user_ptr(addr);						\
144 	__ok = (signed long)(__mask & (__addr | (__addr + __size) |	\
145 		__ua_size(__size)));					\
146 	__ok == 0;							\
147 })
148 
149 #define access_ok(type, addr, size)					\
150 	likely(__access_ok((addr), (size), __access_mask))
151 
152 /*
153  * put_user: - Write a simple value into user space.
154  * @x:	 Value to copy to user space.
155  * @ptr: Destination address, in user space.
156  *
157  * Context: User context only. This function may sleep if pagefaults are
158  *          enabled.
159  *
160  * This macro copies a single simple value from kernel space to user
161  * space.  It supports simple types like char and int, but not larger
162  * data types like structures or arrays.
163  *
164  * @ptr must have pointer-to-simple-variable type, and @x must be assignable
165  * to the result of dereferencing @ptr.
166  *
167  * Returns zero on success, or -EFAULT on error.
168  */
169 #define put_user(x,ptr) \
170 	__put_user_check((x), (ptr), sizeof(*(ptr)))
171 
172 /*
173  * get_user: - Get a simple variable from user space.
174  * @x:	 Variable to store result.
175  * @ptr: Source address, in user space.
176  *
177  * Context: User context only. This function may sleep if pagefaults are
178  *          enabled.
179  *
180  * This macro copies a single simple variable from user space to kernel
181  * space.  It supports simple types like char and int, but not larger
182  * data types like structures or arrays.
183  *
184  * @ptr must have pointer-to-simple-variable type, and the result of
185  * dereferencing @ptr must be assignable to @x without a cast.
186  *
187  * Returns zero on success, or -EFAULT on error.
188  * On error, the variable @x is set to zero.
189  */
190 #define get_user(x,ptr) \
191 	__get_user_check((x), (ptr), sizeof(*(ptr)))
192 
193 /*
194  * __put_user: - Write a simple value into user space, with less checking.
195  * @x:	 Value to copy to user space.
196  * @ptr: Destination address, in user space.
197  *
198  * Context: User context only. This function may sleep if pagefaults are
199  *          enabled.
200  *
201  * This macro copies a single simple value from kernel space to user
202  * space.  It supports simple types like char and int, but not larger
203  * data types like structures or arrays.
204  *
205  * @ptr must have pointer-to-simple-variable type, and @x must be assignable
206  * to the result of dereferencing @ptr.
207  *
208  * Caller must check the pointer with access_ok() before calling this
209  * function.
210  *
211  * Returns zero on success, or -EFAULT on error.
212  */
213 #define __put_user(x,ptr) \
214 	__put_user_nocheck((x), (ptr), sizeof(*(ptr)))
215 
216 /*
217  * __get_user: - Get a simple variable from user space, with less checking.
218  * @x:	 Variable to store result.
219  * @ptr: Source address, in user space.
220  *
221  * Context: User context only. This function may sleep if pagefaults are
222  *          enabled.
223  *
224  * This macro copies a single simple variable from user space to kernel
225  * space.  It supports simple types like char and int, but not larger
226  * data types like structures or arrays.
227  *
228  * @ptr must have pointer-to-simple-variable type, and the result of
229  * dereferencing @ptr must be assignable to @x without a cast.
230  *
231  * Caller must check the pointer with access_ok() before calling this
232  * function.
233  *
234  * Returns zero on success, or -EFAULT on error.
235  * On error, the variable @x is set to zero.
236  */
237 #define __get_user(x,ptr) \
238 	__get_user_nocheck((x), (ptr), sizeof(*(ptr)))
239 
240 struct __large_struct { unsigned long buf[100]; };
241 #define __m(x) (*(struct __large_struct __user *)(x))
242 
243 /*
244  * Yuck.  We need two variants, one for 64bit operation and one
245  * for 32 bit mode and old iron.
246  */
247 #ifndef CONFIG_EVA
248 #define __get_kernel_common(val, size, ptr) __get_user_common(val, size, ptr)
249 #else
250 /*
251  * Kernel specific functions for EVA. We need to use normal load instructions
252  * to read data from kernel when operating in EVA mode. We use these macros to
253  * avoid redefining __get_user_asm for EVA.
254  */
255 #undef _loadd
256 #undef _loadw
257 #undef _loadh
258 #undef _loadb
259 #ifdef CONFIG_32BIT
260 #define _loadd			_loadw
261 #else
262 #define _loadd(reg, addr)	"ld " reg ", " addr
263 #endif
264 #define _loadw(reg, addr)	"lw " reg ", " addr
265 #define _loadh(reg, addr)	"lh " reg ", " addr
266 #define _loadb(reg, addr)	"lb " reg ", " addr
267 
268 #define __get_kernel_common(val, size, ptr)				\
269 do {									\
270 	switch (size) {							\
271 	case 1: __get_data_asm(val, _loadb, ptr); break;		\
272 	case 2: __get_data_asm(val, _loadh, ptr); break;		\
273 	case 4: __get_data_asm(val, _loadw, ptr); break;		\
274 	case 8: __GET_DW(val, _loadd, ptr); break;			\
275 	default: __get_user_unknown(); break;				\
276 	}								\
277 } while (0)
278 #endif
279 
280 #ifdef CONFIG_32BIT
281 #define __GET_DW(val, insn, ptr) __get_data_asm_ll32(val, insn, ptr)
282 #endif
283 #ifdef CONFIG_64BIT
284 #define __GET_DW(val, insn, ptr) __get_data_asm(val, insn, ptr)
285 #endif
286 
287 extern void __get_user_unknown(void);
288 
289 #define __get_user_common(val, size, ptr)				\
290 do {									\
291 	switch (size) {							\
292 	case 1: __get_data_asm(val, user_lb, ptr); break;		\
293 	case 2: __get_data_asm(val, user_lh, ptr); break;		\
294 	case 4: __get_data_asm(val, user_lw, ptr); break;		\
295 	case 8: __GET_DW(val, user_ld, ptr); break;			\
296 	default: __get_user_unknown(); break;				\
297 	}								\
298 } while (0)
299 
300 #define __get_user_nocheck(x, ptr, size)				\
301 ({									\
302 	int __gu_err;							\
303 									\
304 	if (eva_kernel_access()) {					\
305 		__get_kernel_common((x), size, ptr);			\
306 	} else {							\
307 		__chk_user_ptr(ptr);					\
308 		__get_user_common((x), size, ptr);			\
309 	}								\
310 	__gu_err;							\
311 })
312 
313 #define __get_user_check(x, ptr, size)					\
314 ({									\
315 	int __gu_err = -EFAULT;						\
316 	const __typeof__(*(ptr)) __user * __gu_ptr = (ptr);		\
317 									\
318 	might_fault();							\
319 	if (likely(access_ok(VERIFY_READ,  __gu_ptr, size))) {		\
320 		if (eva_kernel_access())				\
321 			__get_kernel_common((x), size, __gu_ptr);	\
322 		else							\
323 			__get_user_common((x), size, __gu_ptr);		\
324 	} else								\
325 		(x) = 0;						\
326 									\
327 	__gu_err;							\
328 })
329 
330 #define __get_data_asm(val, insn, addr)					\
331 {									\
332 	long __gu_tmp;							\
333 									\
334 	__asm__ __volatile__(						\
335 	"1:	"insn("%1", "%3")"				\n"	\
336 	"2:							\n"	\
337 	"	.insn						\n"	\
338 	"	.section .fixup,\"ax\"				\n"	\
339 	"3:	li	%0, %4					\n"	\
340 	"	move	%1, $0					\n"	\
341 	"	j	2b					\n"	\
342 	"	.previous					\n"	\
343 	"	.section __ex_table,\"a\"			\n"	\
344 	"	"__UA_ADDR "\t1b, 3b				\n"	\
345 	"	.previous					\n"	\
346 	: "=r" (__gu_err), "=r" (__gu_tmp)				\
347 	: "0" (0), "o" (__m(addr)), "i" (-EFAULT));			\
348 									\
349 	(val) = (__typeof__(*(addr))) __gu_tmp;				\
350 }
351 
352 /*
353  * Get a long long 64 using 32 bit registers.
354  */
355 #define __get_data_asm_ll32(val, insn, addr)				\
356 {									\
357 	union {								\
358 		unsigned long long	l;				\
359 		__typeof__(*(addr))	t;				\
360 	} __gu_tmp;							\
361 									\
362 	__asm__ __volatile__(						\
363 	"1:	" insn("%1", "(%3)")"				\n"	\
364 	"2:	" insn("%D1", "4(%3)")"				\n"	\
365 	"3:							\n"	\
366 	"	.insn						\n"	\
367 	"	.section	.fixup,\"ax\"			\n"	\
368 	"4:	li	%0, %4					\n"	\
369 	"	move	%1, $0					\n"	\
370 	"	move	%D1, $0					\n"	\
371 	"	j	3b					\n"	\
372 	"	.previous					\n"	\
373 	"	.section	__ex_table,\"a\"		\n"	\
374 	"	" __UA_ADDR "	1b, 4b				\n"	\
375 	"	" __UA_ADDR "	2b, 4b				\n"	\
376 	"	.previous					\n"	\
377 	: "=r" (__gu_err), "=&r" (__gu_tmp.l)				\
378 	: "0" (0), "r" (addr), "i" (-EFAULT));				\
379 									\
380 	(val) = __gu_tmp.t;						\
381 }
382 
383 #ifndef CONFIG_EVA
384 #define __put_kernel_common(ptr, size) __put_user_common(ptr, size)
385 #else
386 /*
387  * Kernel specific functions for EVA. We need to use normal load instructions
388  * to read data from kernel when operating in EVA mode. We use these macros to
389  * avoid redefining __get_data_asm for EVA.
390  */
391 #undef _stored
392 #undef _storew
393 #undef _storeh
394 #undef _storeb
395 #ifdef CONFIG_32BIT
396 #define _stored			_storew
397 #else
398 #define _stored(reg, addr)	"ld " reg ", " addr
399 #endif
400 
401 #define _storew(reg, addr)	"sw " reg ", " addr
402 #define _storeh(reg, addr)	"sh " reg ", " addr
403 #define _storeb(reg, addr)	"sb " reg ", " addr
404 
405 #define __put_kernel_common(ptr, size)					\
406 do {									\
407 	switch (size) {							\
408 	case 1: __put_data_asm(_storeb, ptr); break;			\
409 	case 2: __put_data_asm(_storeh, ptr); break;			\
410 	case 4: __put_data_asm(_storew, ptr); break;			\
411 	case 8: __PUT_DW(_stored, ptr); break;				\
412 	default: __put_user_unknown(); break;				\
413 	}								\
414 } while(0)
415 #endif
416 
417 /*
418  * Yuck.  We need two variants, one for 64bit operation and one
419  * for 32 bit mode and old iron.
420  */
421 #ifdef CONFIG_32BIT
422 #define __PUT_DW(insn, ptr) __put_data_asm_ll32(insn, ptr)
423 #endif
424 #ifdef CONFIG_64BIT
425 #define __PUT_DW(insn, ptr) __put_data_asm(insn, ptr)
426 #endif
427 
428 #define __put_user_common(ptr, size)					\
429 do {									\
430 	switch (size) {							\
431 	case 1: __put_data_asm(user_sb, ptr); break;			\
432 	case 2: __put_data_asm(user_sh, ptr); break;			\
433 	case 4: __put_data_asm(user_sw, ptr); break;			\
434 	case 8: __PUT_DW(user_sd, ptr); break;				\
435 	default: __put_user_unknown(); break;				\
436 	}								\
437 } while (0)
438 
439 #define __put_user_nocheck(x, ptr, size)				\
440 ({									\
441 	__typeof__(*(ptr)) __pu_val;					\
442 	int __pu_err = 0;						\
443 									\
444 	__pu_val = (x);							\
445 	if (eva_kernel_access()) {					\
446 		__put_kernel_common(ptr, size);				\
447 	} else {							\
448 		__chk_user_ptr(ptr);					\
449 		__put_user_common(ptr, size);				\
450 	}								\
451 	__pu_err;							\
452 })
453 
454 #define __put_user_check(x, ptr, size)					\
455 ({									\
456 	__typeof__(*(ptr)) __user *__pu_addr = (ptr);			\
457 	__typeof__(*(ptr)) __pu_val = (x);				\
458 	int __pu_err = -EFAULT;						\
459 									\
460 	might_fault();							\
461 	if (likely(access_ok(VERIFY_WRITE,  __pu_addr, size))) {	\
462 		if (eva_kernel_access())				\
463 			__put_kernel_common(__pu_addr, size);		\
464 		else							\
465 			__put_user_common(__pu_addr, size);		\
466 	}								\
467 									\
468 	__pu_err;							\
469 })
470 
471 #define __put_data_asm(insn, ptr)					\
472 {									\
473 	__asm__ __volatile__(						\
474 	"1:	"insn("%z2", "%3")"	# __put_data_asm	\n"	\
475 	"2:							\n"	\
476 	"	.insn						\n"	\
477 	"	.section	.fixup,\"ax\"			\n"	\
478 	"3:	li	%0, %4					\n"	\
479 	"	j	2b					\n"	\
480 	"	.previous					\n"	\
481 	"	.section	__ex_table,\"a\"		\n"	\
482 	"	" __UA_ADDR "	1b, 3b				\n"	\
483 	"	.previous					\n"	\
484 	: "=r" (__pu_err)						\
485 	: "0" (0), "Jr" (__pu_val), "o" (__m(ptr)),			\
486 	  "i" (-EFAULT));						\
487 }
488 
489 #define __put_data_asm_ll32(insn, ptr)					\
490 {									\
491 	__asm__ __volatile__(						\
492 	"1:	"insn("%2", "(%3)")"	# __put_data_asm_ll32	\n"	\
493 	"2:	"insn("%D2", "4(%3)")"				\n"	\
494 	"3:							\n"	\
495 	"	.insn						\n"	\
496 	"	.section	.fixup,\"ax\"			\n"	\
497 	"4:	li	%0, %4					\n"	\
498 	"	j	3b					\n"	\
499 	"	.previous					\n"	\
500 	"	.section	__ex_table,\"a\"		\n"	\
501 	"	" __UA_ADDR "	1b, 4b				\n"	\
502 	"	" __UA_ADDR "	2b, 4b				\n"	\
503 	"	.previous"						\
504 	: "=r" (__pu_err)						\
505 	: "0" (0), "r" (__pu_val), "r" (ptr),				\
506 	  "i" (-EFAULT));						\
507 }
508 
509 extern void __put_user_unknown(void);
510 
511 /*
512  * ul{b,h,w} are macros and there are no equivalent macros for EVA.
513  * EVA unaligned access is handled in the ADE exception handler.
514  */
515 #ifndef CONFIG_EVA
516 /*
517  * put_user_unaligned: - Write a simple value into user space.
518  * @x:	 Value to copy to user space.
519  * @ptr: Destination address, in user space.
520  *
521  * Context: User context only. This function may sleep if pagefaults are
522  *          enabled.
523  *
524  * This macro copies a single simple value from kernel space to user
525  * space.  It supports simple types like char and int, but not larger
526  * data types like structures or arrays.
527  *
528  * @ptr must have pointer-to-simple-variable type, and @x must be assignable
529  * to the result of dereferencing @ptr.
530  *
531  * Returns zero on success, or -EFAULT on error.
532  */
533 #define put_user_unaligned(x,ptr)	\
534 	__put_user_unaligned_check((x),(ptr),sizeof(*(ptr)))
535 
536 /*
537  * get_user_unaligned: - Get a simple variable from user space.
538  * @x:	 Variable to store result.
539  * @ptr: Source address, in user space.
540  *
541  * Context: User context only. This function may sleep if pagefaults are
542  *          enabled.
543  *
544  * This macro copies a single simple variable from user space to kernel
545  * space.  It supports simple types like char and int, but not larger
546  * data types like structures or arrays.
547  *
548  * @ptr must have pointer-to-simple-variable type, and the result of
549  * dereferencing @ptr must be assignable to @x without a cast.
550  *
551  * Returns zero on success, or -EFAULT on error.
552  * On error, the variable @x is set to zero.
553  */
554 #define get_user_unaligned(x,ptr) \
555 	__get_user_unaligned_check((x),(ptr),sizeof(*(ptr)))
556 
557 /*
558  * __put_user_unaligned: - Write a simple value into user space, with less checking.
559  * @x:	 Value to copy to user space.
560  * @ptr: Destination address, in user space.
561  *
562  * Context: User context only. This function may sleep if pagefaults are
563  *          enabled.
564  *
565  * This macro copies a single simple value from kernel space to user
566  * space.  It supports simple types like char and int, but not larger
567  * data types like structures or arrays.
568  *
569  * @ptr must have pointer-to-simple-variable type, and @x must be assignable
570  * to the result of dereferencing @ptr.
571  *
572  * Caller must check the pointer with access_ok() before calling this
573  * function.
574  *
575  * Returns zero on success, or -EFAULT on error.
576  */
577 #define __put_user_unaligned(x,ptr) \
578 	__put_user_unaligned_nocheck((x),(ptr),sizeof(*(ptr)))
579 
580 /*
581  * __get_user_unaligned: - Get a simple variable from user space, with less checking.
582  * @x:	 Variable to store result.
583  * @ptr: Source address, in user space.
584  *
585  * Context: User context only. This function may sleep if pagefaults are
586  *          enabled.
587  *
588  * This macro copies a single simple variable from user space to kernel
589  * space.  It supports simple types like char and int, but not larger
590  * data types like structures or arrays.
591  *
592  * @ptr must have pointer-to-simple-variable type, and the result of
593  * dereferencing @ptr must be assignable to @x without a cast.
594  *
595  * Caller must check the pointer with access_ok() before calling this
596  * function.
597  *
598  * Returns zero on success, or -EFAULT on error.
599  * On error, the variable @x is set to zero.
600  */
601 #define __get_user_unaligned(x,ptr) \
602 	__get_user_unaligned_nocheck((x),(ptr),sizeof(*(ptr)))
603 
604 /*
605  * Yuck.  We need two variants, one for 64bit operation and one
606  * for 32 bit mode and old iron.
607  */
608 #ifdef CONFIG_32BIT
609 #define __GET_USER_UNALIGNED_DW(val, ptr)				\
610 	__get_user_unaligned_asm_ll32(val, ptr)
611 #endif
612 #ifdef CONFIG_64BIT
613 #define __GET_USER_UNALIGNED_DW(val, ptr)				\
614 	__get_user_unaligned_asm(val, "uld", ptr)
615 #endif
616 
617 extern void __get_user_unaligned_unknown(void);
618 
619 #define __get_user_unaligned_common(val, size, ptr)			\
620 do {									\
621 	switch (size) {							\
622 	case 1: __get_data_asm(val, "lb", ptr); break;			\
623 	case 2: __get_data_unaligned_asm(val, "ulh", ptr); break;	\
624 	case 4: __get_data_unaligned_asm(val, "ulw", ptr); break;	\
625 	case 8: __GET_USER_UNALIGNED_DW(val, ptr); break;		\
626 	default: __get_user_unaligned_unknown(); break;			\
627 	}								\
628 } while (0)
629 
630 #define __get_user_unaligned_nocheck(x,ptr,size)			\
631 ({									\
632 	int __gu_err;							\
633 									\
634 	__get_user_unaligned_common((x), size, ptr);			\
635 	__gu_err;							\
636 })
637 
638 #define __get_user_unaligned_check(x,ptr,size)				\
639 ({									\
640 	int __gu_err = -EFAULT;						\
641 	const __typeof__(*(ptr)) __user * __gu_ptr = (ptr);		\
642 									\
643 	if (likely(access_ok(VERIFY_READ,  __gu_ptr, size)))		\
644 		__get_user_unaligned_common((x), size, __gu_ptr);	\
645 									\
646 	__gu_err;							\
647 })
648 
649 #define __get_data_unaligned_asm(val, insn, addr)			\
650 {									\
651 	long __gu_tmp;							\
652 									\
653 	__asm__ __volatile__(						\
654 	"1:	" insn "	%1, %3				\n"	\
655 	"2:							\n"	\
656 	"	.insn						\n"	\
657 	"	.section .fixup,\"ax\"				\n"	\
658 	"3:	li	%0, %4					\n"	\
659 	"	move	%1, $0					\n"	\
660 	"	j	2b					\n"	\
661 	"	.previous					\n"	\
662 	"	.section __ex_table,\"a\"			\n"	\
663 	"	"__UA_ADDR "\t1b, 3b				\n"	\
664 	"	"__UA_ADDR "\t1b + 4, 3b			\n"	\
665 	"	.previous					\n"	\
666 	: "=r" (__gu_err), "=r" (__gu_tmp)				\
667 	: "0" (0), "o" (__m(addr)), "i" (-EFAULT));			\
668 									\
669 	(val) = (__typeof__(*(addr))) __gu_tmp;				\
670 }
671 
672 /*
673  * Get a long long 64 using 32 bit registers.
674  */
675 #define __get_user_unaligned_asm_ll32(val, addr)			\
676 {									\
677 	unsigned long long __gu_tmp;					\
678 									\
679 	__asm__ __volatile__(						\
680 	"1:	ulw	%1, (%3)				\n"	\
681 	"2:	ulw	%D1, 4(%3)				\n"	\
682 	"	move	%0, $0					\n"	\
683 	"3:							\n"	\
684 	"	.insn						\n"	\
685 	"	.section	.fixup,\"ax\"			\n"	\
686 	"4:	li	%0, %4					\n"	\
687 	"	move	%1, $0					\n"	\
688 	"	move	%D1, $0					\n"	\
689 	"	j	3b					\n"	\
690 	"	.previous					\n"	\
691 	"	.section	__ex_table,\"a\"		\n"	\
692 	"	" __UA_ADDR "	1b, 4b				\n"	\
693 	"	" __UA_ADDR "	1b + 4, 4b			\n"	\
694 	"	" __UA_ADDR "	2b, 4b				\n"	\
695 	"	" __UA_ADDR "	2b + 4, 4b			\n"	\
696 	"	.previous					\n"	\
697 	: "=r" (__gu_err), "=&r" (__gu_tmp)				\
698 	: "0" (0), "r" (addr), "i" (-EFAULT));				\
699 	(val) = (__typeof__(*(addr))) __gu_tmp;				\
700 }
701 
702 /*
703  * Yuck.  We need two variants, one for 64bit operation and one
704  * for 32 bit mode and old iron.
705  */
706 #ifdef CONFIG_32BIT
707 #define __PUT_USER_UNALIGNED_DW(ptr) __put_user_unaligned_asm_ll32(ptr)
708 #endif
709 #ifdef CONFIG_64BIT
710 #define __PUT_USER_UNALIGNED_DW(ptr) __put_user_unaligned_asm("usd", ptr)
711 #endif
712 
713 #define __put_user_unaligned_common(ptr, size)				\
714 do {									\
715 	switch (size) {							\
716 	case 1: __put_data_asm("sb", ptr); break;			\
717 	case 2: __put_user_unaligned_asm("ush", ptr); break;		\
718 	case 4: __put_user_unaligned_asm("usw", ptr); break;		\
719 	case 8: __PUT_USER_UNALIGNED_DW(ptr); break;			\
720 	default: __put_user_unaligned_unknown(); break;			\
721 } while (0)
722 
723 #define __put_user_unaligned_nocheck(x,ptr,size)			\
724 ({									\
725 	__typeof__(*(ptr)) __pu_val;					\
726 	int __pu_err = 0;						\
727 									\
728 	__pu_val = (x);							\
729 	__put_user_unaligned_common(ptr, size);				\
730 	__pu_err;							\
731 })
732 
733 #define __put_user_unaligned_check(x,ptr,size)				\
734 ({									\
735 	__typeof__(*(ptr)) __user *__pu_addr = (ptr);			\
736 	__typeof__(*(ptr)) __pu_val = (x);				\
737 	int __pu_err = -EFAULT;						\
738 									\
739 	if (likely(access_ok(VERIFY_WRITE,  __pu_addr, size)))		\
740 		__put_user_unaligned_common(__pu_addr, size);		\
741 									\
742 	__pu_err;							\
743 })
744 
745 #define __put_user_unaligned_asm(insn, ptr)				\
746 {									\
747 	__asm__ __volatile__(						\
748 	"1:	" insn "	%z2, %3		# __put_user_unaligned_asm\n" \
749 	"2:							\n"	\
750 	"	.insn						\n"	\
751 	"	.section	.fixup,\"ax\"			\n"	\
752 	"3:	li	%0, %4					\n"	\
753 	"	j	2b					\n"	\
754 	"	.previous					\n"	\
755 	"	.section	__ex_table,\"a\"		\n"	\
756 	"	" __UA_ADDR "	1b, 3b				\n"	\
757 	"	.previous					\n"	\
758 	: "=r" (__pu_err)						\
759 	: "0" (0), "Jr" (__pu_val), "o" (__m(ptr)),			\
760 	  "i" (-EFAULT));						\
761 }
762 
763 #define __put_user_unaligned_asm_ll32(ptr)				\
764 {									\
765 	__asm__ __volatile__(						\
766 	"1:	sw	%2, (%3)	# __put_user_unaligned_asm_ll32 \n" \
767 	"2:	sw	%D2, 4(%3)				\n"	\
768 	"3:							\n"	\
769 	"	.insn						\n"	\
770 	"	.section	.fixup,\"ax\"			\n"	\
771 	"4:	li	%0, %4					\n"	\
772 	"	j	3b					\n"	\
773 	"	.previous					\n"	\
774 	"	.section	__ex_table,\"a\"		\n"	\
775 	"	" __UA_ADDR "	1b, 4b				\n"	\
776 	"	" __UA_ADDR "	1b + 4, 4b			\n"	\
777 	"	" __UA_ADDR "	2b, 4b				\n"	\
778 	"	" __UA_ADDR "	2b + 4, 4b			\n"	\
779 	"	.previous"						\
780 	: "=r" (__pu_err)						\
781 	: "0" (0), "r" (__pu_val), "r" (ptr),				\
782 	  "i" (-EFAULT));						\
783 }
784 
785 extern void __put_user_unaligned_unknown(void);
786 #endif
787 
788 /*
789  * We're generating jump to subroutines which will be outside the range of
790  * jump instructions
791  */
792 #ifdef MODULE
793 #define __MODULE_JAL(destination)					\
794 	".set\tnoat\n\t"						\
795 	__UA_LA "\t$1, " #destination "\n\t"				\
796 	"jalr\t$1\n\t"							\
797 	".set\tat\n\t"
798 #else
799 #define __MODULE_JAL(destination)					\
800 	"jal\t" #destination "\n\t"
801 #endif
802 
803 #if defined(CONFIG_CPU_DADDI_WORKAROUNDS) || (defined(CONFIG_EVA) &&	\
804 					      defined(CONFIG_CPU_HAS_PREFETCH))
805 #define DADDI_SCRATCH "$3"
806 #else
807 #define DADDI_SCRATCH "$0"
808 #endif
809 
810 extern size_t __copy_user(void *__to, const void *__from, size_t __n);
811 
812 #ifndef CONFIG_EVA
813 #define __invoke_copy_to_user(to, from, n)				\
814 ({									\
815 	register void __user *__cu_to_r __asm__("$4");			\
816 	register const void *__cu_from_r __asm__("$5");			\
817 	register long __cu_len_r __asm__("$6");				\
818 									\
819 	__cu_to_r = (to);						\
820 	__cu_from_r = (from);						\
821 	__cu_len_r = (n);						\
822 	__asm__ __volatile__(						\
823 	__MODULE_JAL(__copy_user)					\
824 	: "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r)	\
825 	:								\
826 	: "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31",	\
827 	  DADDI_SCRATCH, "memory");					\
828 	__cu_len_r;							\
829 })
830 
831 #define __invoke_copy_to_kernel(to, from, n)				\
832 	__invoke_copy_to_user(to, from, n)
833 
834 #endif
835 
836 /*
837  * __copy_to_user: - Copy a block of data into user space, with less checking.
838  * @to:	  Destination address, in user space.
839  * @from: Source address, in kernel space.
840  * @n:	  Number of bytes to copy.
841  *
842  * Context: User context only. This function may sleep if pagefaults are
843  *          enabled.
844  *
845  * Copy data from kernel space to user space.  Caller must check
846  * the specified block with access_ok() before calling this function.
847  *
848  * Returns number of bytes that could not be copied.
849  * On success, this will be zero.
850  */
851 #define __copy_to_user(to, from, n)					\
852 ({									\
853 	void __user *__cu_to;						\
854 	const void *__cu_from;						\
855 	long __cu_len;							\
856 									\
857 	__cu_to = (to);							\
858 	__cu_from = (from);						\
859 	__cu_len = (n);							\
860 	might_fault();							\
861 	if (eva_kernel_access())					\
862 		__cu_len = __invoke_copy_to_kernel(__cu_to, __cu_from,	\
863 						   __cu_len);		\
864 	else								\
865 		__cu_len = __invoke_copy_to_user(__cu_to, __cu_from,	\
866 						 __cu_len);		\
867 	__cu_len;							\
868 })
869 
870 extern size_t __copy_user_inatomic(void *__to, const void *__from, size_t __n);
871 
872 #define __copy_to_user_inatomic(to, from, n)				\
873 ({									\
874 	void __user *__cu_to;						\
875 	const void *__cu_from;						\
876 	long __cu_len;							\
877 									\
878 	__cu_to = (to);							\
879 	__cu_from = (from);						\
880 	__cu_len = (n);							\
881 	if (eva_kernel_access())					\
882 		__cu_len = __invoke_copy_to_kernel(__cu_to, __cu_from,	\
883 						   __cu_len);		\
884 	else								\
885 		__cu_len = __invoke_copy_to_user(__cu_to, __cu_from,	\
886 						 __cu_len);		\
887 	__cu_len;							\
888 })
889 
890 #define __copy_from_user_inatomic(to, from, n)				\
891 ({									\
892 	void *__cu_to;							\
893 	const void __user *__cu_from;					\
894 	long __cu_len;							\
895 									\
896 	__cu_to = (to);							\
897 	__cu_from = (from);						\
898 	__cu_len = (n);							\
899 	if (eva_kernel_access())					\
900 		__cu_len = __invoke_copy_from_kernel_inatomic(__cu_to,	\
901 							      __cu_from,\
902 							      __cu_len);\
903 	else								\
904 		__cu_len = __invoke_copy_from_user_inatomic(__cu_to,	\
905 							    __cu_from,	\
906 							    __cu_len);	\
907 	__cu_len;							\
908 })
909 
910 /*
911  * copy_to_user: - Copy a block of data into user space.
912  * @to:	  Destination address, in user space.
913  * @from: Source address, in kernel space.
914  * @n:	  Number of bytes to copy.
915  *
916  * Context: User context only. This function may sleep if pagefaults are
917  *          enabled.
918  *
919  * Copy data from kernel space to user space.
920  *
921  * Returns number of bytes that could not be copied.
922  * On success, this will be zero.
923  */
924 #define copy_to_user(to, from, n)					\
925 ({									\
926 	void __user *__cu_to;						\
927 	const void *__cu_from;						\
928 	long __cu_len;							\
929 									\
930 	__cu_to = (to);							\
931 	__cu_from = (from);						\
932 	__cu_len = (n);							\
933 	if (eva_kernel_access()) {					\
934 		__cu_len = __invoke_copy_to_kernel(__cu_to,		\
935 						   __cu_from,		\
936 						   __cu_len);		\
937 	} else {							\
938 		if (access_ok(VERIFY_WRITE, __cu_to, __cu_len)) {       \
939 			might_fault();                                  \
940 			__cu_len = __invoke_copy_to_user(__cu_to,	\
941 							 __cu_from,	\
942 							 __cu_len);     \
943 		}							\
944 	}								\
945 	__cu_len;							\
946 })
947 
948 #ifndef CONFIG_EVA
949 
950 #define __invoke_copy_from_user(to, from, n)				\
951 ({									\
952 	register void *__cu_to_r __asm__("$4");				\
953 	register const void __user *__cu_from_r __asm__("$5");		\
954 	register long __cu_len_r __asm__("$6");				\
955 									\
956 	__cu_to_r = (to);						\
957 	__cu_from_r = (from);						\
958 	__cu_len_r = (n);						\
959 	__asm__ __volatile__(						\
960 	".set\tnoreorder\n\t"						\
961 	__MODULE_JAL(__copy_user)					\
962 	".set\tnoat\n\t"						\
963 	__UA_ADDU "\t$1, %1, %2\n\t"					\
964 	".set\tat\n\t"							\
965 	".set\treorder"							\
966 	: "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r)	\
967 	:								\
968 	: "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31",	\
969 	  DADDI_SCRATCH, "memory");					\
970 	__cu_len_r;							\
971 })
972 
973 #define __invoke_copy_from_kernel(to, from, n)				\
974 	__invoke_copy_from_user(to, from, n)
975 
976 /* For userland <-> userland operations */
977 #define ___invoke_copy_in_user(to, from, n)				\
978 	__invoke_copy_from_user(to, from, n)
979 
980 /* For kernel <-> kernel operations */
981 #define ___invoke_copy_in_kernel(to, from, n)				\
982 	__invoke_copy_from_user(to, from, n)
983 
984 #define __invoke_copy_from_user_inatomic(to, from, n)			\
985 ({									\
986 	register void *__cu_to_r __asm__("$4");				\
987 	register const void __user *__cu_from_r __asm__("$5");		\
988 	register long __cu_len_r __asm__("$6");				\
989 									\
990 	__cu_to_r = (to);						\
991 	__cu_from_r = (from);						\
992 	__cu_len_r = (n);						\
993 	__asm__ __volatile__(						\
994 	".set\tnoreorder\n\t"						\
995 	__MODULE_JAL(__copy_user_inatomic)				\
996 	".set\tnoat\n\t"						\
997 	__UA_ADDU "\t$1, %1, %2\n\t"					\
998 	".set\tat\n\t"							\
999 	".set\treorder"							\
1000 	: "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r)	\
1001 	:								\
1002 	: "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31",	\
1003 	  DADDI_SCRATCH, "memory");					\
1004 	__cu_len_r;							\
1005 })
1006 
1007 #define __invoke_copy_from_kernel_inatomic(to, from, n)			\
1008 	__invoke_copy_from_user_inatomic(to, from, n)			\
1009 
1010 #else
1011 
1012 /* EVA specific functions */
1013 
1014 extern size_t __copy_user_inatomic_eva(void *__to, const void *__from,
1015 				       size_t __n);
1016 extern size_t __copy_from_user_eva(void *__to, const void *__from,
1017 				   size_t __n);
1018 extern size_t __copy_to_user_eva(void *__to, const void *__from,
1019 				 size_t __n);
1020 extern size_t __copy_in_user_eva(void *__to, const void *__from, size_t __n);
1021 
1022 #define __invoke_copy_from_user_eva_generic(to, from, n, func_ptr)	\
1023 ({									\
1024 	register void *__cu_to_r __asm__("$4");				\
1025 	register const void __user *__cu_from_r __asm__("$5");		\
1026 	register long __cu_len_r __asm__("$6");				\
1027 									\
1028 	__cu_to_r = (to);						\
1029 	__cu_from_r = (from);						\
1030 	__cu_len_r = (n);						\
1031 	__asm__ __volatile__(						\
1032 	".set\tnoreorder\n\t"						\
1033 	__MODULE_JAL(func_ptr)						\
1034 	".set\tnoat\n\t"						\
1035 	__UA_ADDU "\t$1, %1, %2\n\t"					\
1036 	".set\tat\n\t"							\
1037 	".set\treorder"							\
1038 	: "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r)	\
1039 	:								\
1040 	: "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31",	\
1041 	  DADDI_SCRATCH, "memory");					\
1042 	__cu_len_r;							\
1043 })
1044 
1045 #define __invoke_copy_to_user_eva_generic(to, from, n, func_ptr)	\
1046 ({									\
1047 	register void *__cu_to_r __asm__("$4");				\
1048 	register const void __user *__cu_from_r __asm__("$5");		\
1049 	register long __cu_len_r __asm__("$6");				\
1050 									\
1051 	__cu_to_r = (to);						\
1052 	__cu_from_r = (from);						\
1053 	__cu_len_r = (n);						\
1054 	__asm__ __volatile__(						\
1055 	__MODULE_JAL(func_ptr)						\
1056 	: "+r" (__cu_to_r), "+r" (__cu_from_r), "+r" (__cu_len_r)	\
1057 	:								\
1058 	: "$8", "$9", "$10", "$11", "$12", "$14", "$15", "$24", "$31",	\
1059 	  DADDI_SCRATCH, "memory");					\
1060 	__cu_len_r;							\
1061 })
1062 
1063 /*
1064  * Source or destination address is in userland. We need to go through
1065  * the TLB
1066  */
1067 #define __invoke_copy_from_user(to, from, n)				\
1068 	__invoke_copy_from_user_eva_generic(to, from, n, __copy_from_user_eva)
1069 
1070 #define __invoke_copy_from_user_inatomic(to, from, n)			\
1071 	__invoke_copy_from_user_eva_generic(to, from, n,		\
1072 					    __copy_user_inatomic_eva)
1073 
1074 #define __invoke_copy_to_user(to, from, n)				\
1075 	__invoke_copy_to_user_eva_generic(to, from, n, __copy_to_user_eva)
1076 
1077 #define ___invoke_copy_in_user(to, from, n)				\
1078 	__invoke_copy_from_user_eva_generic(to, from, n, __copy_in_user_eva)
1079 
1080 /*
1081  * Source or destination address in the kernel. We are not going through
1082  * the TLB
1083  */
1084 #define __invoke_copy_from_kernel(to, from, n)				\
1085 	__invoke_copy_from_user_eva_generic(to, from, n, __copy_user)
1086 
1087 #define __invoke_copy_from_kernel_inatomic(to, from, n)			\
1088 	__invoke_copy_from_user_eva_generic(to, from, n, __copy_user_inatomic)
1089 
1090 #define __invoke_copy_to_kernel(to, from, n)				\
1091 	__invoke_copy_to_user_eva_generic(to, from, n, __copy_user)
1092 
1093 #define ___invoke_copy_in_kernel(to, from, n)				\
1094 	__invoke_copy_from_user_eva_generic(to, from, n, __copy_user)
1095 
1096 #endif /* CONFIG_EVA */
1097 
1098 /*
1099  * __copy_from_user: - Copy a block of data from user space, with less checking.
1100  * @to:	  Destination address, in kernel space.
1101  * @from: Source address, in user space.
1102  * @n:	  Number of bytes to copy.
1103  *
1104  * Context: User context only. This function may sleep if pagefaults are
1105  *          enabled.
1106  *
1107  * Copy data from user space to kernel space.  Caller must check
1108  * the specified block with access_ok() before calling this function.
1109  *
1110  * Returns number of bytes that could not be copied.
1111  * On success, this will be zero.
1112  *
1113  * If some data could not be copied, this function will pad the copied
1114  * data to the requested size using zero bytes.
1115  */
1116 #define __copy_from_user(to, from, n)					\
1117 ({									\
1118 	void *__cu_to;							\
1119 	const void __user *__cu_from;					\
1120 	long __cu_len;							\
1121 									\
1122 	__cu_to = (to);							\
1123 	__cu_from = (from);						\
1124 	__cu_len = (n);							\
1125 	if (eva_kernel_access()) {					\
1126 		__cu_len = __invoke_copy_from_kernel(__cu_to,		\
1127 						     __cu_from,		\
1128 						     __cu_len);		\
1129 	} else {							\
1130 		might_fault();						\
1131 		__cu_len = __invoke_copy_from_user(__cu_to, __cu_from,	\
1132 						   __cu_len);		\
1133 	}								\
1134 	__cu_len;							\
1135 })
1136 
1137 /*
1138  * copy_from_user: - Copy a block of data from user space.
1139  * @to:	  Destination address, in kernel space.
1140  * @from: Source address, in user space.
1141  * @n:	  Number of bytes to copy.
1142  *
1143  * Context: User context only. This function may sleep if pagefaults are
1144  *          enabled.
1145  *
1146  * Copy data from user space to kernel space.
1147  *
1148  * Returns number of bytes that could not be copied.
1149  * On success, this will be zero.
1150  *
1151  * If some data could not be copied, this function will pad the copied
1152  * data to the requested size using zero bytes.
1153  */
1154 #define copy_from_user(to, from, n)					\
1155 ({									\
1156 	void *__cu_to;							\
1157 	const void __user *__cu_from;					\
1158 	long __cu_len;							\
1159 									\
1160 	__cu_to = (to);							\
1161 	__cu_from = (from);						\
1162 	__cu_len = (n);							\
1163 	if (eva_kernel_access()) {					\
1164 		__cu_len = __invoke_copy_from_kernel(__cu_to,		\
1165 						     __cu_from,		\
1166 						     __cu_len);		\
1167 	} else {							\
1168 		if (access_ok(VERIFY_READ, __cu_from, __cu_len)) {	\
1169 			might_fault();                                  \
1170 			__cu_len = __invoke_copy_from_user(__cu_to,	\
1171 							   __cu_from,	\
1172 							   __cu_len);   \
1173 		}							\
1174 	}								\
1175 	__cu_len;							\
1176 })
1177 
1178 #define __copy_in_user(to, from, n)					\
1179 ({									\
1180 	void __user *__cu_to;						\
1181 	const void __user *__cu_from;					\
1182 	long __cu_len;							\
1183 									\
1184 	__cu_to = (to);							\
1185 	__cu_from = (from);						\
1186 	__cu_len = (n);							\
1187 	if (eva_kernel_access()) {					\
1188 		__cu_len = ___invoke_copy_in_kernel(__cu_to, __cu_from,	\
1189 						    __cu_len);		\
1190 	} else {							\
1191 		might_fault();						\
1192 		__cu_len = ___invoke_copy_in_user(__cu_to, __cu_from,	\
1193 						  __cu_len);		\
1194 	}								\
1195 	__cu_len;							\
1196 })
1197 
1198 #define copy_in_user(to, from, n)					\
1199 ({									\
1200 	void __user *__cu_to;						\
1201 	const void __user *__cu_from;					\
1202 	long __cu_len;							\
1203 									\
1204 	__cu_to = (to);							\
1205 	__cu_from = (from);						\
1206 	__cu_len = (n);							\
1207 	if (eva_kernel_access()) {					\
1208 		__cu_len = ___invoke_copy_in_kernel(__cu_to,__cu_from,	\
1209 						    __cu_len);		\
1210 	} else {							\
1211 		if (likely(access_ok(VERIFY_READ, __cu_from, __cu_len) &&\
1212 			   access_ok(VERIFY_WRITE, __cu_to, __cu_len))) {\
1213 			might_fault();					\
1214 			__cu_len = ___invoke_copy_in_user(__cu_to,	\
1215 							  __cu_from,	\
1216 							  __cu_len);	\
1217 		}							\
1218 	}								\
1219 	__cu_len;							\
1220 })
1221 
1222 /*
1223  * __clear_user: - Zero a block of memory in user space, with less checking.
1224  * @to:	  Destination address, in user space.
1225  * @n:	  Number of bytes to zero.
1226  *
1227  * Zero a block of memory in user space.  Caller must check
1228  * the specified block with access_ok() before calling this function.
1229  *
1230  * Returns number of bytes that could not be cleared.
1231  * On success, this will be zero.
1232  */
1233 static inline __kernel_size_t
1234 __clear_user(void __user *addr, __kernel_size_t size)
1235 {
1236 	__kernel_size_t res;
1237 
1238 	if (eva_kernel_access()) {
1239 		__asm__ __volatile__(
1240 			"move\t$4, %1\n\t"
1241 			"move\t$5, $0\n\t"
1242 			"move\t$6, %2\n\t"
1243 			__MODULE_JAL(__bzero_kernel)
1244 			"move\t%0, $6"
1245 			: "=r" (res)
1246 			: "r" (addr), "r" (size)
1247 			: "$4", "$5", "$6", __UA_t0, __UA_t1, "$31");
1248 	} else {
1249 		might_fault();
1250 		__asm__ __volatile__(
1251 			"move\t$4, %1\n\t"
1252 			"move\t$5, $0\n\t"
1253 			"move\t$6, %2\n\t"
1254 			__MODULE_JAL(__bzero)
1255 			"move\t%0, $6"
1256 			: "=r" (res)
1257 			: "r" (addr), "r" (size)
1258 			: "$4", "$5", "$6", __UA_t0, __UA_t1, "$31");
1259 	}
1260 
1261 	return res;
1262 }
1263 
1264 #define clear_user(addr,n)						\
1265 ({									\
1266 	void __user * __cl_addr = (addr);				\
1267 	unsigned long __cl_size = (n);					\
1268 	if (__cl_size && access_ok(VERIFY_WRITE,			\
1269 					__cl_addr, __cl_size))		\
1270 		__cl_size = __clear_user(__cl_addr, __cl_size);		\
1271 	__cl_size;							\
1272 })
1273 
1274 /*
1275  * __strncpy_from_user: - Copy a NUL terminated string from userspace, with less checking.
1276  * @dst:   Destination address, in kernel space.  This buffer must be at
1277  *	   least @count bytes long.
1278  * @src:   Source address, in user space.
1279  * @count: Maximum number of bytes to copy, including the trailing NUL.
1280  *
1281  * Copies a NUL-terminated string from userspace to kernel space.
1282  * Caller must check the specified block with access_ok() before calling
1283  * this function.
1284  *
1285  * On success, returns the length of the string (not including the trailing
1286  * NUL).
1287  *
1288  * If access to userspace fails, returns -EFAULT (some data may have been
1289  * copied).
1290  *
1291  * If @count is smaller than the length of the string, copies @count bytes
1292  * and returns @count.
1293  */
1294 static inline long
1295 __strncpy_from_user(char *__to, const char __user *__from, long __len)
1296 {
1297 	long res;
1298 
1299 	if (eva_kernel_access()) {
1300 		__asm__ __volatile__(
1301 			"move\t$4, %1\n\t"
1302 			"move\t$5, %2\n\t"
1303 			"move\t$6, %3\n\t"
1304 			__MODULE_JAL(__strncpy_from_kernel_nocheck_asm)
1305 			"move\t%0, $2"
1306 			: "=r" (res)
1307 			: "r" (__to), "r" (__from), "r" (__len)
1308 			: "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory");
1309 	} else {
1310 		might_fault();
1311 		__asm__ __volatile__(
1312 			"move\t$4, %1\n\t"
1313 			"move\t$5, %2\n\t"
1314 			"move\t$6, %3\n\t"
1315 			__MODULE_JAL(__strncpy_from_user_nocheck_asm)
1316 			"move\t%0, $2"
1317 			: "=r" (res)
1318 			: "r" (__to), "r" (__from), "r" (__len)
1319 			: "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory");
1320 	}
1321 
1322 	return res;
1323 }
1324 
1325 /*
1326  * strncpy_from_user: - Copy a NUL terminated string from userspace.
1327  * @dst:   Destination address, in kernel space.  This buffer must be at
1328  *	   least @count bytes long.
1329  * @src:   Source address, in user space.
1330  * @count: Maximum number of bytes to copy, including the trailing NUL.
1331  *
1332  * Copies a NUL-terminated string from userspace to kernel space.
1333  *
1334  * On success, returns the length of the string (not including the trailing
1335  * NUL).
1336  *
1337  * If access to userspace fails, returns -EFAULT (some data may have been
1338  * copied).
1339  *
1340  * If @count is smaller than the length of the string, copies @count bytes
1341  * and returns @count.
1342  */
1343 static inline long
1344 strncpy_from_user(char *__to, const char __user *__from, long __len)
1345 {
1346 	long res;
1347 
1348 	if (eva_kernel_access()) {
1349 		__asm__ __volatile__(
1350 			"move\t$4, %1\n\t"
1351 			"move\t$5, %2\n\t"
1352 			"move\t$6, %3\n\t"
1353 			__MODULE_JAL(__strncpy_from_kernel_asm)
1354 			"move\t%0, $2"
1355 			: "=r" (res)
1356 			: "r" (__to), "r" (__from), "r" (__len)
1357 			: "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory");
1358 	} else {
1359 		might_fault();
1360 		__asm__ __volatile__(
1361 			"move\t$4, %1\n\t"
1362 			"move\t$5, %2\n\t"
1363 			"move\t$6, %3\n\t"
1364 			__MODULE_JAL(__strncpy_from_user_asm)
1365 			"move\t%0, $2"
1366 			: "=r" (res)
1367 			: "r" (__to), "r" (__from), "r" (__len)
1368 			: "$2", "$3", "$4", "$5", "$6", __UA_t0, "$31", "memory");
1369 	}
1370 
1371 	return res;
1372 }
1373 
1374 /*
1375  * strlen_user: - Get the size of a string in user space.
1376  * @str: The string to measure.
1377  *
1378  * Context: User context only. This function may sleep if pagefaults are
1379  *          enabled.
1380  *
1381  * Get the size of a NUL-terminated string in user space.
1382  *
1383  * Returns the size of the string INCLUDING the terminating NUL.
1384  * On exception, returns 0.
1385  *
1386  * If there is a limit on the length of a valid string, you may wish to
1387  * consider using strnlen_user() instead.
1388  */
1389 static inline long strlen_user(const char __user *s)
1390 {
1391 	long res;
1392 
1393 	if (eva_kernel_access()) {
1394 		__asm__ __volatile__(
1395 			"move\t$4, %1\n\t"
1396 			__MODULE_JAL(__strlen_kernel_asm)
1397 			"move\t%0, $2"
1398 			: "=r" (res)
1399 			: "r" (s)
1400 			: "$2", "$4", __UA_t0, "$31");
1401 	} else {
1402 		might_fault();
1403 		__asm__ __volatile__(
1404 			"move\t$4, %1\n\t"
1405 			__MODULE_JAL(__strlen_user_asm)
1406 			"move\t%0, $2"
1407 			: "=r" (res)
1408 			: "r" (s)
1409 			: "$2", "$4", __UA_t0, "$31");
1410 	}
1411 
1412 	return res;
1413 }
1414 
1415 /* Returns: 0 if bad, string length+1 (memory size) of string if ok */
1416 static inline long __strnlen_user(const char __user *s, long n)
1417 {
1418 	long res;
1419 
1420 	if (eva_kernel_access()) {
1421 		__asm__ __volatile__(
1422 			"move\t$4, %1\n\t"
1423 			"move\t$5, %2\n\t"
1424 			__MODULE_JAL(__strnlen_kernel_nocheck_asm)
1425 			"move\t%0, $2"
1426 			: "=r" (res)
1427 			: "r" (s), "r" (n)
1428 			: "$2", "$4", "$5", __UA_t0, "$31");
1429 	} else {
1430 		might_fault();
1431 		__asm__ __volatile__(
1432 			"move\t$4, %1\n\t"
1433 			"move\t$5, %2\n\t"
1434 			__MODULE_JAL(__strnlen_user_nocheck_asm)
1435 			"move\t%0, $2"
1436 			: "=r" (res)
1437 			: "r" (s), "r" (n)
1438 			: "$2", "$4", "$5", __UA_t0, "$31");
1439 	}
1440 
1441 	return res;
1442 }
1443 
1444 /*
1445  * strnlen_user: - Get the size of a string in user space.
1446  * @str: The string to measure.
1447  *
1448  * Context: User context only. This function may sleep if pagefaults are
1449  *          enabled.
1450  *
1451  * Get the size of a NUL-terminated string in user space.
1452  *
1453  * Returns the size of the string INCLUDING the terminating NUL.
1454  * On exception, returns 0.
1455  * If the string is too long, returns a value greater than @n.
1456  */
1457 static inline long strnlen_user(const char __user *s, long n)
1458 {
1459 	long res;
1460 
1461 	might_fault();
1462 	if (eva_kernel_access()) {
1463 		__asm__ __volatile__(
1464 			"move\t$4, %1\n\t"
1465 			"move\t$5, %2\n\t"
1466 			__MODULE_JAL(__strnlen_kernel_asm)
1467 			"move\t%0, $2"
1468 			: "=r" (res)
1469 			: "r" (s), "r" (n)
1470 			: "$2", "$4", "$5", __UA_t0, "$31");
1471 	} else {
1472 		__asm__ __volatile__(
1473 			"move\t$4, %1\n\t"
1474 			"move\t$5, %2\n\t"
1475 			__MODULE_JAL(__strnlen_user_asm)
1476 			"move\t%0, $2"
1477 			: "=r" (res)
1478 			: "r" (s), "r" (n)
1479 			: "$2", "$4", "$5", __UA_t0, "$31");
1480 	}
1481 
1482 	return res;
1483 }
1484 
1485 struct exception_table_entry
1486 {
1487 	unsigned long insn;
1488 	unsigned long nextinsn;
1489 };
1490 
1491 extern int fixup_exception(struct pt_regs *regs);
1492 
1493 #endif /* _ASM_UACCESS_H */
1494