xref: /openbmc/linux/arch/parisc/kernel/entry.S (revision 7587eb18)
1/*
2 * Linux/PA-RISC Project (http://www.parisc-linux.org/)
3 *
4 * kernel entry points (interruptions, system call wrappers)
5 *  Copyright (C) 1999,2000 Philipp Rumpf
6 *  Copyright (C) 1999 SuSE GmbH Nuernberg
7 *  Copyright (C) 2000 Hewlett-Packard (John Marvin)
8 *  Copyright (C) 1999 Hewlett-Packard (Frank Rowand)
9 *
10 *    This program is free software; you can redistribute it and/or modify
11 *    it under the terms of the GNU General Public License as published by
12 *    the Free Software Foundation; either version 2, or (at your option)
13 *    any later version.
14 *
15 *    This program is distributed in the hope that it will be useful,
16 *    but WITHOUT ANY WARRANTY; without even the implied warranty of
17 *    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
18 *    GNU General Public License for more details.
19 *
20 *    You should have received a copy of the GNU General Public License
21 *    along with this program; if not, write to the Free Software
22 *    Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
23 */
24
25#include <asm/asm-offsets.h>
26
27/* we have the following possibilities to act on an interruption:
28 *  - handle in assembly and use shadowed registers only
29 *  - save registers to kernel stack and handle in assembly or C */
30
31
32#include <asm/psw.h>
33#include <asm/cache.h>		/* for L1_CACHE_SHIFT */
34#include <asm/assembly.h>	/* for LDREG/STREG defines */
35#include <asm/pgtable.h>
36#include <asm/signal.h>
37#include <asm/unistd.h>
38#include <asm/thread_info.h>
39
40#include <linux/linkage.h>
41
42#ifdef CONFIG_64BIT
43	.level 2.0w
44#else
45	.level 2.0
46#endif
47
48	.import		pa_tlb_lock,data
49
50	/* space_to_prot macro creates a prot id from a space id */
51
52#if (SPACEID_SHIFT) == 0
53	.macro  space_to_prot spc prot
54	depd,z  \spc,62,31,\prot
55	.endm
56#else
57	.macro  space_to_prot spc prot
58	extrd,u \spc,(64 - (SPACEID_SHIFT)),32,\prot
59	.endm
60#endif
61
62	/* Switch to virtual mapping, trashing only %r1 */
63	.macro  virt_map
64	/* pcxt_ssm_bug */
65	rsm	PSW_SM_I, %r0	/* barrier for "Relied upon Translation */
66	mtsp	%r0, %sr4
67	mtsp	%r0, %sr5
68	mtsp	%r0, %sr6
69	tovirt_r1 %r29
70	load32	KERNEL_PSW, %r1
71
72	rsm     PSW_SM_QUIET,%r0	/* second "heavy weight" ctl op */
73	mtctl	%r0, %cr17	/* Clear IIASQ tail */
74	mtctl	%r0, %cr17	/* Clear IIASQ head */
75	mtctl	%r1, %ipsw
76	load32	4f, %r1
77	mtctl	%r1, %cr18	/* Set IIAOQ tail */
78	ldo	4(%r1), %r1
79	mtctl	%r1, %cr18	/* Set IIAOQ head */
80	rfir
81	nop
824:
83	.endm
84
85	/*
86	 * The "get_stack" macros are responsible for determining the
87	 * kernel stack value.
88	 *
89	 *      If sr7 == 0
90	 *          Already using a kernel stack, so call the
91	 *          get_stack_use_r30 macro to push a pt_regs structure
92	 *          on the stack, and store registers there.
93	 *      else
94	 *          Need to set up a kernel stack, so call the
95	 *          get_stack_use_cr30 macro to set up a pointer
96	 *          to the pt_regs structure contained within the
97	 *          task pointer pointed to by cr30. Set the stack
98	 *          pointer to point to the end of the task structure.
99	 *
100	 * Note that we use shadowed registers for temps until
101	 * we can save %r26 and %r29. %r26 is used to preserve
102	 * %r8 (a shadowed register) which temporarily contained
103	 * either the fault type ("code") or the eirr. We need
104	 * to use a non-shadowed register to carry the value over
105	 * the rfir in virt_map. We use %r26 since this value winds
106	 * up being passed as the argument to either do_cpu_irq_mask
107	 * or handle_interruption. %r29 is used to hold a pointer
108	 * the register save area, and once again, it needs to
109	 * be a non-shadowed register so that it survives the rfir.
110	 *
111	 * N.B. TASK_SZ_ALGN and PT_SZ_ALGN include space for a stack frame.
112	 */
113
114	.macro  get_stack_use_cr30
115
116	/* we save the registers in the task struct */
117
118	copy	%r30, %r17
119	mfctl   %cr30, %r1
120	ldo	THREAD_SZ_ALGN(%r1), %r30
121	mtsp	%r0,%sr7
122	mtsp	%r16,%sr3
123	tophys  %r1,%r9
124	LDREG	TI_TASK(%r9), %r1	/* thread_info -> task_struct */
125	tophys  %r1,%r9
126	ldo     TASK_REGS(%r9),%r9
127	STREG   %r17,PT_GR30(%r9)
128	STREG   %r29,PT_GR29(%r9)
129	STREG   %r26,PT_GR26(%r9)
130	STREG	%r16,PT_SR7(%r9)
131	copy    %r9,%r29
132	.endm
133
134	.macro  get_stack_use_r30
135
136	/* we put a struct pt_regs on the stack and save the registers there */
137
138	tophys  %r30,%r9
139	copy	%r30,%r1
140	ldo	PT_SZ_ALGN(%r30),%r30
141	STREG   %r1,PT_GR30(%r9)
142	STREG   %r29,PT_GR29(%r9)
143	STREG   %r26,PT_GR26(%r9)
144	STREG	%r16,PT_SR7(%r9)
145	copy    %r9,%r29
146	.endm
147
148	.macro  rest_stack
149	LDREG   PT_GR1(%r29), %r1
150	LDREG   PT_GR30(%r29),%r30
151	LDREG   PT_GR29(%r29),%r29
152	.endm
153
154	/* default interruption handler
155	 * (calls traps.c:handle_interruption) */
156	.macro	def code
157	b	intr_save
158	ldi     \code, %r8
159	.align	32
160	.endm
161
162	/* Interrupt interruption handler
163	 * (calls irq.c:do_cpu_irq_mask) */
164	.macro	extint code
165	b	intr_extint
166	mfsp    %sr7,%r16
167	.align	32
168	.endm
169
170	.import	os_hpmc, code
171
172	/* HPMC handler */
173	.macro	hpmc code
174	nop			/* must be a NOP, will be patched later */
175	load32	PA(os_hpmc), %r3
176	bv,n	0(%r3)
177	nop
178	.word	0		/* checksum (will be patched) */
179	.word	PA(os_hpmc)	/* address of handler */
180	.word	0		/* length of handler */
181	.endm
182
183	/*
184	 * Performance Note: Instructions will be moved up into
185	 * this part of the code later on, once we are sure
186	 * that the tlb miss handlers are close to final form.
187	 */
188
189	/* Register definitions for tlb miss handler macros */
190
191	va  = r8	/* virtual address for which the trap occurred */
192	spc = r24	/* space for which the trap occurred */
193
194#ifndef CONFIG_64BIT
195
196	/*
197	 * itlb miss interruption handler (parisc 1.1 - 32 bit)
198	 */
199
200	.macro	itlb_11 code
201
202	mfctl	%pcsq, spc
203	b	itlb_miss_11
204	mfctl	%pcoq, va
205
206	.align		32
207	.endm
208#endif
209
210	/*
211	 * itlb miss interruption handler (parisc 2.0)
212	 */
213
214	.macro	itlb_20 code
215	mfctl	%pcsq, spc
216#ifdef CONFIG_64BIT
217	b       itlb_miss_20w
218#else
219	b	itlb_miss_20
220#endif
221	mfctl	%pcoq, va
222
223	.align		32
224	.endm
225
226#ifndef CONFIG_64BIT
227	/*
228	 * naitlb miss interruption handler (parisc 1.1 - 32 bit)
229	 */
230
231	.macro	naitlb_11 code
232
233	mfctl	%isr,spc
234	b	naitlb_miss_11
235	mfctl 	%ior,va
236
237	.align		32
238	.endm
239#endif
240
241	/*
242	 * naitlb miss interruption handler (parisc 2.0)
243	 */
244
245	.macro	naitlb_20 code
246
247	mfctl	%isr,spc
248#ifdef CONFIG_64BIT
249	b       naitlb_miss_20w
250#else
251	b	naitlb_miss_20
252#endif
253	mfctl 	%ior,va
254
255	.align		32
256	.endm
257
258#ifndef CONFIG_64BIT
259	/*
260	 * dtlb miss interruption handler (parisc 1.1 - 32 bit)
261	 */
262
263	.macro	dtlb_11 code
264
265	mfctl	%isr, spc
266	b	dtlb_miss_11
267	mfctl	%ior, va
268
269	.align		32
270	.endm
271#endif
272
273	/*
274	 * dtlb miss interruption handler (parisc 2.0)
275	 */
276
277	.macro	dtlb_20 code
278
279	mfctl	%isr, spc
280#ifdef CONFIG_64BIT
281	b       dtlb_miss_20w
282#else
283	b	dtlb_miss_20
284#endif
285	mfctl	%ior, va
286
287	.align		32
288	.endm
289
290#ifndef CONFIG_64BIT
291	/* nadtlb miss interruption handler (parisc 1.1 - 32 bit) */
292
293	.macro	nadtlb_11 code
294
295	mfctl	%isr,spc
296	b       nadtlb_miss_11
297	mfctl	%ior,va
298
299	.align		32
300	.endm
301#endif
302
303	/* nadtlb miss interruption handler (parisc 2.0) */
304
305	.macro	nadtlb_20 code
306
307	mfctl	%isr,spc
308#ifdef CONFIG_64BIT
309	b       nadtlb_miss_20w
310#else
311	b       nadtlb_miss_20
312#endif
313	mfctl	%ior,va
314
315	.align		32
316	.endm
317
318#ifndef CONFIG_64BIT
319	/*
320	 * dirty bit trap interruption handler (parisc 1.1 - 32 bit)
321	 */
322
323	.macro	dbit_11 code
324
325	mfctl	%isr,spc
326	b	dbit_trap_11
327	mfctl	%ior,va
328
329	.align		32
330	.endm
331#endif
332
333	/*
334	 * dirty bit trap interruption handler (parisc 2.0)
335	 */
336
337	.macro	dbit_20 code
338
339	mfctl	%isr,spc
340#ifdef CONFIG_64BIT
341	b       dbit_trap_20w
342#else
343	b	dbit_trap_20
344#endif
345	mfctl	%ior,va
346
347	.align		32
348	.endm
349
350	/* In LP64, the space contains part of the upper 32 bits of the
351	 * fault.  We have to extract this and place it in the va,
352	 * zeroing the corresponding bits in the space register */
353	.macro		space_adjust	spc,va,tmp
354#ifdef CONFIG_64BIT
355	extrd,u		\spc,63,SPACEID_SHIFT,\tmp
356	depd		%r0,63,SPACEID_SHIFT,\spc
357	depd		\tmp,31,SPACEID_SHIFT,\va
358#endif
359	.endm
360
361	.import		swapper_pg_dir,code
362
363	/* Get the pgd.  For faults on space zero (kernel space), this
364	 * is simply swapper_pg_dir.  For user space faults, the
365	 * pgd is stored in %cr25 */
366	.macro		get_pgd		spc,reg
367	ldil		L%PA(swapper_pg_dir),\reg
368	ldo		R%PA(swapper_pg_dir)(\reg),\reg
369	or,COND(=)	%r0,\spc,%r0
370	mfctl		%cr25,\reg
371	.endm
372
373	/*
374		space_check(spc,tmp,fault)
375
376		spc - The space we saw the fault with.
377		tmp - The place to store the current space.
378		fault - Function to call on failure.
379
380		Only allow faults on different spaces from the
381		currently active one if we're the kernel
382
383	*/
384	.macro		space_check	spc,tmp,fault
385	mfsp		%sr7,\tmp
386	or,COND(<>)	%r0,\spc,%r0	/* user may execute gateway page
387					 * as kernel, so defeat the space
388					 * check if it is */
389	copy		\spc,\tmp
390	or,COND(=)	%r0,\tmp,%r0	/* nullify if executing as kernel */
391	cmpb,COND(<>),n	\tmp,\spc,\fault
392	.endm
393
394	/* Look up a PTE in a 2-Level scheme (faulting at each
395	 * level if the entry isn't present
396	 *
397	 * NOTE: we use ldw even for LP64, since the short pointers
398	 * can address up to 1TB
399	 */
400	.macro		L2_ptep	pmd,pte,index,va,fault
401#if CONFIG_PGTABLE_LEVELS == 3
402	extru		\va,31-ASM_PMD_SHIFT,ASM_BITS_PER_PMD,\index
403#else
404# if defined(CONFIG_64BIT)
405	extrd,u		\va,63-ASM_PGDIR_SHIFT,ASM_BITS_PER_PGD,\index
406  #else
407  # if PAGE_SIZE > 4096
408	extru		\va,31-ASM_PGDIR_SHIFT,32-ASM_PGDIR_SHIFT,\index
409  # else
410	extru		\va,31-ASM_PGDIR_SHIFT,ASM_BITS_PER_PGD,\index
411  # endif
412# endif
413#endif
414	dep             %r0,31,PAGE_SHIFT,\pmd  /* clear offset */
415	copy		%r0,\pte
416	ldw,s		\index(\pmd),\pmd
417	bb,>=,n		\pmd,_PxD_PRESENT_BIT,\fault
418	dep		%r0,31,PxD_FLAG_SHIFT,\pmd /* clear flags */
419	copy		\pmd,%r9
420	SHLREG		%r9,PxD_VALUE_SHIFT,\pmd
421	extru		\va,31-PAGE_SHIFT,ASM_BITS_PER_PTE,\index
422	dep		%r0,31,PAGE_SHIFT,\pmd  /* clear offset */
423	shladd		\index,BITS_PER_PTE_ENTRY,\pmd,\pmd /* pmd is now pte */
424	LDREG		%r0(\pmd),\pte
425	bb,>=,n		\pte,_PAGE_PRESENT_BIT,\fault
426	.endm
427
428	/* Look up PTE in a 3-Level scheme.
429	 *
430	 * Here we implement a Hybrid L2/L3 scheme: we allocate the
431	 * first pmd adjacent to the pgd.  This means that we can
432	 * subtract a constant offset to get to it.  The pmd and pgd
433	 * sizes are arranged so that a single pmd covers 4GB (giving
434	 * a full LP64 process access to 8TB) so our lookups are
435	 * effectively L2 for the first 4GB of the kernel (i.e. for
436	 * all ILP32 processes and all the kernel for machines with
437	 * under 4GB of memory) */
438	.macro		L3_ptep pgd,pte,index,va,fault
439#if CONFIG_PGTABLE_LEVELS == 3 /* we might have a 2-Level scheme, e.g. with 16kb page size */
440	extrd,u		\va,63-ASM_PGDIR_SHIFT,ASM_BITS_PER_PGD,\index
441	copy		%r0,\pte
442	extrd,u,*=	\va,63-ASM_PGDIR_SHIFT,64-ASM_PGDIR_SHIFT,%r0
443	ldw,s		\index(\pgd),\pgd
444	extrd,u,*=	\va,63-ASM_PGDIR_SHIFT,64-ASM_PGDIR_SHIFT,%r0
445	bb,>=,n		\pgd,_PxD_PRESENT_BIT,\fault
446	extrd,u,*=	\va,63-ASM_PGDIR_SHIFT,64-ASM_PGDIR_SHIFT,%r0
447	shld		\pgd,PxD_VALUE_SHIFT,\index
448	extrd,u,*=	\va,63-ASM_PGDIR_SHIFT,64-ASM_PGDIR_SHIFT,%r0
449	copy		\index,\pgd
450	extrd,u,*<>	\va,63-ASM_PGDIR_SHIFT,64-ASM_PGDIR_SHIFT,%r0
451	ldo		ASM_PGD_PMD_OFFSET(\pgd),\pgd
452#endif
453	L2_ptep		\pgd,\pte,\index,\va,\fault
454	.endm
455
456	/* Acquire pa_tlb_lock lock and recheck page is still present. */
457	.macro		tlb_lock	spc,ptp,pte,tmp,tmp1,fault
458#ifdef CONFIG_SMP
459	cmpib,COND(=),n	0,\spc,2f
460	load32		PA(pa_tlb_lock),\tmp
4611:	LDCW		0(\tmp),\tmp1
462	cmpib,COND(=)	0,\tmp1,1b
463	nop
464	LDREG		0(\ptp),\pte
465	bb,<,n		\pte,_PAGE_PRESENT_BIT,2f
466	b		\fault
467	stw		 \spc,0(\tmp)
4682:
469#endif
470	.endm
471
472	/* Release pa_tlb_lock lock without reloading lock address. */
473	.macro		tlb_unlock0	spc,tmp
474#ifdef CONFIG_SMP
475	or,COND(=)	%r0,\spc,%r0
476	stw             \spc,0(\tmp)
477#endif
478	.endm
479
480	/* Release pa_tlb_lock lock. */
481	.macro		tlb_unlock1	spc,tmp
482#ifdef CONFIG_SMP
483	load32		PA(pa_tlb_lock),\tmp
484	tlb_unlock0	\spc,\tmp
485#endif
486	.endm
487
488	/* Set the _PAGE_ACCESSED bit of the PTE.  Be clever and
489	 * don't needlessly dirty the cache line if it was already set */
490	.macro		update_accessed	ptp,pte,tmp,tmp1
491	ldi		_PAGE_ACCESSED,\tmp1
492	or		\tmp1,\pte,\tmp
493	and,COND(<>)	\tmp1,\pte,%r0
494	STREG		\tmp,0(\ptp)
495	.endm
496
497	/* Set the dirty bit (and accessed bit).  No need to be
498	 * clever, this is only used from the dirty fault */
499	.macro		update_dirty	ptp,pte,tmp
500	ldi		_PAGE_ACCESSED|_PAGE_DIRTY,\tmp
501	or		\tmp,\pte,\pte
502	STREG		\pte,0(\ptp)
503	.endm
504
505	/* We have (depending on the page size):
506	 * - 38 to 52-bit Physical Page Number
507	 * - 12 to 26-bit page offset
508	 */
509	/* bitshift difference between a PFN (based on kernel's PAGE_SIZE)
510	 * to a CPU TLB 4k PFN (4k => 12 bits to shift) */
511	#define PAGE_ADD_SHIFT		(PAGE_SHIFT-12)
512	#define PAGE_ADD_HUGE_SHIFT	(REAL_HPAGE_SHIFT-12)
513
514	/* Drop prot bits and convert to page addr for iitlbt and idtlbt */
515	.macro		convert_for_tlb_insert20 pte,tmp
516#ifdef CONFIG_HUGETLB_PAGE
517	copy		\pte,\tmp
518	extrd,u		\tmp,(63-ASM_PFN_PTE_SHIFT)+(63-58)+PAGE_ADD_SHIFT,\
519				64-PAGE_SHIFT-PAGE_ADD_SHIFT,\pte
520
521	depdi		_PAGE_SIZE_ENCODING_DEFAULT,63,\
522				(63-58)+PAGE_ADD_SHIFT,\pte
523	extrd,u,*=	\tmp,_PAGE_HPAGE_BIT+32,1,%r0
524	depdi		_HUGE_PAGE_SIZE_ENCODING_DEFAULT,63,\
525				(63-58)+PAGE_ADD_HUGE_SHIFT,\pte
526#else /* Huge pages disabled */
527	extrd,u		\pte,(63-ASM_PFN_PTE_SHIFT)+(63-58)+PAGE_ADD_SHIFT,\
528				64-PAGE_SHIFT-PAGE_ADD_SHIFT,\pte
529	depdi		_PAGE_SIZE_ENCODING_DEFAULT,63,\
530				(63-58)+PAGE_ADD_SHIFT,\pte
531#endif
532	.endm
533
534	/* Convert the pte and prot to tlb insertion values.  How
535	 * this happens is quite subtle, read below */
536	.macro		make_insert_tlb	spc,pte,prot,tmp
537	space_to_prot   \spc \prot        /* create prot id from space */
538	/* The following is the real subtlety.  This is depositing
539	 * T <-> _PAGE_REFTRAP
540	 * D <-> _PAGE_DIRTY
541	 * B <-> _PAGE_DMB (memory break)
542	 *
543	 * Then incredible subtlety: The access rights are
544	 * _PAGE_GATEWAY, _PAGE_EXEC and _PAGE_WRITE
545	 * See 3-14 of the parisc 2.0 manual
546	 *
547	 * Finally, _PAGE_READ goes in the top bit of PL1 (so we
548	 * trigger an access rights trap in user space if the user
549	 * tries to read an unreadable page */
550	depd            \pte,8,7,\prot
551
552	/* PAGE_USER indicates the page can be read with user privileges,
553	 * so deposit X1|11 to PL1|PL2 (remember the upper bit of PL1
554	 * contains _PAGE_READ) */
555	extrd,u,*=      \pte,_PAGE_USER_BIT+32,1,%r0
556	depdi		7,11,3,\prot
557	/* If we're a gateway page, drop PL2 back to zero for promotion
558	 * to kernel privilege (so we can execute the page as kernel).
559	 * Any privilege promotion page always denys read and write */
560	extrd,u,*= 	\pte,_PAGE_GATEWAY_BIT+32,1,%r0
561	depd		%r0,11,2,\prot	/* If Gateway, Set PL2 to 0 */
562
563	/* Enforce uncacheable pages.
564	 * This should ONLY be use for MMIO on PA 2.0 machines.
565	 * Memory/DMA is cache coherent on all PA2.0 machines we support
566	 * (that means T-class is NOT supported) and the memory controllers
567	 * on most of those machines only handles cache transactions.
568	 */
569	extrd,u,*=	\pte,_PAGE_NO_CACHE_BIT+32,1,%r0
570	depdi		1,12,1,\prot
571
572	/* Drop prot bits and convert to page addr for iitlbt and idtlbt */
573	convert_for_tlb_insert20 \pte \tmp
574	.endm
575
576	/* Identical macro to make_insert_tlb above, except it
577	 * makes the tlb entry for the differently formatted pa11
578	 * insertion instructions */
579	.macro		make_insert_tlb_11	spc,pte,prot
580	zdep		\spc,30,15,\prot
581	dep		\pte,8,7,\prot
582	extru,=		\pte,_PAGE_NO_CACHE_BIT,1,%r0
583	depi		1,12,1,\prot
584	extru,=         \pte,_PAGE_USER_BIT,1,%r0
585	depi		7,11,3,\prot   /* Set for user space (1 rsvd for read) */
586	extru,= 	\pte,_PAGE_GATEWAY_BIT,1,%r0
587	depi		0,11,2,\prot	/* If Gateway, Set PL2 to 0 */
588
589	/* Get rid of prot bits and convert to page addr for iitlba */
590
591	depi		0,31,ASM_PFN_PTE_SHIFT,\pte
592	SHRREG		\pte,(ASM_PFN_PTE_SHIFT-(31-26)),\pte
593	.endm
594
595	/* This is for ILP32 PA2.0 only.  The TLB insertion needs
596	 * to extend into I/O space if the address is 0xfXXXXXXX
597	 * so we extend the f's into the top word of the pte in
598	 * this case */
599	.macro		f_extend	pte,tmp
600	extrd,s		\pte,42,4,\tmp
601	addi,<>		1,\tmp,%r0
602	extrd,s		\pte,63,25,\pte
603	.endm
604
605	/* The alias region is an 8MB aligned 16MB to do clear and
606	 * copy user pages at addresses congruent with the user
607	 * virtual address.
608	 *
609	 * To use the alias page, you set %r26 up with the to TLB
610	 * entry (identifying the physical page) and %r23 up with
611	 * the from tlb entry (or nothing if only a to entry---for
612	 * clear_user_page_asm) */
613	.macro		do_alias	spc,tmp,tmp1,va,pte,prot,fault,patype
614	cmpib,COND(<>),n 0,\spc,\fault
615	ldil		L%(TMPALIAS_MAP_START),\tmp
616#if defined(CONFIG_64BIT) && (TMPALIAS_MAP_START >= 0x80000000)
617	/* on LP64, ldi will sign extend into the upper 32 bits,
618	 * which is behaviour we don't want */
619	depdi		0,31,32,\tmp
620#endif
621	copy		\va,\tmp1
622	depi		0,31,23,\tmp1
623	cmpb,COND(<>),n	\tmp,\tmp1,\fault
624	mfctl		%cr19,\tmp	/* iir */
625	/* get the opcode (first six bits) into \tmp */
626	extrw,u		\tmp,5,6,\tmp
627	/*
628	 * Only setting the T bit prevents data cache movein
629	 * Setting access rights to zero prevents instruction cache movein
630	 *
631	 * Note subtlety here: _PAGE_GATEWAY, _PAGE_EXEC and _PAGE_WRITE go
632	 * to type field and _PAGE_READ goes to top bit of PL1
633	 */
634	ldi		(_PAGE_REFTRAP|_PAGE_READ|_PAGE_WRITE),\prot
635	/*
636	 * so if the opcode is one (i.e. this is a memory management
637	 * instruction) nullify the next load so \prot is only T.
638	 * Otherwise this is a normal data operation
639	 */
640	cmpiclr,=	0x01,\tmp,%r0
641	ldi		(_PAGE_DIRTY|_PAGE_READ|_PAGE_WRITE),\prot
642.ifc \patype,20
643	depd,z		\prot,8,7,\prot
644.else
645.ifc \patype,11
646	depw,z		\prot,8,7,\prot
647.else
648	.error "undefined PA type to do_alias"
649.endif
650.endif
651	/*
652	 * OK, it is in the temp alias region, check whether "from" or "to".
653	 * Check "subtle" note in pacache.S re: r23/r26.
654	 */
655#ifdef CONFIG_64BIT
656	extrd,u,*=	\va,41,1,%r0
657#else
658	extrw,u,=	\va,9,1,%r0
659#endif
660	or,COND(tr)	%r23,%r0,\pte
661	or		%r26,%r0,\pte
662	.endm
663
664
665	/*
666	 * Fault_vectors are architecturally required to be aligned on a 2K
667	 * boundary
668	 */
669
670	.section .text.hot
671	.align 2048
672
673ENTRY(fault_vector_20)
674	/* First vector is invalid (0) */
675	.ascii	"cows can fly"
676	.byte 0
677	.align 32
678
679	hpmc		 1
680	def		 2
681	def		 3
682	extint		 4
683	def		 5
684	itlb_20		 6
685	def		 7
686	def		 8
687	def              9
688	def		10
689	def		11
690	def		12
691	def		13
692	def		14
693	dtlb_20		15
694	naitlb_20	16
695	nadtlb_20	17
696	def		18
697	def		19
698	dbit_20		20
699	def		21
700	def		22
701	def		23
702	def		24
703	def		25
704	def		26
705	def		27
706	def		28
707	def		29
708	def		30
709	def		31
710END(fault_vector_20)
711
712#ifndef CONFIG_64BIT
713
714	.align 2048
715
716ENTRY(fault_vector_11)
717	/* First vector is invalid (0) */
718	.ascii	"cows can fly"
719	.byte 0
720	.align 32
721
722	hpmc		 1
723	def		 2
724	def		 3
725	extint		 4
726	def		 5
727	itlb_11		 6
728	def		 7
729	def		 8
730	def              9
731	def		10
732	def		11
733	def		12
734	def		13
735	def		14
736	dtlb_11		15
737	naitlb_11	16
738	nadtlb_11	17
739	def		18
740	def		19
741	dbit_11		20
742	def		21
743	def		22
744	def		23
745	def		24
746	def		25
747	def		26
748	def		27
749	def		28
750	def		29
751	def		30
752	def		31
753END(fault_vector_11)
754
755#endif
756	/* Fault vector is separately protected and *must* be on its own page */
757	.align		PAGE_SIZE
758ENTRY(end_fault_vector)
759
760	.import		handle_interruption,code
761	.import		do_cpu_irq_mask,code
762
763	/*
764	 * Child Returns here
765	 *
766	 * copy_thread moved args into task save area.
767	 */
768
769ENTRY(ret_from_kernel_thread)
770
771	/* Call schedule_tail first though */
772	BL	schedule_tail, %r2
773	nop
774
775	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
776	LDREG	TASK_PT_GR25(%r1), %r26
777#ifdef CONFIG_64BIT
778	LDREG	TASK_PT_GR27(%r1), %r27
779#endif
780	LDREG	TASK_PT_GR26(%r1), %r1
781	ble	0(%sr7, %r1)
782	copy	%r31, %r2
783	b	finish_child_return
784	nop
785ENDPROC(ret_from_kernel_thread)
786
787
788	/*
789	 * struct task_struct *_switch_to(struct task_struct *prev,
790	 *	struct task_struct *next)
791	 *
792	 * switch kernel stacks and return prev */
793ENTRY(_switch_to)
794	STREG	 %r2, -RP_OFFSET(%r30)
795
796	callee_save_float
797	callee_save
798
799	load32	_switch_to_ret, %r2
800
801	STREG	%r2, TASK_PT_KPC(%r26)
802	LDREG	TASK_PT_KPC(%r25), %r2
803
804	STREG	%r30, TASK_PT_KSP(%r26)
805	LDREG	TASK_PT_KSP(%r25), %r30
806	LDREG	TASK_THREAD_INFO(%r25), %r25
807	bv	%r0(%r2)
808	mtctl   %r25,%cr30
809
810_switch_to_ret:
811	mtctl	%r0, %cr0		/* Needed for single stepping */
812	callee_rest
813	callee_rest_float
814
815	LDREG	-RP_OFFSET(%r30), %r2
816	bv	%r0(%r2)
817	copy	%r26, %r28
818ENDPROC(_switch_to)
819
820	/*
821	 * Common rfi return path for interruptions, kernel execve, and
822	 * sys_rt_sigreturn (sometimes).  The sys_rt_sigreturn syscall will
823	 * return via this path if the signal was received when the process
824	 * was running; if the process was blocked on a syscall then the
825	 * normal syscall_exit path is used.  All syscalls for traced
826	 * proceses exit via intr_restore.
827	 *
828	 * XXX If any syscalls that change a processes space id ever exit
829	 * this way, then we will need to copy %sr3 in to PT_SR[3..7], and
830	 * adjust IASQ[0..1].
831	 *
832	 */
833
834	.align	PAGE_SIZE
835
836ENTRY(syscall_exit_rfi)
837	mfctl   %cr30,%r16
838	LDREG	TI_TASK(%r16), %r16	/* thread_info -> task_struct */
839	ldo	TASK_REGS(%r16),%r16
840	/* Force iaoq to userspace, as the user has had access to our current
841	 * context via sigcontext. Also Filter the PSW for the same reason.
842	 */
843	LDREG	PT_IAOQ0(%r16),%r19
844	depi	3,31,2,%r19
845	STREG	%r19,PT_IAOQ0(%r16)
846	LDREG	PT_IAOQ1(%r16),%r19
847	depi	3,31,2,%r19
848	STREG	%r19,PT_IAOQ1(%r16)
849	LDREG   PT_PSW(%r16),%r19
850	load32	USER_PSW_MASK,%r1
851#ifdef CONFIG_64BIT
852	load32	USER_PSW_HI_MASK,%r20
853	depd    %r20,31,32,%r1
854#endif
855	and     %r19,%r1,%r19 /* Mask out bits that user shouldn't play with */
856	load32	USER_PSW,%r1
857	or      %r19,%r1,%r19 /* Make sure default USER_PSW bits are set */
858	STREG   %r19,PT_PSW(%r16)
859
860	/*
861	 * If we aren't being traced, we never saved space registers
862	 * (we don't store them in the sigcontext), so set them
863	 * to "proper" values now (otherwise we'll wind up restoring
864	 * whatever was last stored in the task structure, which might
865	 * be inconsistent if an interrupt occurred while on the gateway
866	 * page). Note that we may be "trashing" values the user put in
867	 * them, but we don't support the user changing them.
868	 */
869
870	STREG   %r0,PT_SR2(%r16)
871	mfsp    %sr3,%r19
872	STREG   %r19,PT_SR0(%r16)
873	STREG   %r19,PT_SR1(%r16)
874	STREG   %r19,PT_SR3(%r16)
875	STREG   %r19,PT_SR4(%r16)
876	STREG   %r19,PT_SR5(%r16)
877	STREG   %r19,PT_SR6(%r16)
878	STREG   %r19,PT_SR7(%r16)
879
880intr_return:
881	/* check for reschedule */
882	mfctl   %cr30,%r1
883	LDREG   TI_FLAGS(%r1),%r19	/* sched.h: TIF_NEED_RESCHED */
884	bb,<,n	%r19,31-TIF_NEED_RESCHED,intr_do_resched /* forward */
885
886	.import do_notify_resume,code
887intr_check_sig:
888	/* As above */
889	mfctl   %cr30,%r1
890	LDREG	TI_FLAGS(%r1),%r19
891	ldi	(_TIF_SIGPENDING|_TIF_NOTIFY_RESUME), %r20
892	and,COND(<>)	%r19, %r20, %r0
893	b,n	intr_restore	/* skip past if we've nothing to do */
894
895	/* This check is critical to having LWS
896	 * working. The IASQ is zero on the gateway
897	 * page and we cannot deliver any signals until
898	 * we get off the gateway page.
899	 *
900	 * Only do signals if we are returning to user space
901	 */
902	LDREG	PT_IASQ0(%r16), %r20
903	cmpib,COND(=),n 0,%r20,intr_restore /* backward */
904	LDREG	PT_IASQ1(%r16), %r20
905	cmpib,COND(=),n 0,%r20,intr_restore /* backward */
906
907	/* NOTE: We need to enable interrupts if we have to deliver
908	 * signals. We used to do this earlier but it caused kernel
909	 * stack overflows. */
910	ssm     PSW_SM_I, %r0
911
912	copy	%r0, %r25			/* long in_syscall = 0 */
913#ifdef CONFIG_64BIT
914	ldo	-16(%r30),%r29			/* Reference param save area */
915#endif
916
917	BL	do_notify_resume,%r2
918	copy	%r16, %r26			/* struct pt_regs *regs */
919
920	b,n	intr_check_sig
921
922intr_restore:
923	copy            %r16,%r29
924	ldo             PT_FR31(%r29),%r1
925	rest_fp         %r1
926	rest_general    %r29
927
928	/* inverse of virt_map */
929	pcxt_ssm_bug
930	rsm             PSW_SM_QUIET,%r0	/* prepare for rfi */
931	tophys_r1       %r29
932
933	/* Restore space id's and special cr's from PT_REGS
934	 * structure pointed to by r29
935	 */
936	rest_specials	%r29
937
938	/* IMPORTANT: rest_stack restores r29 last (we are using it)!
939	 * It also restores r1 and r30.
940	 */
941	rest_stack
942
943	rfi
944	nop
945
946#ifndef CONFIG_PREEMPT
947# define intr_do_preempt	intr_restore
948#endif /* !CONFIG_PREEMPT */
949
950	.import schedule,code
951intr_do_resched:
952	/* Only call schedule on return to userspace. If we're returning
953	 * to kernel space, we may schedule if CONFIG_PREEMPT, otherwise
954	 * we jump back to intr_restore.
955	 */
956	LDREG	PT_IASQ0(%r16), %r20
957	cmpib,COND(=)	0, %r20, intr_do_preempt
958	nop
959	LDREG	PT_IASQ1(%r16), %r20
960	cmpib,COND(=)	0, %r20, intr_do_preempt
961	nop
962
963	/* NOTE: We need to enable interrupts if we schedule.  We used
964	 * to do this earlier but it caused kernel stack overflows. */
965	ssm     PSW_SM_I, %r0
966
967#ifdef CONFIG_64BIT
968	ldo	-16(%r30),%r29		/* Reference param save area */
969#endif
970
971	ldil	L%intr_check_sig, %r2
972#ifndef CONFIG_64BIT
973	b	schedule
974#else
975	load32	schedule, %r20
976	bv	%r0(%r20)
977#endif
978	ldo	R%intr_check_sig(%r2), %r2
979
980	/* preempt the current task on returning to kernel
981	 * mode from an interrupt, iff need_resched is set,
982	 * and preempt_count is 0. otherwise, we continue on
983	 * our merry way back to the current running task.
984	 */
985#ifdef CONFIG_PREEMPT
986	.import preempt_schedule_irq,code
987intr_do_preempt:
988	rsm	PSW_SM_I, %r0		/* disable interrupts */
989
990	/* current_thread_info()->preempt_count */
991	mfctl	%cr30, %r1
992	LDREG	TI_PRE_COUNT(%r1), %r19
993	cmpib,COND(<>)	0, %r19, intr_restore	/* if preempt_count > 0 */
994	nop				/* prev insn branched backwards */
995
996	/* check if we interrupted a critical path */
997	LDREG	PT_PSW(%r16), %r20
998	bb,<,n	%r20, 31 - PSW_SM_I, intr_restore
999	nop
1000
1001	BL	preempt_schedule_irq, %r2
1002	nop
1003
1004	b,n	intr_restore		/* ssm PSW_SM_I done by intr_restore */
1005#endif /* CONFIG_PREEMPT */
1006
1007	/*
1008	 * External interrupts.
1009	 */
1010
1011intr_extint:
1012	cmpib,COND(=),n 0,%r16,1f
1013
1014	get_stack_use_cr30
1015	b,n 2f
1016
10171:
1018	get_stack_use_r30
10192:
1020	save_specials	%r29
1021	virt_map
1022	save_general	%r29
1023
1024	ldo	PT_FR0(%r29), %r24
1025	save_fp	%r24
1026
1027	loadgp
1028
1029	copy	%r29, %r26	/* arg0 is pt_regs */
1030	copy	%r29, %r16	/* save pt_regs */
1031
1032	ldil	L%intr_return, %r2
1033
1034#ifdef CONFIG_64BIT
1035	ldo	-16(%r30),%r29	/* Reference param save area */
1036#endif
1037
1038	b	do_cpu_irq_mask
1039	ldo	R%intr_return(%r2), %r2	/* return to intr_return, not here */
1040ENDPROC(syscall_exit_rfi)
1041
1042
1043	/* Generic interruptions (illegal insn, unaligned, page fault, etc) */
1044
1045ENTRY(intr_save)		/* for os_hpmc */
1046	mfsp    %sr7,%r16
1047	cmpib,COND(=),n 0,%r16,1f
1048	get_stack_use_cr30
1049	b	2f
1050	copy    %r8,%r26
1051
10521:
1053	get_stack_use_r30
1054	copy    %r8,%r26
1055
10562:
1057	save_specials	%r29
1058
1059	/* If this trap is a itlb miss, skip saving/adjusting isr/ior */
1060
1061	/*
1062	 * FIXME: 1) Use a #define for the hardwired "6" below (and in
1063	 *           traps.c.
1064	 *        2) Once we start executing code above 4 Gb, we need
1065	 *           to adjust iasq/iaoq here in the same way we
1066	 *           adjust isr/ior below.
1067	 */
1068
1069	cmpib,COND(=),n        6,%r26,skip_save_ior
1070
1071
1072	mfctl           %cr20, %r16 /* isr */
1073	nop		/* serialize mfctl on PA 2.0 to avoid 4 cycle penalty */
1074	mfctl           %cr21, %r17 /* ior */
1075
1076
1077#ifdef CONFIG_64BIT
1078	/*
1079	 * If the interrupted code was running with W bit off (32 bit),
1080	 * clear the b bits (bits 0 & 1) in the ior.
1081	 * save_specials left ipsw value in r8 for us to test.
1082	 */
1083	extrd,u,*<>     %r8,PSW_W_BIT,1,%r0
1084	depdi           0,1,2,%r17
1085
1086	/*
1087	 * FIXME: This code has hardwired assumptions about the split
1088	 *        between space bits and offset bits. This will change
1089	 *        when we allow alternate page sizes.
1090	 */
1091
1092	/* adjust isr/ior. */
1093	extrd,u         %r16,63,SPACEID_SHIFT,%r1	/* get high bits from isr for ior */
1094	depd            %r1,31,SPACEID_SHIFT,%r17	/* deposit them into ior */
1095	depdi           0,63,SPACEID_SHIFT,%r16		/* clear them from isr */
1096#endif
1097	STREG           %r16, PT_ISR(%r29)
1098	STREG           %r17, PT_IOR(%r29)
1099
1100
1101skip_save_ior:
1102	virt_map
1103	save_general	%r29
1104
1105	ldo		PT_FR0(%r29), %r25
1106	save_fp		%r25
1107
1108	loadgp
1109
1110	copy		%r29, %r25	/* arg1 is pt_regs */
1111#ifdef CONFIG_64BIT
1112	ldo		-16(%r30),%r29	/* Reference param save area */
1113#endif
1114
1115	ldil		L%intr_check_sig, %r2
1116	copy		%r25, %r16	/* save pt_regs */
1117
1118	b		handle_interruption
1119	ldo		R%intr_check_sig(%r2), %r2
1120ENDPROC(intr_save)
1121
1122
1123	/*
1124	 * Note for all tlb miss handlers:
1125	 *
1126	 * cr24 contains a pointer to the kernel address space
1127	 * page directory.
1128	 *
1129	 * cr25 contains a pointer to the current user address
1130	 * space page directory.
1131	 *
1132	 * sr3 will contain the space id of the user address space
1133	 * of the current running thread while that thread is
1134	 * running in the kernel.
1135	 */
1136
1137	/*
1138	 * register number allocations.  Note that these are all
1139	 * in the shadowed registers
1140	 */
1141
1142	t0 = r1		/* temporary register 0 */
1143	va = r8		/* virtual address for which the trap occurred */
1144	t1 = r9		/* temporary register 1 */
1145	pte  = r16	/* pte/phys page # */
1146	prot = r17	/* prot bits */
1147	spc  = r24	/* space for which the trap occurred */
1148	ptp = r25	/* page directory/page table pointer */
1149
1150#ifdef CONFIG_64BIT
1151
1152dtlb_miss_20w:
1153	space_adjust	spc,va,t0
1154	get_pgd		spc,ptp
1155	space_check	spc,t0,dtlb_fault
1156
1157	L3_ptep		ptp,pte,t0,va,dtlb_check_alias_20w
1158
1159	tlb_lock	spc,ptp,pte,t0,t1,dtlb_check_alias_20w
1160	update_accessed	ptp,pte,t0,t1
1161
1162	make_insert_tlb	spc,pte,prot,t1
1163
1164	idtlbt          pte,prot
1165
1166	tlb_unlock1	spc,t0
1167	rfir
1168	nop
1169
1170dtlb_check_alias_20w:
1171	do_alias	spc,t0,t1,va,pte,prot,dtlb_fault,20
1172
1173	idtlbt          pte,prot
1174
1175	rfir
1176	nop
1177
1178nadtlb_miss_20w:
1179	space_adjust	spc,va,t0
1180	get_pgd		spc,ptp
1181	space_check	spc,t0,nadtlb_fault
1182
1183	L3_ptep		ptp,pte,t0,va,nadtlb_check_alias_20w
1184
1185	tlb_lock	spc,ptp,pte,t0,t1,nadtlb_check_alias_20w
1186	update_accessed	ptp,pte,t0,t1
1187
1188	make_insert_tlb	spc,pte,prot,t1
1189
1190	idtlbt          pte,prot
1191
1192	tlb_unlock1	spc,t0
1193	rfir
1194	nop
1195
1196nadtlb_check_alias_20w:
1197	do_alias	spc,t0,t1,va,pte,prot,nadtlb_emulate,20
1198
1199	idtlbt          pte,prot
1200
1201	rfir
1202	nop
1203
1204#else
1205
1206dtlb_miss_11:
1207	get_pgd		spc,ptp
1208
1209	space_check	spc,t0,dtlb_fault
1210
1211	L2_ptep		ptp,pte,t0,va,dtlb_check_alias_11
1212
1213	tlb_lock	spc,ptp,pte,t0,t1,dtlb_check_alias_11
1214	update_accessed	ptp,pte,t0,t1
1215
1216	make_insert_tlb_11	spc,pte,prot
1217
1218	mfsp		%sr1,t1  /* Save sr1 so we can use it in tlb inserts */
1219	mtsp		spc,%sr1
1220
1221	idtlba		pte,(%sr1,va)
1222	idtlbp		prot,(%sr1,va)
1223
1224	mtsp		t1, %sr1	/* Restore sr1 */
1225
1226	tlb_unlock1	spc,t0
1227	rfir
1228	nop
1229
1230dtlb_check_alias_11:
1231	do_alias	spc,t0,t1,va,pte,prot,dtlb_fault,11
1232
1233	idtlba          pte,(va)
1234	idtlbp          prot,(va)
1235
1236	rfir
1237	nop
1238
1239nadtlb_miss_11:
1240	get_pgd		spc,ptp
1241
1242	space_check	spc,t0,nadtlb_fault
1243
1244	L2_ptep		ptp,pte,t0,va,nadtlb_check_alias_11
1245
1246	tlb_lock	spc,ptp,pte,t0,t1,nadtlb_check_alias_11
1247	update_accessed	ptp,pte,t0,t1
1248
1249	make_insert_tlb_11	spc,pte,prot
1250
1251	mfsp		%sr1,t1  /* Save sr1 so we can use it in tlb inserts */
1252	mtsp		spc,%sr1
1253
1254	idtlba		pte,(%sr1,va)
1255	idtlbp		prot,(%sr1,va)
1256
1257	mtsp		t1, %sr1	/* Restore sr1 */
1258
1259	tlb_unlock1	spc,t0
1260	rfir
1261	nop
1262
1263nadtlb_check_alias_11:
1264	do_alias	spc,t0,t1,va,pte,prot,nadtlb_emulate,11
1265
1266	idtlba          pte,(va)
1267	idtlbp          prot,(va)
1268
1269	rfir
1270	nop
1271
1272dtlb_miss_20:
1273	space_adjust	spc,va,t0
1274	get_pgd		spc,ptp
1275	space_check	spc,t0,dtlb_fault
1276
1277	L2_ptep		ptp,pte,t0,va,dtlb_check_alias_20
1278
1279	tlb_lock	spc,ptp,pte,t0,t1,dtlb_check_alias_20
1280	update_accessed	ptp,pte,t0,t1
1281
1282	make_insert_tlb	spc,pte,prot,t1
1283
1284	f_extend	pte,t1
1285
1286	idtlbt          pte,prot
1287
1288	tlb_unlock1	spc,t0
1289	rfir
1290	nop
1291
1292dtlb_check_alias_20:
1293	do_alias	spc,t0,t1,va,pte,prot,dtlb_fault,20
1294
1295	idtlbt          pte,prot
1296
1297	rfir
1298	nop
1299
1300nadtlb_miss_20:
1301	get_pgd		spc,ptp
1302
1303	space_check	spc,t0,nadtlb_fault
1304
1305	L2_ptep		ptp,pte,t0,va,nadtlb_check_alias_20
1306
1307	tlb_lock	spc,ptp,pte,t0,t1,nadtlb_check_alias_20
1308	update_accessed	ptp,pte,t0,t1
1309
1310	make_insert_tlb	spc,pte,prot,t1
1311
1312	f_extend	pte,t1
1313
1314	idtlbt		pte,prot
1315
1316	tlb_unlock1	spc,t0
1317	rfir
1318	nop
1319
1320nadtlb_check_alias_20:
1321	do_alias	spc,t0,t1,va,pte,prot,nadtlb_emulate,20
1322
1323	idtlbt          pte,prot
1324
1325	rfir
1326	nop
1327
1328#endif
1329
1330nadtlb_emulate:
1331
1332	/*
1333	 * Non access misses can be caused by fdc,fic,pdc,lpa,probe and
1334	 * probei instructions. We don't want to fault for these
1335	 * instructions (not only does it not make sense, it can cause
1336	 * deadlocks, since some flushes are done with the mmap
1337	 * semaphore held). If the translation doesn't exist, we can't
1338	 * insert a translation, so have to emulate the side effects
1339	 * of the instruction. Since we don't insert a translation
1340	 * we can get a lot of faults during a flush loop, so it makes
1341	 * sense to try to do it here with minimum overhead. We only
1342	 * emulate fdc,fic,pdc,probew,prober instructions whose base
1343	 * and index registers are not shadowed. We defer everything
1344	 * else to the "slow" path.
1345	 */
1346
1347	mfctl           %cr19,%r9 /* Get iir */
1348
1349	/* PA 2.0 Arch Ref. Book pg 382 has a good description of the insn bits.
1350	   Checks for fdc,fdce,pdc,"fic,4f",prober,probeir,probew, probeiw */
1351
1352	/* Checks for fdc,fdce,pdc,"fic,4f" only */
1353	ldi             0x280,%r16
1354	and             %r9,%r16,%r17
1355	cmpb,<>,n       %r16,%r17,nadtlb_probe_check
1356	bb,>=,n         %r9,26,nadtlb_nullify  /* m bit not set, just nullify */
1357	BL		get_register,%r25
1358	extrw,u         %r9,15,5,%r8           /* Get index register # */
1359	cmpib,COND(=),n        -1,%r1,nadtlb_fault    /* have to use slow path */
1360	copy            %r1,%r24
1361	BL		get_register,%r25
1362	extrw,u         %r9,10,5,%r8           /* Get base register # */
1363	cmpib,COND(=),n        -1,%r1,nadtlb_fault    /* have to use slow path */
1364	BL		set_register,%r25
1365	add,l           %r1,%r24,%r1           /* doesn't affect c/b bits */
1366
1367nadtlb_nullify:
1368	mfctl           %ipsw,%r8
1369	ldil            L%PSW_N,%r9
1370	or              %r8,%r9,%r8            /* Set PSW_N */
1371	mtctl           %r8,%ipsw
1372
1373	rfir
1374	nop
1375
1376	/*
1377		When there is no translation for the probe address then we
1378		must nullify the insn and return zero in the target regsiter.
1379		This will indicate to the calling code that it does not have
1380		write/read privileges to this address.
1381
1382		This should technically work for prober and probew in PA 1.1,
1383		and also probe,r and probe,w in PA 2.0
1384
1385		WARNING: USE ONLY NON-SHADOW REGISTERS WITH PROBE INSN!
1386		THE SLOW-PATH EMULATION HAS NOT BEEN WRITTEN YET.
1387
1388	*/
1389nadtlb_probe_check:
1390	ldi             0x80,%r16
1391	and             %r9,%r16,%r17
1392	cmpb,<>,n       %r16,%r17,nadtlb_fault /* Must be probe,[rw]*/
1393	BL              get_register,%r25      /* Find the target register */
1394	extrw,u         %r9,31,5,%r8           /* Get target register */
1395	cmpib,COND(=),n        -1,%r1,nadtlb_fault    /* have to use slow path */
1396	BL		set_register,%r25
1397	copy            %r0,%r1                /* Write zero to target register */
1398	b nadtlb_nullify                       /* Nullify return insn */
1399	nop
1400
1401
1402#ifdef CONFIG_64BIT
1403itlb_miss_20w:
1404
1405	/*
1406	 * I miss is a little different, since we allow users to fault
1407	 * on the gateway page which is in the kernel address space.
1408	 */
1409
1410	space_adjust	spc,va,t0
1411	get_pgd		spc,ptp
1412	space_check	spc,t0,itlb_fault
1413
1414	L3_ptep		ptp,pte,t0,va,itlb_fault
1415
1416	tlb_lock	spc,ptp,pte,t0,t1,itlb_fault
1417	update_accessed	ptp,pte,t0,t1
1418
1419	make_insert_tlb	spc,pte,prot,t1
1420
1421	iitlbt          pte,prot
1422
1423	tlb_unlock1	spc,t0
1424	rfir
1425	nop
1426
1427naitlb_miss_20w:
1428
1429	/*
1430	 * I miss is a little different, since we allow users to fault
1431	 * on the gateway page which is in the kernel address space.
1432	 */
1433
1434	space_adjust	spc,va,t0
1435	get_pgd		spc,ptp
1436	space_check	spc,t0,naitlb_fault
1437
1438	L3_ptep		ptp,pte,t0,va,naitlb_check_alias_20w
1439
1440	tlb_lock	spc,ptp,pte,t0,t1,naitlb_check_alias_20w
1441	update_accessed	ptp,pte,t0,t1
1442
1443	make_insert_tlb	spc,pte,prot,t1
1444
1445	iitlbt          pte,prot
1446
1447	tlb_unlock1	spc,t0
1448	rfir
1449	nop
1450
1451naitlb_check_alias_20w:
1452	do_alias	spc,t0,t1,va,pte,prot,naitlb_fault,20
1453
1454	iitlbt		pte,prot
1455
1456	rfir
1457	nop
1458
1459#else
1460
1461itlb_miss_11:
1462	get_pgd		spc,ptp
1463
1464	space_check	spc,t0,itlb_fault
1465
1466	L2_ptep		ptp,pte,t0,va,itlb_fault
1467
1468	tlb_lock	spc,ptp,pte,t0,t1,itlb_fault
1469	update_accessed	ptp,pte,t0,t1
1470
1471	make_insert_tlb_11	spc,pte,prot
1472
1473	mfsp		%sr1,t1  /* Save sr1 so we can use it in tlb inserts */
1474	mtsp		spc,%sr1
1475
1476	iitlba		pte,(%sr1,va)
1477	iitlbp		prot,(%sr1,va)
1478
1479	mtsp		t1, %sr1	/* Restore sr1 */
1480
1481	tlb_unlock1	spc,t0
1482	rfir
1483	nop
1484
1485naitlb_miss_11:
1486	get_pgd		spc,ptp
1487
1488	space_check	spc,t0,naitlb_fault
1489
1490	L2_ptep		ptp,pte,t0,va,naitlb_check_alias_11
1491
1492	tlb_lock	spc,ptp,pte,t0,t1,naitlb_check_alias_11
1493	update_accessed	ptp,pte,t0,t1
1494
1495	make_insert_tlb_11	spc,pte,prot
1496
1497	mfsp		%sr1,t1  /* Save sr1 so we can use it in tlb inserts */
1498	mtsp		spc,%sr1
1499
1500	iitlba		pte,(%sr1,va)
1501	iitlbp		prot,(%sr1,va)
1502
1503	mtsp		t1, %sr1	/* Restore sr1 */
1504
1505	tlb_unlock1	spc,t0
1506	rfir
1507	nop
1508
1509naitlb_check_alias_11:
1510	do_alias	spc,t0,t1,va,pte,prot,itlb_fault,11
1511
1512	iitlba          pte,(%sr0, va)
1513	iitlbp          prot,(%sr0, va)
1514
1515	rfir
1516	nop
1517
1518
1519itlb_miss_20:
1520	get_pgd		spc,ptp
1521
1522	space_check	spc,t0,itlb_fault
1523
1524	L2_ptep		ptp,pte,t0,va,itlb_fault
1525
1526	tlb_lock	spc,ptp,pte,t0,t1,itlb_fault
1527	update_accessed	ptp,pte,t0,t1
1528
1529	make_insert_tlb	spc,pte,prot,t1
1530
1531	f_extend	pte,t1
1532
1533	iitlbt          pte,prot
1534
1535	tlb_unlock1	spc,t0
1536	rfir
1537	nop
1538
1539naitlb_miss_20:
1540	get_pgd		spc,ptp
1541
1542	space_check	spc,t0,naitlb_fault
1543
1544	L2_ptep		ptp,pte,t0,va,naitlb_check_alias_20
1545
1546	tlb_lock	spc,ptp,pte,t0,t1,naitlb_check_alias_20
1547	update_accessed	ptp,pte,t0,t1
1548
1549	make_insert_tlb	spc,pte,prot,t1
1550
1551	f_extend	pte,t1
1552
1553	iitlbt          pte,prot
1554
1555	tlb_unlock1	spc,t0
1556	rfir
1557	nop
1558
1559naitlb_check_alias_20:
1560	do_alias	spc,t0,t1,va,pte,prot,naitlb_fault,20
1561
1562	iitlbt          pte,prot
1563
1564	rfir
1565	nop
1566
1567#endif
1568
1569#ifdef CONFIG_64BIT
1570
1571dbit_trap_20w:
1572	space_adjust	spc,va,t0
1573	get_pgd		spc,ptp
1574	space_check	spc,t0,dbit_fault
1575
1576	L3_ptep		ptp,pte,t0,va,dbit_fault
1577
1578	tlb_lock	spc,ptp,pte,t0,t1,dbit_fault
1579	update_dirty	ptp,pte,t1
1580
1581	make_insert_tlb	spc,pte,prot,t1
1582
1583	idtlbt          pte,prot
1584
1585	tlb_unlock0	spc,t0
1586	rfir
1587	nop
1588#else
1589
1590dbit_trap_11:
1591
1592	get_pgd		spc,ptp
1593
1594	space_check	spc,t0,dbit_fault
1595
1596	L2_ptep		ptp,pte,t0,va,dbit_fault
1597
1598	tlb_lock	spc,ptp,pte,t0,t1,dbit_fault
1599	update_dirty	ptp,pte,t1
1600
1601	make_insert_tlb_11	spc,pte,prot
1602
1603	mfsp            %sr1,t1  /* Save sr1 so we can use it in tlb inserts */
1604	mtsp		spc,%sr1
1605
1606	idtlba		pte,(%sr1,va)
1607	idtlbp		prot,(%sr1,va)
1608
1609	mtsp            t1, %sr1     /* Restore sr1 */
1610
1611	tlb_unlock0	spc,t0
1612	rfir
1613	nop
1614
1615dbit_trap_20:
1616	get_pgd		spc,ptp
1617
1618	space_check	spc,t0,dbit_fault
1619
1620	L2_ptep		ptp,pte,t0,va,dbit_fault
1621
1622	tlb_lock	spc,ptp,pte,t0,t1,dbit_fault
1623	update_dirty	ptp,pte,t1
1624
1625	make_insert_tlb	spc,pte,prot,t1
1626
1627	f_extend	pte,t1
1628
1629	idtlbt		pte,prot
1630
1631	tlb_unlock0	spc,t0
1632	rfir
1633	nop
1634#endif
1635
1636	.import handle_interruption,code
1637
1638kernel_bad_space:
1639	b               intr_save
1640	ldi             31,%r8  /* Use an unused code */
1641
1642dbit_fault:
1643	b               intr_save
1644	ldi             20,%r8
1645
1646itlb_fault:
1647	b               intr_save
1648	ldi             6,%r8
1649
1650nadtlb_fault:
1651	b               intr_save
1652	ldi             17,%r8
1653
1654naitlb_fault:
1655	b               intr_save
1656	ldi             16,%r8
1657
1658dtlb_fault:
1659	b               intr_save
1660	ldi             15,%r8
1661
1662	/* Register saving semantics for system calls:
1663
1664	   %r1		   clobbered by system call macro in userspace
1665	   %r2		   saved in PT_REGS by gateway page
1666	   %r3  - %r18	   preserved by C code (saved by signal code)
1667	   %r19 - %r20	   saved in PT_REGS by gateway page
1668	   %r21 - %r22	   non-standard syscall args
1669			   stored in kernel stack by gateway page
1670	   %r23 - %r26	   arg3-arg0, saved in PT_REGS by gateway page
1671	   %r27 - %r30	   saved in PT_REGS by gateway page
1672	   %r31		   syscall return pointer
1673	 */
1674
1675	/* Floating point registers (FIXME: what do we do with these?)
1676
1677	   %fr0  - %fr3	   status/exception, not preserved
1678	   %fr4  - %fr7	   arguments
1679	   %fr8	 - %fr11   not preserved by C code
1680	   %fr12 - %fr21   preserved by C code
1681	   %fr22 - %fr31   not preserved by C code
1682	 */
1683
1684	.macro	reg_save regs
1685	STREG	%r3, PT_GR3(\regs)
1686	STREG	%r4, PT_GR4(\regs)
1687	STREG	%r5, PT_GR5(\regs)
1688	STREG	%r6, PT_GR6(\regs)
1689	STREG	%r7, PT_GR7(\regs)
1690	STREG	%r8, PT_GR8(\regs)
1691	STREG	%r9, PT_GR9(\regs)
1692	STREG   %r10,PT_GR10(\regs)
1693	STREG   %r11,PT_GR11(\regs)
1694	STREG   %r12,PT_GR12(\regs)
1695	STREG   %r13,PT_GR13(\regs)
1696	STREG   %r14,PT_GR14(\regs)
1697	STREG   %r15,PT_GR15(\regs)
1698	STREG   %r16,PT_GR16(\regs)
1699	STREG   %r17,PT_GR17(\regs)
1700	STREG   %r18,PT_GR18(\regs)
1701	.endm
1702
1703	.macro	reg_restore regs
1704	LDREG	PT_GR3(\regs), %r3
1705	LDREG	PT_GR4(\regs), %r4
1706	LDREG	PT_GR5(\regs), %r5
1707	LDREG	PT_GR6(\regs), %r6
1708	LDREG	PT_GR7(\regs), %r7
1709	LDREG	PT_GR8(\regs), %r8
1710	LDREG	PT_GR9(\regs), %r9
1711	LDREG   PT_GR10(\regs),%r10
1712	LDREG   PT_GR11(\regs),%r11
1713	LDREG   PT_GR12(\regs),%r12
1714	LDREG   PT_GR13(\regs),%r13
1715	LDREG   PT_GR14(\regs),%r14
1716	LDREG   PT_GR15(\regs),%r15
1717	LDREG   PT_GR16(\regs),%r16
1718	LDREG   PT_GR17(\regs),%r17
1719	LDREG   PT_GR18(\regs),%r18
1720	.endm
1721
1722	.macro	fork_like name
1723ENTRY(sys_\name\()_wrapper)
1724	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
1725	ldo	TASK_REGS(%r1),%r1
1726	reg_save %r1
1727	mfctl	%cr27, %r28
1728	ldil	L%sys_\name, %r31
1729	be	R%sys_\name(%sr4,%r31)
1730	STREG	%r28, PT_CR27(%r1)
1731ENDPROC(sys_\name\()_wrapper)
1732	.endm
1733
1734fork_like clone
1735fork_like fork
1736fork_like vfork
1737
1738	/* Set the return value for the child */
1739ENTRY(child_return)
1740	BL	schedule_tail, %r2
1741	nop
1742finish_child_return:
1743	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30), %r1
1744	ldo	TASK_REGS(%r1),%r1	 /* get pt regs */
1745
1746	LDREG	PT_CR27(%r1), %r3
1747	mtctl	%r3, %cr27
1748	reg_restore %r1
1749	b	syscall_exit
1750	copy	%r0,%r28
1751ENDPROC(child_return)
1752
1753ENTRY(sys_rt_sigreturn_wrapper)
1754	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r26
1755	ldo	TASK_REGS(%r26),%r26	/* get pt regs */
1756	/* Don't save regs, we are going to restore them from sigcontext. */
1757	STREG	%r2, -RP_OFFSET(%r30)
1758#ifdef CONFIG_64BIT
1759	ldo	FRAME_SIZE(%r30), %r30
1760	BL	sys_rt_sigreturn,%r2
1761	ldo	-16(%r30),%r29		/* Reference param save area */
1762#else
1763	BL	sys_rt_sigreturn,%r2
1764	ldo	FRAME_SIZE(%r30), %r30
1765#endif
1766
1767	ldo	-FRAME_SIZE(%r30), %r30
1768	LDREG	-RP_OFFSET(%r30), %r2
1769
1770	/* FIXME: I think we need to restore a few more things here. */
1771	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1772	ldo	TASK_REGS(%r1),%r1	/* get pt regs */
1773	reg_restore %r1
1774
1775	/* If the signal was received while the process was blocked on a
1776	 * syscall, then r2 will take us to syscall_exit; otherwise r2 will
1777	 * take us to syscall_exit_rfi and on to intr_return.
1778	 */
1779	bv	%r0(%r2)
1780	LDREG	PT_GR28(%r1),%r28  /* reload original r28 for syscall_exit */
1781ENDPROC(sys_rt_sigreturn_wrapper)
1782
1783ENTRY(syscall_exit)
1784	/* NOTE: Not all syscalls exit this way.  rt_sigreturn will exit
1785	 * via syscall_exit_rfi if the signal was received while the process
1786	 * was running.
1787	 */
1788
1789	/* save return value now */
1790
1791	mfctl     %cr30, %r1
1792	LDREG     TI_TASK(%r1),%r1
1793	STREG     %r28,TASK_PT_GR28(%r1)
1794
1795	/* Seems to me that dp could be wrong here, if the syscall involved
1796	 * calling a module, and nothing got round to restoring dp on return.
1797	 */
1798	loadgp
1799
1800syscall_check_resched:
1801
1802	/* check for reschedule */
1803
1804	LDREG	TI_FLAGS-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r19	/* long */
1805	bb,<,n	%r19, 31-TIF_NEED_RESCHED, syscall_do_resched /* forward */
1806
1807	.import do_signal,code
1808syscall_check_sig:
1809	LDREG	TI_FLAGS-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r19
1810	ldi	(_TIF_SIGPENDING|_TIF_NOTIFY_RESUME), %r26
1811	and,COND(<>)	%r19, %r26, %r0
1812	b,n	syscall_restore	/* skip past if we've nothing to do */
1813
1814syscall_do_signal:
1815	/* Save callee-save registers (for sigcontext).
1816	 * FIXME: After this point the process structure should be
1817	 * consistent with all the relevant state of the process
1818	 * before the syscall.  We need to verify this.
1819	 */
1820	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1821	ldo	TASK_REGS(%r1), %r26		/* struct pt_regs *regs */
1822	reg_save %r26
1823
1824#ifdef CONFIG_64BIT
1825	ldo	-16(%r30),%r29			/* Reference param save area */
1826#endif
1827
1828	BL	do_notify_resume,%r2
1829	ldi	1, %r25				/* long in_syscall = 1 */
1830
1831	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1832	ldo	TASK_REGS(%r1), %r20		/* reload pt_regs */
1833	reg_restore %r20
1834
1835	b,n     syscall_check_sig
1836
1837syscall_restore:
1838	LDREG	TI_TASK-THREAD_SZ_ALGN-FRAME_SIZE(%r30),%r1
1839
1840	/* Are we being ptraced? */
1841	ldw	TASK_FLAGS(%r1),%r19
1842	ldi	_TIF_SYSCALL_TRACE_MASK,%r2
1843	and,COND(=)	%r19,%r2,%r0
1844	b,n	syscall_restore_rfi
1845
1846	ldo	TASK_PT_FR31(%r1),%r19		   /* reload fpregs */
1847	rest_fp	%r19
1848
1849	LDREG	TASK_PT_SAR(%r1),%r19		   /* restore SAR */
1850	mtsar	%r19
1851
1852	LDREG	TASK_PT_GR2(%r1),%r2		   /* restore user rp */
1853	LDREG	TASK_PT_GR19(%r1),%r19
1854	LDREG   TASK_PT_GR20(%r1),%r20
1855	LDREG	TASK_PT_GR21(%r1),%r21
1856	LDREG	TASK_PT_GR22(%r1),%r22
1857	LDREG	TASK_PT_GR23(%r1),%r23
1858	LDREG	TASK_PT_GR24(%r1),%r24
1859	LDREG	TASK_PT_GR25(%r1),%r25
1860	LDREG	TASK_PT_GR26(%r1),%r26
1861	LDREG	TASK_PT_GR27(%r1),%r27	   /* restore user dp */
1862	LDREG	TASK_PT_GR28(%r1),%r28	   /* syscall return value */
1863	LDREG	TASK_PT_GR29(%r1),%r29
1864	LDREG	TASK_PT_GR31(%r1),%r31	   /* restore syscall rp */
1865
1866	/* NOTE: We use rsm/ssm pair to make this operation atomic */
1867	LDREG   TASK_PT_GR30(%r1),%r1              /* Get user sp */
1868	rsm     PSW_SM_I, %r0
1869	copy    %r1,%r30                           /* Restore user sp */
1870	mfsp    %sr3,%r1                           /* Get user space id */
1871	mtsp    %r1,%sr7                           /* Restore sr7 */
1872	ssm     PSW_SM_I, %r0
1873
1874	/* Set sr2 to zero for userspace syscalls to work. */
1875	mtsp	%r0,%sr2
1876	mtsp	%r1,%sr4			   /* Restore sr4 */
1877	mtsp	%r1,%sr5			   /* Restore sr5 */
1878	mtsp	%r1,%sr6			   /* Restore sr6 */
1879
1880	depi	3,31,2,%r31			   /* ensure return to user mode. */
1881
1882#ifdef CONFIG_64BIT
1883	/* decide whether to reset the wide mode bit
1884	 *
1885	 * For a syscall, the W bit is stored in the lowest bit
1886	 * of sp.  Extract it and reset W if it is zero */
1887	extrd,u,*<>	%r30,63,1,%r1
1888	rsm	PSW_SM_W, %r0
1889	/* now reset the lowest bit of sp if it was set */
1890	xor	%r30,%r1,%r30
1891#endif
1892	be,n    0(%sr3,%r31)                       /* return to user space */
1893
1894	/* We have to return via an RFI, so that PSW T and R bits can be set
1895	 * appropriately.
1896	 * This sets up pt_regs so we can return via intr_restore, which is not
1897	 * the most efficient way of doing things, but it works.
1898	 */
1899syscall_restore_rfi:
1900	ldo	-1(%r0),%r2			   /* Set recovery cntr to -1 */
1901	mtctl	%r2,%cr0			   /*   for immediate trap */
1902	LDREG	TASK_PT_PSW(%r1),%r2		   /* Get old PSW */
1903	ldi	0x0b,%r20			   /* Create new PSW */
1904	depi	-1,13,1,%r20			   /* C, Q, D, and I bits */
1905
1906	/* The values of SINGLESTEP_BIT and BLOCKSTEP_BIT are
1907	 * set in thread_info.h and converted to PA bitmap
1908	 * numbers in asm-offsets.c */
1909
1910	/* if ((%r19.SINGLESTEP_BIT)) { %r20.27=1} */
1911	extru,=	%r19,TIF_SINGLESTEP_PA_BIT,1,%r0
1912	depi	-1,27,1,%r20			   /* R bit */
1913
1914	/* if ((%r19.BLOCKSTEP_BIT)) { %r20.7=1} */
1915	extru,= %r19,TIF_BLOCKSTEP_PA_BIT,1,%r0
1916	depi	-1,7,1,%r20			   /* T bit */
1917
1918	STREG	%r20,TASK_PT_PSW(%r1)
1919
1920	/* Always store space registers, since sr3 can be changed (e.g. fork) */
1921
1922	mfsp    %sr3,%r25
1923	STREG   %r25,TASK_PT_SR3(%r1)
1924	STREG   %r25,TASK_PT_SR4(%r1)
1925	STREG   %r25,TASK_PT_SR5(%r1)
1926	STREG   %r25,TASK_PT_SR6(%r1)
1927	STREG   %r25,TASK_PT_SR7(%r1)
1928	STREG   %r25,TASK_PT_IASQ0(%r1)
1929	STREG   %r25,TASK_PT_IASQ1(%r1)
1930
1931	/* XXX W bit??? */
1932	/* Now if old D bit is clear, it means we didn't save all registers
1933	 * on syscall entry, so do that now.  This only happens on TRACEME
1934	 * calls, or if someone attached to us while we were on a syscall.
1935	 * We could make this more efficient by not saving r3-r18, but
1936	 * then we wouldn't be able to use the common intr_restore path.
1937	 * It is only for traced processes anyway, so performance is not
1938	 * an issue.
1939	 */
1940	bb,<	%r2,30,pt_regs_ok		   /* Branch if D set */
1941	ldo	TASK_REGS(%r1),%r25
1942	reg_save %r25				   /* Save r3 to r18 */
1943
1944	/* Save the current sr */
1945	mfsp	%sr0,%r2
1946	STREG	%r2,TASK_PT_SR0(%r1)
1947
1948	/* Save the scratch sr */
1949	mfsp	%sr1,%r2
1950	STREG	%r2,TASK_PT_SR1(%r1)
1951
1952	/* sr2 should be set to zero for userspace syscalls */
1953	STREG	%r0,TASK_PT_SR2(%r1)
1954
1955	LDREG	TASK_PT_GR31(%r1),%r2
1956	depi	3,31,2,%r2		   /* ensure return to user mode. */
1957	STREG   %r2,TASK_PT_IAOQ0(%r1)
1958	ldo	4(%r2),%r2
1959	STREG	%r2,TASK_PT_IAOQ1(%r1)
1960	b	intr_restore
1961	copy	%r25,%r16
1962
1963pt_regs_ok:
1964	LDREG	TASK_PT_IAOQ0(%r1),%r2
1965	depi	3,31,2,%r2		   /* ensure return to user mode. */
1966	STREG	%r2,TASK_PT_IAOQ0(%r1)
1967	LDREG	TASK_PT_IAOQ1(%r1),%r2
1968	depi	3,31,2,%r2
1969	STREG	%r2,TASK_PT_IAOQ1(%r1)
1970	b	intr_restore
1971	copy	%r25,%r16
1972
1973syscall_do_resched:
1974	load32	syscall_check_resched,%r2 /* if resched, we start over again */
1975	load32	schedule,%r19
1976	bv	%r0(%r19)		/* jumps to schedule() */
1977#ifdef CONFIG_64BIT
1978	ldo	-16(%r30),%r29		/* Reference param save area */
1979#else
1980	nop
1981#endif
1982ENDPROC(syscall_exit)
1983
1984
1985#ifdef CONFIG_FUNCTION_TRACER
1986
1987	.import ftrace_function_trampoline,code
1988	.align L1_CACHE_BYTES
1989	.globl mcount
1990	.type  mcount, @function
1991ENTRY(mcount)
1992_mcount:
1993	.export _mcount,data
1994	.proc
1995	.callinfo caller,frame=0
1996	.entry
1997	/*
1998	 * The 64bit mcount() function pointer needs 4 dwords, of which the
1999	 * first two are free.  We optimize it here and put 2 instructions for
2000	 * calling mcount(), and 2 instructions for ftrace_stub().  That way we
2001	 * have all on one L1 cacheline.
2002	 */
2003	b	ftrace_function_trampoline
2004	copy	%r3, %arg2	/* caller original %sp */
2005ftrace_stub:
2006	.globl ftrace_stub
2007        .type  ftrace_stub, @function
2008#ifdef CONFIG_64BIT
2009	bve	(%rp)
2010#else
2011	bv	%r0(%rp)
2012#endif
2013	nop
2014#ifdef CONFIG_64BIT
2015	.dword mcount
2016	.dword 0 /* code in head.S puts value of global gp here */
2017#endif
2018	.exit
2019	.procend
2020ENDPROC(mcount)
2021
2022#ifdef CONFIG_FUNCTION_GRAPH_TRACER
2023	.align 8
2024	.globl return_to_handler
2025	.type  return_to_handler, @function
2026ENTRY(return_to_handler)
2027	.proc
2028	.callinfo caller,frame=FRAME_SIZE
2029	.entry
2030	.export parisc_return_to_handler,data
2031parisc_return_to_handler:
2032	copy %r3,%r1
2033	STREG %r0,-RP_OFFSET(%sp)	/* store 0 as %rp */
2034	copy %sp,%r3
2035	STREGM %r1,FRAME_SIZE(%sp)
2036	STREG %ret0,8(%r3)
2037	STREG %ret1,16(%r3)
2038
2039#ifdef CONFIG_64BIT
2040	loadgp
2041#endif
2042
2043	/* call ftrace_return_to_handler(0) */
2044	.import ftrace_return_to_handler,code
2045	load32 ftrace_return_to_handler,%ret0
2046	load32 .Lftrace_ret,%r2
2047#ifdef CONFIG_64BIT
2048	ldo -16(%sp),%ret1		/* Reference param save area */
2049	bve	(%ret0)
2050#else
2051	bv	%r0(%ret0)
2052#endif
2053	ldi 0,%r26
2054.Lftrace_ret:
2055	copy %ret0,%rp
2056
2057	/* restore original return values */
2058	LDREG 8(%r3),%ret0
2059	LDREG 16(%r3),%ret1
2060
2061	/* return from function */
2062#ifdef CONFIG_64BIT
2063	bve	(%rp)
2064#else
2065	bv	%r0(%rp)
2066#endif
2067	LDREGM -FRAME_SIZE(%sp),%r3
2068	.exit
2069	.procend
2070ENDPROC(return_to_handler)
2071
2072#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
2073
2074#endif	/* CONFIG_FUNCTION_TRACER */
2075
2076#ifdef CONFIG_IRQSTACKS
2077/* void call_on_stack(unsigned long param1, void *func,
2078		      unsigned long new_stack) */
2079ENTRY(call_on_stack)
2080	copy	%sp, %r1
2081
2082	/* Regarding the HPPA calling conventions for function pointers,
2083	   we assume the PIC register is not changed across call.  For
2084	   CONFIG_64BIT, the argument pointer is left to point at the
2085	   argument region allocated for the call to call_on_stack. */
2086# ifdef CONFIG_64BIT
2087	/* Switch to new stack.  We allocate two 128 byte frames.  */
2088	ldo	256(%arg2), %sp
2089	/* Save previous stack pointer and return pointer in frame marker */
2090	STREG	%rp, -144(%sp)
2091	/* Calls always use function descriptor */
2092	LDREG	16(%arg1), %arg1
2093	bve,l	(%arg1), %rp
2094	STREG	%r1, -136(%sp)
2095	LDREG	-144(%sp), %rp
2096	bve	(%rp)
2097	LDREG	-136(%sp), %sp
2098# else
2099	/* Switch to new stack.  We allocate two 64 byte frames.  */
2100	ldo	128(%arg2), %sp
2101	/* Save previous stack pointer and return pointer in frame marker */
2102	STREG	%r1, -68(%sp)
2103	STREG	%rp, -84(%sp)
2104	/* Calls use function descriptor if PLABEL bit is set */
2105	bb,>=,n	%arg1, 30, 1f
2106	depwi	0,31,2, %arg1
2107	LDREG	0(%arg1), %arg1
21081:
2109	be,l	0(%sr4,%arg1), %sr0, %r31
2110	copy	%r31, %rp
2111	LDREG	-84(%sp), %rp
2112	bv	(%rp)
2113	LDREG	-68(%sp), %sp
2114# endif /* CONFIG_64BIT */
2115ENDPROC(call_on_stack)
2116#endif /* CONFIG_IRQSTACKS */
2117
2118get_register:
2119	/*
2120	 * get_register is used by the non access tlb miss handlers to
2121	 * copy the value of the general register specified in r8 into
2122	 * r1. This routine can't be used for shadowed registers, since
2123	 * the rfir will restore the original value. So, for the shadowed
2124	 * registers we put a -1 into r1 to indicate that the register
2125	 * should not be used (the register being copied could also have
2126	 * a -1 in it, but that is OK, it just means that we will have
2127	 * to use the slow path instead).
2128	 */
2129	blr     %r8,%r0
2130	nop
2131	bv      %r0(%r25)    /* r0 */
2132	copy    %r0,%r1
2133	bv      %r0(%r25)    /* r1 - shadowed */
2134	ldi     -1,%r1
2135	bv      %r0(%r25)    /* r2 */
2136	copy    %r2,%r1
2137	bv      %r0(%r25)    /* r3 */
2138	copy    %r3,%r1
2139	bv      %r0(%r25)    /* r4 */
2140	copy    %r4,%r1
2141	bv      %r0(%r25)    /* r5 */
2142	copy    %r5,%r1
2143	bv      %r0(%r25)    /* r6 */
2144	copy    %r6,%r1
2145	bv      %r0(%r25)    /* r7 */
2146	copy    %r7,%r1
2147	bv      %r0(%r25)    /* r8 - shadowed */
2148	ldi     -1,%r1
2149	bv      %r0(%r25)    /* r9 - shadowed */
2150	ldi     -1,%r1
2151	bv      %r0(%r25)    /* r10 */
2152	copy    %r10,%r1
2153	bv      %r0(%r25)    /* r11 */
2154	copy    %r11,%r1
2155	bv      %r0(%r25)    /* r12 */
2156	copy    %r12,%r1
2157	bv      %r0(%r25)    /* r13 */
2158	copy    %r13,%r1
2159	bv      %r0(%r25)    /* r14 */
2160	copy    %r14,%r1
2161	bv      %r0(%r25)    /* r15 */
2162	copy    %r15,%r1
2163	bv      %r0(%r25)    /* r16 - shadowed */
2164	ldi     -1,%r1
2165	bv      %r0(%r25)    /* r17 - shadowed */
2166	ldi     -1,%r1
2167	bv      %r0(%r25)    /* r18 */
2168	copy    %r18,%r1
2169	bv      %r0(%r25)    /* r19 */
2170	copy    %r19,%r1
2171	bv      %r0(%r25)    /* r20 */
2172	copy    %r20,%r1
2173	bv      %r0(%r25)    /* r21 */
2174	copy    %r21,%r1
2175	bv      %r0(%r25)    /* r22 */
2176	copy    %r22,%r1
2177	bv      %r0(%r25)    /* r23 */
2178	copy    %r23,%r1
2179	bv      %r0(%r25)    /* r24 - shadowed */
2180	ldi     -1,%r1
2181	bv      %r0(%r25)    /* r25 - shadowed */
2182	ldi     -1,%r1
2183	bv      %r0(%r25)    /* r26 */
2184	copy    %r26,%r1
2185	bv      %r0(%r25)    /* r27 */
2186	copy    %r27,%r1
2187	bv      %r0(%r25)    /* r28 */
2188	copy    %r28,%r1
2189	bv      %r0(%r25)    /* r29 */
2190	copy    %r29,%r1
2191	bv      %r0(%r25)    /* r30 */
2192	copy    %r30,%r1
2193	bv      %r0(%r25)    /* r31 */
2194	copy    %r31,%r1
2195
2196
2197set_register:
2198	/*
2199	 * set_register is used by the non access tlb miss handlers to
2200	 * copy the value of r1 into the general register specified in
2201	 * r8.
2202	 */
2203	blr     %r8,%r0
2204	nop
2205	bv      %r0(%r25)    /* r0 (silly, but it is a place holder) */
2206	copy    %r1,%r0
2207	bv      %r0(%r25)    /* r1 */
2208	copy    %r1,%r1
2209	bv      %r0(%r25)    /* r2 */
2210	copy    %r1,%r2
2211	bv      %r0(%r25)    /* r3 */
2212	copy    %r1,%r3
2213	bv      %r0(%r25)    /* r4 */
2214	copy    %r1,%r4
2215	bv      %r0(%r25)    /* r5 */
2216	copy    %r1,%r5
2217	bv      %r0(%r25)    /* r6 */
2218	copy    %r1,%r6
2219	bv      %r0(%r25)    /* r7 */
2220	copy    %r1,%r7
2221	bv      %r0(%r25)    /* r8 */
2222	copy    %r1,%r8
2223	bv      %r0(%r25)    /* r9 */
2224	copy    %r1,%r9
2225	bv      %r0(%r25)    /* r10 */
2226	copy    %r1,%r10
2227	bv      %r0(%r25)    /* r11 */
2228	copy    %r1,%r11
2229	bv      %r0(%r25)    /* r12 */
2230	copy    %r1,%r12
2231	bv      %r0(%r25)    /* r13 */
2232	copy    %r1,%r13
2233	bv      %r0(%r25)    /* r14 */
2234	copy    %r1,%r14
2235	bv      %r0(%r25)    /* r15 */
2236	copy    %r1,%r15
2237	bv      %r0(%r25)    /* r16 */
2238	copy    %r1,%r16
2239	bv      %r0(%r25)    /* r17 */
2240	copy    %r1,%r17
2241	bv      %r0(%r25)    /* r18 */
2242	copy    %r1,%r18
2243	bv      %r0(%r25)    /* r19 */
2244	copy    %r1,%r19
2245	bv      %r0(%r25)    /* r20 */
2246	copy    %r1,%r20
2247	bv      %r0(%r25)    /* r21 */
2248	copy    %r1,%r21
2249	bv      %r0(%r25)    /* r22 */
2250	copy    %r1,%r22
2251	bv      %r0(%r25)    /* r23 */
2252	copy    %r1,%r23
2253	bv      %r0(%r25)    /* r24 */
2254	copy    %r1,%r24
2255	bv      %r0(%r25)    /* r25 */
2256	copy    %r1,%r25
2257	bv      %r0(%r25)    /* r26 */
2258	copy    %r1,%r26
2259	bv      %r0(%r25)    /* r27 */
2260	copy    %r1,%r27
2261	bv      %r0(%r25)    /* r28 */
2262	copy    %r1,%r28
2263	bv      %r0(%r25)    /* r29 */
2264	copy    %r1,%r29
2265	bv      %r0(%r25)    /* r30 */
2266	copy    %r1,%r30
2267	bv      %r0(%r25)    /* r31 */
2268	copy    %r1,%r31
2269
2270