xref: /openbmc/linux/arch/xtensa/kernel/align.S (revision a83b02e9bd0c28d27b6c6e5b184585f7a1b8bf86)
15a0015d6SChris Zankel/*
25a0015d6SChris Zankel * arch/xtensa/kernel/align.S
35a0015d6SChris Zankel *
45a0015d6SChris Zankel * Handle unalignment exceptions in kernel space.
55a0015d6SChris Zankel *
65a0015d6SChris Zankel * This file is subject to the terms and conditions of the GNU General
75a0015d6SChris Zankel * Public License.  See the file "COPYING" in the main directory of
85a0015d6SChris Zankel * this archive for more details.
95a0015d6SChris Zankel *
105a0015d6SChris Zankel * Copyright (C) 2001 - 2005 Tensilica, Inc.
11c3ef1f4dSMax Filippov * Copyright (C) 2014 Cadence Design Systems Inc.
125a0015d6SChris Zankel *
135a0015d6SChris Zankel * Rewritten by Chris Zankel <chris@zankel.net>
145a0015d6SChris Zankel *
155a0015d6SChris Zankel * Based on work from Joe Taylor <joe@tensilica.com, joetylr@yahoo.com>
165a0015d6SChris Zankel * and Marc Gauthier <marc@tensilica.com, marc@alimni.uwaterloo.ca>
175a0015d6SChris Zankel */
185a0015d6SChris Zankel
195a0015d6SChris Zankel#include <linux/linkage.h>
205a0015d6SChris Zankel#include <asm/current.h>
210013a854SSam Ravnborg#include <asm/asm-offsets.h>
225a0015d6SChris Zankel#include <asm/processor.h>
235a0015d6SChris Zankel
245a0015d6SChris Zankel#if XCHAL_UNALIGNED_LOAD_EXCEPTION || XCHAL_UNALIGNED_STORE_EXCEPTION
255a0015d6SChris Zankel
265a0015d6SChris Zankel/*  First-level exception handler for unaligned exceptions.
275a0015d6SChris Zankel *
285a0015d6SChris Zankel *  Note: This handler works only for kernel exceptions.  Unaligned user
295a0015d6SChris Zankel *        access should get a seg fault.
305a0015d6SChris Zankel */
315a0015d6SChris Zankel
325a0015d6SChris Zankel/* Big and little endian 16-bit values are located in
335a0015d6SChris Zankel * different halves of a register.  HWORD_START helps to
345a0015d6SChris Zankel * abstract the notion of extracting a 16-bit value from a
355a0015d6SChris Zankel * register.
365a0015d6SChris Zankel * We also have to define new shifting instructions because
375a0015d6SChris Zankel * lsb and msb are on 'opposite' ends in a register for
385a0015d6SChris Zankel * different endian machines.
395a0015d6SChris Zankel *
405a0015d6SChris Zankel * Assume a memory region in ascending address:
415a0015d6SChris Zankel *   	0 1 2 3|4 5 6 7
425a0015d6SChris Zankel *
435a0015d6SChris Zankel * When loading one word into a register, the content of that register is:
445a0015d6SChris Zankel *  LE	3 2 1 0, 7 6 5 4
455a0015d6SChris Zankel *  BE  0 1 2 3, 4 5 6 7
465a0015d6SChris Zankel *
475a0015d6SChris Zankel * Masking the bits of the higher/lower address means:
485a0015d6SChris Zankel *  LE  X X 0 0, 0 0 X X
495a0015d6SChris Zankel *  BE	0 0 X X, X X 0 0
505a0015d6SChris Zankel *
515a0015d6SChris Zankel * Shifting to higher/lower addresses, means:
525a0015d6SChris Zankel *  LE  shift left / shift right
535a0015d6SChris Zankel *  BE  shift right / shift left
545a0015d6SChris Zankel *
555a0015d6SChris Zankel * Extracting 16 bits from a 32 bit reg. value to higher/lower address means:
565a0015d6SChris Zankel *  LE  mask 0 0 X X / shift left
575a0015d6SChris Zankel *  BE  shift left / mask 0 0 X X
585a0015d6SChris Zankel */
595a0015d6SChris Zankel
605a0015d6SChris Zankel#define UNALIGNED_USER_EXCEPTION
615a0015d6SChris Zankel
625a0015d6SChris Zankel#if XCHAL_HAVE_BE
635a0015d6SChris Zankel
645a0015d6SChris Zankel#define HWORD_START	16
655a0015d6SChris Zankel#define	INSN_OP0	28
665a0015d6SChris Zankel#define	INSN_T		24
675a0015d6SChris Zankel#define	INSN_OP1	16
685a0015d6SChris Zankel
695a0015d6SChris Zankel.macro __src_b	r, w0, w1;	src	\r, \w0, \w1;	.endm
705a0015d6SChris Zankel.macro __ssa8	r;		ssa8b	\r;		.endm
715a0015d6SChris Zankel.macro __ssa8r	r;		ssa8l	\r;		.endm
725a0015d6SChris Zankel.macro __sh	r, s;		srl	\r, \s;		.endm
735a0015d6SChris Zankel.macro __sl	r, s;		sll	\r, \s;		.endm
745a0015d6SChris Zankel.macro __exth	r, s;		extui	\r, \s, 0, 16;	.endm
755a0015d6SChris Zankel.macro __extl	r, s;		slli	\r, \s, 16;	.endm
765a0015d6SChris Zankel
775a0015d6SChris Zankel#else
785a0015d6SChris Zankel
795a0015d6SChris Zankel#define HWORD_START	0
805a0015d6SChris Zankel#define	INSN_OP0	0
815a0015d6SChris Zankel#define	INSN_T		4
825a0015d6SChris Zankel#define	INSN_OP1	12
835a0015d6SChris Zankel
845a0015d6SChris Zankel.macro __src_b	r, w0, w1;	src	\r, \w1, \w0;	.endm
855a0015d6SChris Zankel.macro __ssa8	r;		ssa8l	\r;		.endm
865a0015d6SChris Zankel.macro __ssa8r	r;		ssa8b	\r;		.endm
875a0015d6SChris Zankel.macro __sh	r, s;		sll	\r, \s;		.endm
885a0015d6SChris Zankel.macro __sl	r, s;		srl	\r, \s;		.endm
895a0015d6SChris Zankel.macro __exth	r, s;		slli	\r, \s, 16;	.endm
905a0015d6SChris Zankel.macro __extl	r, s;		extui	\r, \s, 0, 16;	.endm
915a0015d6SChris Zankel
925a0015d6SChris Zankel#endif
935a0015d6SChris Zankel
945a0015d6SChris Zankel/*
955a0015d6SChris Zankel *	xxxx xxxx = imm8 field
965a0015d6SChris Zankel *	     yyyy = imm4 field
975a0015d6SChris Zankel *	     ssss = s field
985a0015d6SChris Zankel *	     tttt = t field
995a0015d6SChris Zankel *
1005a0015d6SChris Zankel *	       		 16		    0
1015a0015d6SChris Zankel *		          -------------------
1025a0015d6SChris Zankel *	L32I.N		  yyyy ssss tttt 1000
1035a0015d6SChris Zankel *	S32I.N	          yyyy ssss tttt 1001
1045a0015d6SChris Zankel *
1055a0015d6SChris Zankel *	       23			    0
1065a0015d6SChris Zankel *		-----------------------------
1075a0015d6SChris Zankel *	res	          0000           0010
1085a0015d6SChris Zankel *	L16UI	xxxx xxxx 0001 ssss tttt 0010
1095a0015d6SChris Zankel *	L32I	xxxx xxxx 0010 ssss tttt 0010
1105a0015d6SChris Zankel *	XXX	          0011 ssss tttt 0010
1115a0015d6SChris Zankel *	XXX	          0100 ssss tttt 0010
1125a0015d6SChris Zankel *	S16I	xxxx xxxx 0101 ssss tttt 0010
1135a0015d6SChris Zankel *	S32I	xxxx xxxx 0110 ssss tttt 0010
1145a0015d6SChris Zankel *	XXX	          0111 ssss tttt 0010
1155a0015d6SChris Zankel *	XXX	          1000 ssss tttt 0010
1165a0015d6SChris Zankel *	L16SI	xxxx xxxx 1001 ssss tttt 0010
1175a0015d6SChris Zankel *	XXX	          1010           0010
1185a0015d6SChris Zankel *      **L32AI	xxxx xxxx 1011 ssss tttt 0010 unsupported
1195a0015d6SChris Zankel *	XXX	          1100           0010
1205a0015d6SChris Zankel *	XXX	          1101           0010
1215a0015d6SChris Zankel *	XXX	          1110           0010
1225a0015d6SChris Zankel *	**S32RI	xxxx xxxx 1111 ssss tttt 0010 unsupported
1235a0015d6SChris Zankel *		-----------------------------
1245a0015d6SChris Zankel *                           ^         ^    ^
1255a0015d6SChris Zankel *    sub-opcode (NIBBLE_R) -+         |    |
1265a0015d6SChris Zankel *       t field (NIBBLE_T) -----------+    |
1275a0015d6SChris Zankel *  major opcode (NIBBLE_OP0) --------------+
1285a0015d6SChris Zankel */
1295a0015d6SChris Zankel
1305a0015d6SChris Zankel#define OP0_L32I_N	0x8		/* load immediate narrow */
1315a0015d6SChris Zankel#define OP0_S32I_N	0x9		/* store immediate narrow */
1325a0015d6SChris Zankel#define OP1_SI_MASK	0x4		/* OP1 bit set for stores */
1335a0015d6SChris Zankel#define OP1_SI_BIT	2		/* OP1 bit number for stores */
1345a0015d6SChris Zankel
1355a0015d6SChris Zankel#define OP1_L32I	0x2
1365a0015d6SChris Zankel#define OP1_L16UI	0x1
1375a0015d6SChris Zankel#define OP1_L16SI	0x9
1385a0015d6SChris Zankel#define OP1_L32AI	0xb
1395a0015d6SChris Zankel
1405a0015d6SChris Zankel#define OP1_S32I	0x6
1415a0015d6SChris Zankel#define OP1_S16I	0x5
1425a0015d6SChris Zankel#define OP1_S32RI	0xf
1435a0015d6SChris Zankel
1445a0015d6SChris Zankel/*
1455a0015d6SChris Zankel * Entry condition:
1465a0015d6SChris Zankel *
1475a0015d6SChris Zankel *   a0:	trashed, original value saved on stack (PT_AREG0)
1485a0015d6SChris Zankel *   a1:	a1
1495a0015d6SChris Zankel *   a2:	new stack pointer, original in DEPC
15099d5040eSMax Filippov *   a3:	a3
1515a0015d6SChris Zankel *   depc:	a2, original value saved on stack (PT_DEPC)
15299d5040eSMax Filippov *   excsave_1:	dispatch table
1535a0015d6SChris Zankel *
1545a0015d6SChris Zankel *   PT_DEPC >= VALID_DOUBLE_EXCEPTION_ADDRESS: double exception, DEPC
1555a0015d6SChris Zankel *	     <  VALID_DOUBLE_EXCEPTION_ADDRESS: regular exception
1565a0015d6SChris Zankel */
1575a0015d6SChris Zankel
1585a0015d6SChris Zankel
1595a0015d6SChris ZankelENTRY(fast_unaligned)
1605a0015d6SChris Zankel
1615a0015d6SChris Zankel	/* Note: We don't expect the address to be aligned on a word
1625a0015d6SChris Zankel	 *       boundary. After all, the processor generated that exception
1635a0015d6SChris Zankel	 *       and it would be a hardware fault.
1645a0015d6SChris Zankel	 */
1655a0015d6SChris Zankel
1665a0015d6SChris Zankel	/* Save some working register */
1675a0015d6SChris Zankel
1685a0015d6SChris Zankel	s32i	a4, a2, PT_AREG4
1695a0015d6SChris Zankel	s32i	a5, a2, PT_AREG5
1705a0015d6SChris Zankel	s32i	a6, a2, PT_AREG6
1715a0015d6SChris Zankel	s32i	a7, a2, PT_AREG7
1725a0015d6SChris Zankel	s32i	a8, a2, PT_AREG8
1735a0015d6SChris Zankel
174bc5378fcSMax Filippov	rsr	a0, depc
1755a0015d6SChris Zankel	s32i	a0, a2, PT_AREG2
1765a0015d6SChris Zankel	s32i	a3, a2, PT_AREG3
1775a0015d6SChris Zankel
178c3ef1f4dSMax Filippov	rsr	a3, excsave1
179c3ef1f4dSMax Filippov	movi	a4, fast_unaligned_fixup
180c3ef1f4dSMax Filippov	s32i	a4, a3, EXC_TABLE_FIXUP
181c3ef1f4dSMax Filippov
1825a0015d6SChris Zankel	/* Keep value of SAR in a0 */
1835a0015d6SChris Zankel
184bc5378fcSMax Filippov	rsr	a0, sar
185bc5378fcSMax Filippov	rsr	a8, excvaddr		# load unaligned memory address
1865a0015d6SChris Zankel
1875a0015d6SChris Zankel	/* Now, identify one of the following load/store instructions.
1885a0015d6SChris Zankel	 *
1895a0015d6SChris Zankel	 * The only possible danger of a double exception on the
1905a0015d6SChris Zankel	 * following l32i instructions is kernel code in vmalloc
1915a0015d6SChris Zankel	 * memory. The processor was just executing at the EPC_1
1925a0015d6SChris Zankel	 * address, and indeed, already fetched the instruction.  That
1935a0015d6SChris Zankel	 * guarantees a TLB mapping, which hasn't been replaced by
1945a0015d6SChris Zankel	 * this unaligned exception handler that uses only static TLB
1955a0015d6SChris Zankel	 * mappings. However, high-level interrupt handlers might
1965a0015d6SChris Zankel	 * modify TLB entries, so for the generic case, we register a
1975a0015d6SChris Zankel	 * TABLE_FIXUP handler here, too.
1985a0015d6SChris Zankel	 */
1995a0015d6SChris Zankel
2005a0015d6SChris Zankel	/* a3...a6 saved on stack, a2 = SP */
2015a0015d6SChris Zankel
2025a0015d6SChris Zankel	/* Extract the instruction that caused the unaligned access. */
2035a0015d6SChris Zankel
204bc5378fcSMax Filippov	rsr	a7, epc1	# load exception address
2055a0015d6SChris Zankel	movi	a3, ~3
2065a0015d6SChris Zankel	and	a3, a3, a7	# mask lower bits
2075a0015d6SChris Zankel
2085a0015d6SChris Zankel	l32i	a4, a3, 0	# load 2 words
2095a0015d6SChris Zankel	l32i	a5, a3, 4
2105a0015d6SChris Zankel
2115a0015d6SChris Zankel	__ssa8	a7
2125a0015d6SChris Zankel	__src_b	a4, a4, a5	# a4 has the instruction
2135a0015d6SChris Zankel
2145a0015d6SChris Zankel	/* Analyze the instruction (load or store?). */
2155a0015d6SChris Zankel
2165a0015d6SChris Zankel	extui	a5, a4, INSN_OP0, 4	# get insn.op0 nibble
2175a0015d6SChris Zankel
218173d6681SChris Zankel#if XCHAL_HAVE_DENSITY
2195a0015d6SChris Zankel	_beqi	a5, OP0_L32I_N, .Lload	# L32I.N, jump
2205a0015d6SChris Zankel	addi	a6, a5, -OP0_S32I_N
2215a0015d6SChris Zankel	_beqz	a6, .Lstore		# S32I.N, do a store
2225a0015d6SChris Zankel#endif
2235a0015d6SChris Zankel	/* 'store indicator bit' not set, jump */
2245a0015d6SChris Zankel	_bbci.l	a4, OP1_SI_BIT + INSN_OP1, .Lload
2255a0015d6SChris Zankel
2265a0015d6SChris Zankel	/* Store: Jump to table entry to get the value in the source register.*/
2275a0015d6SChris Zankel
2285a0015d6SChris Zankel.Lstore:movi	a5, .Lstore_table	# table
2295a0015d6SChris Zankel	extui	a6, a4, INSN_T, 4	# get source register
2305a0015d6SChris Zankel	addx8	a5, a6, a5
2315a0015d6SChris Zankel	jx	a5			# jump into table
2325a0015d6SChris Zankel
2335a0015d6SChris Zankel	/* Load: Load memory address. */
2345a0015d6SChris Zankel
2355a0015d6SChris Zankel.Lload: movi	a3, ~3
2365a0015d6SChris Zankel	and	a3, a3, a8		# align memory address
2375a0015d6SChris Zankel
2385a0015d6SChris Zankel	__ssa8	a8
2395a0015d6SChris Zankel#ifdef UNALIGNED_USER_EXCEPTION
2405a0015d6SChris Zankel	addi	a3, a3, 8
2415a0015d6SChris Zankel	l32e	a5, a3, -8
2425a0015d6SChris Zankel	l32e	a6, a3, -4
2435a0015d6SChris Zankel#else
2445a0015d6SChris Zankel	l32i	a5, a3, 0
2455a0015d6SChris Zankel	l32i	a6, a3, 4
2465a0015d6SChris Zankel#endif
2475a0015d6SChris Zankel	__src_b	a3, a5, a6		# a3 has the data word
2485a0015d6SChris Zankel
249173d6681SChris Zankel#if XCHAL_HAVE_DENSITY
2505a0015d6SChris Zankel	addi	a7, a7, 2		# increment PC (assume 16-bit insn)
2515a0015d6SChris Zankel
2525a0015d6SChris Zankel	extui	a5, a4, INSN_OP0, 4
2535a0015d6SChris Zankel	_beqi	a5, OP0_L32I_N, 1f	# l32i.n: jump
2545a0015d6SChris Zankel
2555a0015d6SChris Zankel	addi	a7, a7, 1
2565a0015d6SChris Zankel#else
2575a0015d6SChris Zankel	addi	a7, a7, 3
2585a0015d6SChris Zankel#endif
2595a0015d6SChris Zankel
2605a0015d6SChris Zankel	extui	a5, a4, INSN_OP1, 4
2615a0015d6SChris Zankel	_beqi	a5, OP1_L32I, 1f	# l32i: jump
2625a0015d6SChris Zankel
2635a0015d6SChris Zankel	extui	a3, a3, 0, 16		# extract lower 16 bits
2645a0015d6SChris Zankel	_beqi	a5, OP1_L16UI, 1f
2655a0015d6SChris Zankel	addi	a5, a5, -OP1_L16SI
2665a0015d6SChris Zankel	_bnez	a5, .Linvalid_instruction_load
2675a0015d6SChris Zankel
2685a0015d6SChris Zankel	/* sign extend value */
2695a0015d6SChris Zankel
2705a0015d6SChris Zankel	slli	a3, a3, 16
2715a0015d6SChris Zankel	srai	a3, a3, 16
2725a0015d6SChris Zankel
2735a0015d6SChris Zankel	/* Set target register. */
2745a0015d6SChris Zankel
2755a0015d6SChris Zankel1:
2765a0015d6SChris Zankel	extui	a4, a4, INSN_T, 4	# extract target register
2775a0015d6SChris Zankel	movi	a5, .Lload_table
2785a0015d6SChris Zankel	addx8	a4, a4, a5
2795a0015d6SChris Zankel	jx	a4			# jump to entry for target register
2805a0015d6SChris Zankel
2815a0015d6SChris Zankel	.align	8
2825a0015d6SChris Zankel.Lload_table:
2835a0015d6SChris Zankel	s32i	a3, a2, PT_AREG0;	_j .Lexit;	.align 8
2845a0015d6SChris Zankel	mov	a1, a3;			_j .Lexit;	.align 8 # fishy??
2855a0015d6SChris Zankel	s32i	a3, a2, PT_AREG2;	_j .Lexit;	.align 8
2865a0015d6SChris Zankel	s32i	a3, a2, PT_AREG3;	_j .Lexit;	.align 8
2875a0015d6SChris Zankel	s32i	a3, a2, PT_AREG4;	_j .Lexit;	.align 8
2885a0015d6SChris Zankel	s32i	a3, a2, PT_AREG5;	_j .Lexit;	.align 8
2895a0015d6SChris Zankel	s32i	a3, a2, PT_AREG6;	_j .Lexit;	.align 8
2905a0015d6SChris Zankel	s32i	a3, a2, PT_AREG7;	_j .Lexit;	.align 8
2915a0015d6SChris Zankel	s32i	a3, a2, PT_AREG8;	_j .Lexit;	.align 8
2925a0015d6SChris Zankel	mov	a9, a3		;	_j .Lexit;	.align 8
2935a0015d6SChris Zankel	mov	a10, a3		;	_j .Lexit;	.align 8
2945a0015d6SChris Zankel	mov	a11, a3		;	_j .Lexit;	.align 8
2955a0015d6SChris Zankel	mov	a12, a3		;	_j .Lexit;	.align 8
2965a0015d6SChris Zankel	mov	a13, a3		;	_j .Lexit;	.align 8
2975a0015d6SChris Zankel	mov	a14, a3		;	_j .Lexit;	.align 8
2985a0015d6SChris Zankel	mov	a15, a3		;	_j .Lexit;	.align 8
2995a0015d6SChris Zankel
3005a0015d6SChris Zankel.Lstore_table:
3015a0015d6SChris Zankel	l32i	a3, a2, PT_AREG0;	_j 1f;	.align 8
3025a0015d6SChris Zankel	mov	a3, a1;			_j 1f;	.align 8	# fishy??
3035a0015d6SChris Zankel	l32i	a3, a2, PT_AREG2;	_j 1f;	.align 8
3045a0015d6SChris Zankel	l32i	a3, a2, PT_AREG3;	_j 1f;	.align 8
3055a0015d6SChris Zankel	l32i	a3, a2, PT_AREG4;	_j 1f;	.align 8
3065a0015d6SChris Zankel	l32i	a3, a2, PT_AREG5;	_j 1f;	.align 8
3075a0015d6SChris Zankel	l32i	a3, a2, PT_AREG6;	_j 1f;	.align 8
3085a0015d6SChris Zankel	l32i	a3, a2, PT_AREG7;	_j 1f;	.align 8
3095a0015d6SChris Zankel	l32i	a3, a2, PT_AREG8;	_j 1f;	.align 8
3105a0015d6SChris Zankel	mov	a3, a9		;	_j 1f;	.align 8
3115a0015d6SChris Zankel	mov	a3, a10		;	_j 1f;	.align 8
3125a0015d6SChris Zankel	mov	a3, a11		;	_j 1f;	.align 8
3135a0015d6SChris Zankel	mov	a3, a12		;	_j 1f;	.align 8
3145a0015d6SChris Zankel	mov	a3, a13		;	_j 1f;	.align 8
3155a0015d6SChris Zankel	mov	a3, a14		;	_j 1f;	.align 8
3165a0015d6SChris Zankel	mov	a3, a15		;	_j 1f;	.align 8
3175a0015d6SChris Zankel
31821570465SMax Filippov	/* We cannot handle this exception. */
31921570465SMax Filippov
32021570465SMax Filippov	.extern _kernel_exception
32121570465SMax Filippov.Linvalid_instruction_load:
32221570465SMax Filippov.Linvalid_instruction_store:
32321570465SMax Filippov
32421570465SMax Filippov	movi	a4, 0
32521570465SMax Filippov	rsr	a3, excsave1
32621570465SMax Filippov	s32i	a4, a3, EXC_TABLE_FIXUP
32721570465SMax Filippov
32821570465SMax Filippov	/* Restore a4...a8 and SAR, set SP, and jump to default exception. */
32921570465SMax Filippov
33021570465SMax Filippov	l32i	a8, a2, PT_AREG8
33121570465SMax Filippov	l32i	a7, a2, PT_AREG7
33221570465SMax Filippov	l32i	a6, a2, PT_AREG6
33321570465SMax Filippov	l32i	a5, a2, PT_AREG5
33421570465SMax Filippov	l32i	a4, a2, PT_AREG4
33521570465SMax Filippov	wsr	a0, sar
33621570465SMax Filippov	mov	a1, a2
33721570465SMax Filippov
33821570465SMax Filippov	rsr	a0, ps
33921570465SMax Filippov	bbsi.l  a0, PS_UM_BIT, 2f     # jump if user mode
34021570465SMax Filippov
34121570465SMax Filippov	movi	a0, _kernel_exception
34221570465SMax Filippov	jx	a0
34321570465SMax Filippov
34421570465SMax Filippov2:	movi	a0, _user_exception
34521570465SMax Filippov	jx	a0
34621570465SMax Filippov
3475a0015d6SChris Zankel1: 	# a7: instruction pointer, a4: instruction, a3: value
3485a0015d6SChris Zankel
3495a0015d6SChris Zankel	movi	a6, 0			# mask: ffffffff:00000000
3505a0015d6SChris Zankel
351173d6681SChris Zankel#if XCHAL_HAVE_DENSITY
3525a0015d6SChris Zankel	addi	a7, a7, 2		# incr. PC,assume 16-bit instruction
3535a0015d6SChris Zankel
3545a0015d6SChris Zankel	extui	a5, a4, INSN_OP0, 4	# extract OP0
3555a0015d6SChris Zankel	addi	a5, a5, -OP0_S32I_N
3565a0015d6SChris Zankel	_beqz	a5, 1f			# s32i.n: jump
3575a0015d6SChris Zankel
3585a0015d6SChris Zankel	addi	a7, a7, 1		# increment PC, 32-bit instruction
3595a0015d6SChris Zankel#else
3605a0015d6SChris Zankel	addi	a7, a7, 3		# increment PC, 32-bit instruction
3615a0015d6SChris Zankel#endif
3625a0015d6SChris Zankel
3635a0015d6SChris Zankel	extui	a5, a4, INSN_OP1, 4	# extract OP1
3645a0015d6SChris Zankel	_beqi	a5, OP1_S32I, 1f	# jump if 32 bit store
3655a0015d6SChris Zankel	_bnei	a5, OP1_S16I, .Linvalid_instruction_store
3665a0015d6SChris Zankel
3675a0015d6SChris Zankel	movi	a5, -1
3685a0015d6SChris Zankel	__extl	a3, a3			# get 16-bit value
3695a0015d6SChris Zankel	__exth	a6, a5			# get 16-bit mask ffffffff:ffff0000
3705a0015d6SChris Zankel
3715a0015d6SChris Zankel	/* Get memory address */
3725a0015d6SChris Zankel
3735a0015d6SChris Zankel1:
3745a0015d6SChris Zankel	movi	a4, ~3
3755a0015d6SChris Zankel	and	a4, a4, a8		# align memory address
3765a0015d6SChris Zankel
3775a0015d6SChris Zankel	/* Insert value into memory */
3785a0015d6SChris Zankel
3795a0015d6SChris Zankel	movi	a5, -1			# mask: ffffffff:XXXX0000
3805a0015d6SChris Zankel#ifdef UNALIGNED_USER_EXCEPTION
3815a0015d6SChris Zankel	addi	a4, a4, 8
3825a0015d6SChris Zankel#endif
3835a0015d6SChris Zankel
3845a0015d6SChris Zankel	__ssa8r a8
385e9500dd8SMax Filippov	__src_b	a8, a5, a6		# lo-mask  F..F0..0 (BE) 0..0F..F (LE)
3865a0015d6SChris Zankel	__src_b	a6, a6, a5		# hi-mask  0..0F..F (BE) F..F0..0 (LE)
3875a0015d6SChris Zankel#ifdef UNALIGNED_USER_EXCEPTION
3885a0015d6SChris Zankel	l32e	a5, a4, -8
3895a0015d6SChris Zankel#else
3905a0015d6SChris Zankel	l32i	a5, a4, 0		# load lower address word
3915a0015d6SChris Zankel#endif
392e9500dd8SMax Filippov	and	a5, a5, a8		# mask
393e9500dd8SMax Filippov	__sh	a8, a3 			# shift value
394e9500dd8SMax Filippov	or	a5, a5, a8		# or with original value
3955a0015d6SChris Zankel#ifdef UNALIGNED_USER_EXCEPTION
3965a0015d6SChris Zankel	s32e	a5, a4, -8
397e9500dd8SMax Filippov	l32e	a8, a4, -4
3985a0015d6SChris Zankel#else
3995a0015d6SChris Zankel	s32i	a5, a4, 0		# store
400e9500dd8SMax Filippov	l32i	a8, a4, 4		# same for upper address word
4015a0015d6SChris Zankel#endif
4025a0015d6SChris Zankel	__sl	a5, a3
403e9500dd8SMax Filippov	and	a6, a8, a6
4045a0015d6SChris Zankel	or	a6, a6, a5
4055a0015d6SChris Zankel#ifdef UNALIGNED_USER_EXCEPTION
4065a0015d6SChris Zankel	s32e	a6, a4, -4
4075a0015d6SChris Zankel#else
4085a0015d6SChris Zankel	s32i	a6, a4, 4
4095a0015d6SChris Zankel#endif
4105a0015d6SChris Zankel
4115a0015d6SChris Zankel.Lexit:
412e9500dd8SMax Filippov#if XCHAL_HAVE_LOOPS
413e9500dd8SMax Filippov	rsr	a4, lend		# check if we reached LEND
414e9500dd8SMax Filippov	bne	a7, a4, 1f
415e9500dd8SMax Filippov	rsr	a4, lcount		# and LCOUNT != 0
416e9500dd8SMax Filippov	beqz	a4, 1f
417e9500dd8SMax Filippov	addi	a4, a4, -1		# decrement LCOUNT and set
418e9500dd8SMax Filippov	rsr	a7, lbeg		# set PC to LBEGIN
419e9500dd8SMax Filippov	wsr	a4, lcount
420e9500dd8SMax Filippov#endif
421e9500dd8SMax Filippov
422e9500dd8SMax Filippov1:	wsr	a7, epc1		# skip emulated instruction
423e9500dd8SMax Filippov
424*a83b02e9SMax Filippov	/* Update icount if we're single-stepping in userspace. */
425*a83b02e9SMax Filippov	rsr	a4, icountlevel
426*a83b02e9SMax Filippov	beqz	a4, 1f
427*a83b02e9SMax Filippov	bgeui	a4, LOCKLEVEL + 1, 1f
428*a83b02e9SMax Filippov	rsr	a4, icount
429*a83b02e9SMax Filippov	addi	a4, a4, 1
430*a83b02e9SMax Filippov	wsr	a4, icount
431*a83b02e9SMax Filippov1:
4325a0015d6SChris Zankel	movi	a4, 0
433bc5378fcSMax Filippov	rsr	a3, excsave1
4345a0015d6SChris Zankel	s32i	a4, a3, EXC_TABLE_FIXUP
4355a0015d6SChris Zankel
4365a0015d6SChris Zankel	/* Restore working register */
4375a0015d6SChris Zankel
438173d6681SChris Zankel	l32i	a8, a2, PT_AREG8
4395a0015d6SChris Zankel	l32i	a7, a2, PT_AREG7
4405a0015d6SChris Zankel	l32i	a6, a2, PT_AREG6
4415a0015d6SChris Zankel	l32i	a5, a2, PT_AREG5
4425a0015d6SChris Zankel	l32i	a4, a2, PT_AREG4
4435a0015d6SChris Zankel	l32i	a3, a2, PT_AREG3
4445a0015d6SChris Zankel
4455a0015d6SChris Zankel	/* restore SAR and return */
4465a0015d6SChris Zankel
447bc5378fcSMax Filippov	wsr	a0, sar
4485a0015d6SChris Zankel	l32i	a0, a2, PT_AREG0
4495a0015d6SChris Zankel	l32i	a2, a2, PT_AREG2
4505a0015d6SChris Zankel	rfe
4515a0015d6SChris Zankel
452d1538c46SChris ZankelENDPROC(fast_unaligned)
4535a0015d6SChris Zankel
454c3ef1f4dSMax FilippovENTRY(fast_unaligned_fixup)
455c3ef1f4dSMax Filippov
456c3ef1f4dSMax Filippov	l32i	a2, a3, EXC_TABLE_DOUBLE_SAVE
457c3ef1f4dSMax Filippov	wsr	a3, excsave1
458c3ef1f4dSMax Filippov
459c3ef1f4dSMax Filippov	l32i	a8, a2, PT_AREG8
460c3ef1f4dSMax Filippov	l32i	a7, a2, PT_AREG7
461c3ef1f4dSMax Filippov	l32i	a6, a2, PT_AREG6
462c3ef1f4dSMax Filippov	l32i	a5, a2, PT_AREG5
463c3ef1f4dSMax Filippov	l32i	a4, a2, PT_AREG4
464c3ef1f4dSMax Filippov	l32i	a0, a2, PT_AREG2
465c3ef1f4dSMax Filippov	xsr	a0, depc			# restore depc and a0
466c3ef1f4dSMax Filippov	wsr	a0, sar
467c3ef1f4dSMax Filippov
468c3ef1f4dSMax Filippov	rsr	a0, exccause
469c3ef1f4dSMax Filippov	s32i	a0, a2, PT_DEPC			# mark as a regular exception
470c3ef1f4dSMax Filippov
471c3ef1f4dSMax Filippov	rsr	a0, ps
472c3ef1f4dSMax Filippov	bbsi.l  a0, PS_UM_BIT, 1f		# jump if user mode
473c3ef1f4dSMax Filippov
474c3ef1f4dSMax Filippov	rsr	a0, exccause
475c3ef1f4dSMax Filippov	addx4	a0, a0, a3              	# find entry in table
476c3ef1f4dSMax Filippov	l32i	a0, a0, EXC_TABLE_FAST_KERNEL   # load handler
477c3ef1f4dSMax Filippov	l32i	a3, a2, PT_AREG3
478c3ef1f4dSMax Filippov	jx	a0
479c3ef1f4dSMax Filippov1:
480c3ef1f4dSMax Filippov	rsr	a0, exccause
481c3ef1f4dSMax Filippov	addx4	a0, a0, a3              	# find entry in table
482c3ef1f4dSMax Filippov	l32i	a0, a0, EXC_TABLE_FAST_USER     # load handler
483c3ef1f4dSMax Filippov	l32i	a3, a2, PT_AREG3
484c3ef1f4dSMax Filippov	jx	a0
485c3ef1f4dSMax Filippov
486c3ef1f4dSMax FilippovENDPROC(fast_unaligned_fixup)
487c3ef1f4dSMax Filippov
4885a0015d6SChris Zankel#endif /* XCHAL_UNALIGNED_LOAD_EXCEPTION || XCHAL_UNALIGNED_STORE_EXCEPTION */
489