1 /* 2 * arch/arm/include/asm/assembler.h 3 * 4 * Copyright (C) 1996-2000 Russell King 5 * 6 * This program is free software; you can redistribute it and/or modify 7 * it under the terms of the GNU General Public License version 2 as 8 * published by the Free Software Foundation. 9 * 10 * This file contains arm architecture specific defines 11 * for the different processors. 12 * 13 * Do not include any C declarations in this file - it is included by 14 * assembler source. 15 */ 16 #ifndef __ASSEMBLY__ 17 #error "Only include this from assembly code" 18 #endif 19 20 #include <asm/ptrace.h> 21 22 /* 23 * Endian independent macros for shifting bytes within registers. 24 */ 25 #ifndef __ARMEB__ 26 #define pull lsr 27 #define push lsl 28 #define get_byte_0 lsl #0 29 #define get_byte_1 lsr #8 30 #define get_byte_2 lsr #16 31 #define get_byte_3 lsr #24 32 #define put_byte_0 lsl #0 33 #define put_byte_1 lsl #8 34 #define put_byte_2 lsl #16 35 #define put_byte_3 lsl #24 36 #else 37 #define pull lsl 38 #define push lsr 39 #define get_byte_0 lsr #24 40 #define get_byte_1 lsr #16 41 #define get_byte_2 lsr #8 42 #define get_byte_3 lsl #0 43 #define put_byte_0 lsl #24 44 #define put_byte_1 lsl #16 45 #define put_byte_2 lsl #8 46 #define put_byte_3 lsl #0 47 #endif 48 49 /* 50 * Data preload for architectures that support it 51 */ 52 #if __LINUX_ARM_ARCH__ >= 5 53 #define PLD(code...) code 54 #else 55 #define PLD(code...) 56 #endif 57 58 /* 59 * This can be used to enable code to cacheline align the destination 60 * pointer when bulk writing to memory. Experiments on StrongARM and 61 * XScale didn't show this a worthwhile thing to do when the cache is not 62 * set to write-allocate (this would need further testing on XScale when WA 63 * is used). 64 * 65 * On Feroceon there is much to gain however, regardless of cache mode. 66 */ 67 #ifdef CONFIG_CPU_FEROCEON 68 #define CALGN(code...) code 69 #else 70 #define CALGN(code...) 71 #endif 72 73 /* 74 * Enable and disable interrupts 75 */ 76 #if __LINUX_ARM_ARCH__ >= 6 77 .macro disable_irq 78 cpsid i 79 .endm 80 81 .macro enable_irq 82 cpsie i 83 .endm 84 #else 85 .macro disable_irq 86 msr cpsr_c, #PSR_I_BIT | SVC_MODE 87 .endm 88 89 .macro enable_irq 90 msr cpsr_c, #SVC_MODE 91 .endm 92 #endif 93 94 /* 95 * Save the current IRQ state and disable IRQs. Note that this macro 96 * assumes FIQs are enabled, and that the processor is in SVC mode. 97 */ 98 .macro save_and_disable_irqs, oldcpsr 99 mrs \oldcpsr, cpsr 100 disable_irq 101 .endm 102 103 /* 104 * Restore interrupt state previously stored in a register. We don't 105 * guarantee that this will preserve the flags. 106 */ 107 .macro restore_irqs, oldcpsr 108 msr cpsr_c, \oldcpsr 109 .endm 110 111 #define USER(x...) \ 112 9999: x; \ 113 .section __ex_table,"a"; \ 114 .align 3; \ 115 .long 9999b,9001f; \ 116 .previous 117 118 /* 119 * SMP data memory barrier 120 */ 121 .macro smp_dmb 122 #ifdef CONFIG_SMP 123 #if __LINUX_ARM_ARCH__ >= 7 124 dmb 125 #elif __LINUX_ARM_ARCH__ == 6 126 mcr p15, 0, r0, c7, c10, 5 @ dmb 127 #endif 128 #endif 129 .endm 130