1 /* SPDX-License-Identifier: GPL-2.0 */ 2 3 #ifndef _ASM_X86_CPU_ENTRY_AREA_H 4 #define _ASM_X86_CPU_ENTRY_AREA_H 5 6 #include <linux/percpu-defs.h> 7 #include <asm/processor.h> 8 #include <asm/intel_ds.h> 9 #include <asm/pgtable_areas.h> 10 11 #ifdef CONFIG_X86_64 12 13 #ifdef CONFIG_AMD_MEM_ENCRYPT 14 #define VC_EXCEPTION_STKSZ EXCEPTION_STKSZ 15 #else 16 #define VC_EXCEPTION_STKSZ 0 17 #endif 18 19 /* Macro to enforce the same ordering and stack sizes */ 20 #define ESTACKS_MEMBERS(guardsize, optional_stack_size) \ 21 char DF_stack_guard[guardsize]; \ 22 char DF_stack[EXCEPTION_STKSZ]; \ 23 char NMI_stack_guard[guardsize]; \ 24 char NMI_stack[EXCEPTION_STKSZ]; \ 25 char DB_stack_guard[guardsize]; \ 26 char DB_stack[EXCEPTION_STKSZ]; \ 27 char MCE_stack_guard[guardsize]; \ 28 char MCE_stack[EXCEPTION_STKSZ]; \ 29 char VC_stack_guard[guardsize]; \ 30 char VC_stack[optional_stack_size]; \ 31 char VC2_stack_guard[guardsize]; \ 32 char VC2_stack[optional_stack_size]; \ 33 char IST_top_guard[guardsize]; \ 34 35 /* The exception stacks' physical storage. No guard pages required */ 36 struct exception_stacks { 37 ESTACKS_MEMBERS(0, VC_EXCEPTION_STKSZ) 38 }; 39 40 /* The effective cpu entry area mapping with guard pages. */ 41 struct cea_exception_stacks { 42 ESTACKS_MEMBERS(PAGE_SIZE, EXCEPTION_STKSZ) 43 }; 44 45 /* 46 * The exception stack ordering in [cea_]exception_stacks 47 */ 48 enum exception_stack_ordering { 49 ESTACK_DF, 50 ESTACK_NMI, 51 ESTACK_DB, 52 ESTACK_MCE, 53 ESTACK_VC, 54 ESTACK_VC2, 55 N_EXCEPTION_STACKS 56 }; 57 58 #define CEA_ESTACK_SIZE(st) \ 59 sizeof(((struct cea_exception_stacks *)0)->st## _stack) 60 61 #define CEA_ESTACK_BOT(ceastp, st) \ 62 ((unsigned long)&(ceastp)->st## _stack) 63 64 #define CEA_ESTACK_TOP(ceastp, st) \ 65 (CEA_ESTACK_BOT(ceastp, st) + CEA_ESTACK_SIZE(st)) 66 67 #define CEA_ESTACK_OFFS(st) \ 68 offsetof(struct cea_exception_stacks, st## _stack) 69 70 #define CEA_ESTACK_PAGES \ 71 (sizeof(struct cea_exception_stacks) / PAGE_SIZE) 72 73 #endif 74 75 #ifdef CONFIG_X86_32 76 struct doublefault_stack { 77 unsigned long stack[(PAGE_SIZE - sizeof(struct x86_hw_tss)) / sizeof(unsigned long)]; 78 struct x86_hw_tss tss; 79 } __aligned(PAGE_SIZE); 80 #endif 81 82 /* 83 * cpu_entry_area is a percpu region that contains things needed by the CPU 84 * and early entry/exit code. Real types aren't used for all fields here 85 * to avoid circular header dependencies. 86 * 87 * Every field is a virtual alias of some other allocated backing store. 88 * There is no direct allocation of a struct cpu_entry_area. 89 */ 90 struct cpu_entry_area { 91 char gdt[PAGE_SIZE]; 92 93 /* 94 * The GDT is just below entry_stack and thus serves (on x86_64) as 95 * a read-only guard page. On 32-bit the GDT must be writeable, so 96 * it needs an extra guard page. 97 */ 98 #ifdef CONFIG_X86_32 99 char guard_entry_stack[PAGE_SIZE]; 100 #endif 101 struct entry_stack_page entry_stack_page; 102 103 #ifdef CONFIG_X86_32 104 char guard_doublefault_stack[PAGE_SIZE]; 105 struct doublefault_stack doublefault_stack; 106 #endif 107 108 /* 109 * On x86_64, the TSS is mapped RO. On x86_32, it's mapped RW because 110 * we need task switches to work, and task switches write to the TSS. 111 */ 112 struct tss_struct tss; 113 114 #ifdef CONFIG_X86_64 115 /* 116 * Exception stacks used for IST entries with guard pages. 117 */ 118 struct cea_exception_stacks estacks; 119 #endif 120 /* 121 * Per CPU debug store for Intel performance monitoring. Wastes a 122 * full page at the moment. 123 */ 124 struct debug_store cpu_debug_store; 125 /* 126 * The actual PEBS/BTS buffers must be mapped to user space 127 * Reserve enough fixmap PTEs. 128 */ 129 struct debug_store_buffers cpu_debug_buffers; 130 }; 131 132 #define CPU_ENTRY_AREA_SIZE (sizeof(struct cpu_entry_area)) 133 134 DECLARE_PER_CPU(struct cpu_entry_area *, cpu_entry_area); 135 DECLARE_PER_CPU(struct cea_exception_stacks *, cea_exception_stacks); 136 137 extern void setup_cpu_entry_areas(void); 138 extern void cea_set_pte(void *cea_vaddr, phys_addr_t pa, pgprot_t flags); 139 140 extern struct cpu_entry_area *get_cpu_entry_area(int cpu); 141 142 static __always_inline struct entry_stack *cpu_entry_stack(int cpu) 143 { 144 return &get_cpu_entry_area(cpu)->entry_stack_page.stack; 145 } 146 147 #define __this_cpu_ist_top_va(name) \ 148 CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name) 149 150 #define __this_cpu_ist_bottom_va(name) \ 151 CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name) 152 153 #endif 154