1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 2669ebabbSIngo Molnar #ifndef __ASM_X86_XSAVE_H 3669ebabbSIngo Molnar #define __ASM_X86_XSAVE_H 4669ebabbSIngo Molnar 559a36d16SIngo Molnar #include <linux/uaccess.h> 60cecca9dSRik van Riel #include <linux/types.h> 70cecca9dSRik van Riel 80cecca9dSRik van Riel #include <asm/processor.h> 9784a4661SDave Hansen #include <asm/fpu/api.h> 100cecca9dSRik van Riel #include <asm/user.h> 11669ebabbSIngo Molnar 12669ebabbSIngo Molnar /* Bit 63 of XCR0 is reserved for future expansion */ 13d91cab78SDave Hansen #define XFEATURE_MASK_EXTEND (~(XFEATURE_MASK_FPSSE | (1ULL << 63))) 14669ebabbSIngo Molnar 15677b98bdSIngo Molnar #define XSTATE_CPUID 0x0000000d 16677b98bdSIngo Molnar 17669ebabbSIngo Molnar #define FXSAVE_SIZE 512 18669ebabbSIngo Molnar 19669ebabbSIngo Molnar #define XSAVE_HDR_SIZE 64 20669ebabbSIngo Molnar #define XSAVE_HDR_OFFSET FXSAVE_SIZE 21669ebabbSIngo Molnar 22669ebabbSIngo Molnar #define XSAVE_YMM_SIZE 256 23669ebabbSIngo Molnar #define XSAVE_YMM_OFFSET (XSAVE_HDR_SIZE + XSAVE_HDR_OFFSET) 24669ebabbSIngo Molnar 25ce711ea3SKan Liang #define XSAVE_ALIGNMENT 64 26ce711ea3SKan Liang 278ab22804SFenghua Yu /* All currently supported user features */ 288ab22804SFenghua Yu #define XFEATURE_MASK_USER_SUPPORTED (XFEATURE_MASK_FP | \ 29a65050c6SYu-cheng Yu XFEATURE_MASK_SSE | \ 30d91cab78SDave Hansen XFEATURE_MASK_YMM | \ 31d91cab78SDave Hansen XFEATURE_MASK_OPMASK | \ 32d91cab78SDave Hansen XFEATURE_MASK_ZMM_Hi256 | \ 33c8df4009SDave Hansen XFEATURE_MASK_Hi16_ZMM | \ 342f7fada2SAndy Lutomirski XFEATURE_MASK_PKRU | \ 352f7fada2SAndy Lutomirski XFEATURE_MASK_BNDREGS | \ 362f7fada2SAndy Lutomirski XFEATURE_MASK_BNDCSR) 37669ebabbSIngo Molnar 382ebe81c6SThomas Gleixner /* 392ebe81c6SThomas Gleixner * Features which are restored when returning to user space. 402ebe81c6SThomas Gleixner * PKRU is not restored on return to user space because PKRU 412ebe81c6SThomas Gleixner * is switched eagerly in switch_to() and flush_thread() 422ebe81c6SThomas Gleixner */ 432ebe81c6SThomas Gleixner #define XFEATURE_MASK_USER_RESTORE \ 442ebe81c6SThomas Gleixner (XFEATURE_MASK_USER_SUPPORTED & ~XFEATURE_MASK_PKRU) 452ebe81c6SThomas Gleixner 46*23686ef2SThomas Gleixner /* Features which are dynamically enabled for a process on request */ 47*23686ef2SThomas Gleixner #define XFEATURE_MASK_USER_DYNAMIC 0ULL 48*23686ef2SThomas Gleixner 498ab22804SFenghua Yu /* All currently supported supervisor features */ 50b454feb9SYu-cheng Yu #define XFEATURE_MASK_SUPERVISOR_SUPPORTED (XFEATURE_MASK_PASID) 518ab22804SFenghua Yu 528ab22804SFenghua Yu /* 53f0dccc9dSKan Liang * A supervisor state component may not always contain valuable information, 54f0dccc9dSKan Liang * and its size may be huge. Saving/restoring such supervisor state components 55f0dccc9dSKan Liang * at each context switch can cause high CPU and space overhead, which should 56f0dccc9dSKan Liang * be avoided. Such supervisor state components should only be saved/restored 5701707b66SAndy Lutomirski * on demand. The on-demand supervisor features are set in this mask. 58f0dccc9dSKan Liang * 5901707b66SAndy Lutomirski * Unlike the existing supported supervisor features, an independent supervisor 60f0dccc9dSKan Liang * feature does not allocate a buffer in task->fpu, and the corresponding 61f0dccc9dSKan Liang * supervisor state component cannot be saved/restored at each context switch. 62f0dccc9dSKan Liang * 6301707b66SAndy Lutomirski * To support an independent supervisor feature, a developer should follow the 64f0dccc9dSKan Liang * dos and don'ts as below: 65f0dccc9dSKan Liang * - Do dynamically allocate a buffer for the supervisor state component. 66f0dccc9dSKan Liang * - Do manually invoke the XSAVES/XRSTORS instruction to save/restore the 67f0dccc9dSKan Liang * state component to/from the buffer. 6801707b66SAndy Lutomirski * - Don't set the bit corresponding to the independent supervisor feature in 69f0dccc9dSKan Liang * IA32_XSS at run time, since it has been set at boot time. 70f0dccc9dSKan Liang */ 7101707b66SAndy Lutomirski #define XFEATURE_MASK_INDEPENDENT (XFEATURE_MASK_LBR) 72f0dccc9dSKan Liang 73f0dccc9dSKan Liang /* 748ab22804SFenghua Yu * Unsupported supervisor features. When a supervisor feature in this mask is 758ab22804SFenghua Yu * supported in the future, move it to the supported supervisor feature mask. 768ab22804SFenghua Yu */ 778ab22804SFenghua Yu #define XFEATURE_MASK_SUPERVISOR_UNSUPPORTED (XFEATURE_MASK_PT) 788ab22804SFenghua Yu 798ab22804SFenghua Yu /* All supervisor states including supported and unsupported states. */ 808ab22804SFenghua Yu #define XFEATURE_MASK_SUPERVISOR_ALL (XFEATURE_MASK_SUPERVISOR_SUPPORTED | \ 8101707b66SAndy Lutomirski XFEATURE_MASK_INDEPENDENT | \ 828ab22804SFenghua Yu XFEATURE_MASK_SUPERVISOR_UNSUPPORTED) 838ab22804SFenghua Yu 84eda32f4fSThomas Gleixner /* 85eda32f4fSThomas Gleixner * The feature mask required to restore FPU state: 86eda32f4fSThomas Gleixner * - All user states which are not eagerly switched in switch_to()/exec() 87eda32f4fSThomas Gleixner * - The suporvisor states 88eda32f4fSThomas Gleixner */ 89eda32f4fSThomas Gleixner #define XFEATURE_MASK_FPSTATE (XFEATURE_MASK_USER_RESTORE | \ 90eda32f4fSThomas Gleixner XFEATURE_MASK_SUPERVISOR_SUPPORTED) 91eda32f4fSThomas Gleixner 92669ebabbSIngo Molnar extern u64 xstate_fx_sw_bytes[USER_XSTATE_FX_SW_WORDS]; 93669ebabbSIngo Molnar 94404f6aacSKees Cook extern void __init update_regset_xstate_info(unsigned int size, 95404f6aacSKees Cook u64 xstate_mask); 96669ebabbSIngo Molnar 97ce711ea3SKan Liang int xfeature_size(int xfeature_nr); 98a75c5289SThomas Gleixner 99a75c5289SThomas Gleixner void xsaves(struct xregs_state *xsave, u64 mask); 100a75c5289SThomas Gleixner void xrstors(struct xregs_state *xsave, u64 mask); 10150f408d9SKan Liang 102*23686ef2SThomas Gleixner #ifdef CONFIG_X86_64 103*23686ef2SThomas Gleixner DECLARE_STATIC_KEY_FALSE(__fpu_state_size_dynamic); 104*23686ef2SThomas Gleixner #endif 105*23686ef2SThomas Gleixner 106*23686ef2SThomas Gleixner #ifdef CONFIG_X86_64 107*23686ef2SThomas Gleixner DECLARE_STATIC_KEY_FALSE(__fpu_state_size_dynamic); 108*23686ef2SThomas Gleixner 109*23686ef2SThomas Gleixner static __always_inline __pure bool fpu_state_size_dynamic(void) 110*23686ef2SThomas Gleixner { 111*23686ef2SThomas Gleixner return static_branch_unlikely(&__fpu_state_size_dynamic); 112*23686ef2SThomas Gleixner } 113*23686ef2SThomas Gleixner #else 114*23686ef2SThomas Gleixner static __always_inline __pure bool fpu_state_size_dynamic(void) 115*23686ef2SThomas Gleixner { 116*23686ef2SThomas Gleixner return false; 117*23686ef2SThomas Gleixner } 118*23686ef2SThomas Gleixner #endif 119*23686ef2SThomas Gleixner 120669ebabbSIngo Molnar #endif 121