1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 2e08cae41SH. Peter Anvin #ifndef _ASM_X86_MSHYPER_H 3e08cae41SH. Peter Anvin #define _ASM_X86_MSHYPER_H 4a2a47c6cSKy Srinivasan 5e08cae41SH. Peter Anvin #include <linux/types.h> 626fcd952SThomas Gleixner #include <linux/atomic.h> 7806c8927SVitaly Kuznetsov #include <linux/nmi.h> 8fc53662fSVitaly Kuznetsov #include <asm/io.h> 95a485803SVitaly Kuznetsov #include <asm/hyperv-tlfs.h> 10e70e5892SDavid Woodhouse #include <asm/nospec-branch.h> 11e08cae41SH. Peter Anvin 121268ed0cSK. Y. Srinivasan #define VP_INVAL U32_MAX 131268ed0cSK. Y. Srinivasan 14e08cae41SH. Peter Anvin struct ms_hyperv_info { 15e08cae41SH. Peter Anvin u32 features; 16cc2dd402SDenis V. Lunev u32 misc_features; 17e08cae41SH. Peter Anvin u32 hints; 185431390bSVitaly Kuznetsov u32 nested_features; 19dd018597SVitaly Kuznetsov u32 max_vp_index; 20dd018597SVitaly Kuznetsov u32 max_lp_index; 21e08cae41SH. Peter Anvin }; 22e08cae41SH. Peter Anvin 23e08cae41SH. Peter Anvin extern struct ms_hyperv_info ms_hyperv; 24a2a47c6cSKy Srinivasan 25*cc4edae4SLan Tianyu 26*cc4edae4SLan Tianyu typedef int (*hyperv_fill_flush_list_func)( 27*cc4edae4SLan Tianyu struct hv_guest_mapping_flush_list *flush, 28*cc4edae4SLan Tianyu void *data); 29*cc4edae4SLan Tianyu 30352c9624SK. Y. Srinivasan /* 31415bd1cdSVitaly Kuznetsov * Generate the guest ID. 32352c9624SK. Y. Srinivasan */ 33352c9624SK. Y. Srinivasan 34352c9624SK. Y. Srinivasan static inline __u64 generate_guest_id(__u64 d_info1, __u64 kernel_version, 35352c9624SK. Y. Srinivasan __u64 d_info2) 36352c9624SK. Y. Srinivasan { 37352c9624SK. Y. Srinivasan __u64 guest_id = 0; 38352c9624SK. Y. Srinivasan 399b06e101SK. Y. Srinivasan guest_id = (((__u64)HV_LINUX_VENDOR_ID) << 48); 40352c9624SK. Y. Srinivasan guest_id |= (d_info1 << 48); 41352c9624SK. Y. Srinivasan guest_id |= (kernel_version << 16); 42352c9624SK. Y. Srinivasan guest_id |= d_info2; 43352c9624SK. Y. Srinivasan 44352c9624SK. Y. Srinivasan return guest_id; 45352c9624SK. Y. Srinivasan } 46352c9624SK. Y. Srinivasan 47e810e48cSK. Y. Srinivasan 48e810e48cSK. Y. Srinivasan /* Free the message slot and signal end-of-message if required */ 49e810e48cSK. Y. Srinivasan static inline void vmbus_signal_eom(struct hv_message *msg, u32 old_msg_type) 50e810e48cSK. Y. Srinivasan { 51e810e48cSK. Y. Srinivasan /* 52e810e48cSK. Y. Srinivasan * On crash we're reading some other CPU's message page and we need 53e810e48cSK. Y. Srinivasan * to be careful: this other CPU may already had cleared the header 54e810e48cSK. Y. Srinivasan * and the host may already had delivered some other message there. 55e810e48cSK. Y. Srinivasan * In case we blindly write msg->header.message_type we're going 56e810e48cSK. Y. Srinivasan * to lose it. We can still lose a message of the same type but 57e810e48cSK. Y. Srinivasan * we count on the fact that there can only be one 58e810e48cSK. Y. Srinivasan * CHANNELMSG_UNLOAD_RESPONSE and we don't care about other messages 59e810e48cSK. Y. Srinivasan * on crash. 60e810e48cSK. Y. Srinivasan */ 61e810e48cSK. Y. Srinivasan if (cmpxchg(&msg->header.message_type, old_msg_type, 62e810e48cSK. Y. Srinivasan HVMSG_NONE) != old_msg_type) 63e810e48cSK. Y. Srinivasan return; 64e810e48cSK. Y. Srinivasan 65e810e48cSK. Y. Srinivasan /* 66e810e48cSK. Y. Srinivasan * Make sure the write to MessageType (ie set to 67e810e48cSK. Y. Srinivasan * HVMSG_NONE) happens before we read the 68e810e48cSK. Y. Srinivasan * MessagePending and EOMing. Otherwise, the EOMing 69e810e48cSK. Y. Srinivasan * will not deliver any more messages since there is 70e810e48cSK. Y. Srinivasan * no empty slot 71e810e48cSK. Y. Srinivasan */ 72e810e48cSK. Y. Srinivasan mb(); 73e810e48cSK. Y. Srinivasan 74e810e48cSK. Y. Srinivasan if (msg->header.message_flags.msg_pending) { 75e810e48cSK. Y. Srinivasan /* 76e810e48cSK. Y. Srinivasan * This will cause message queue rescan to 77e810e48cSK. Y. Srinivasan * possibly deliver another msg from the 78e810e48cSK. Y. Srinivasan * hypervisor 79e810e48cSK. Y. Srinivasan */ 80e810e48cSK. Y. Srinivasan wrmsrl(HV_X64_MSR_EOM, 0); 81e810e48cSK. Y. Srinivasan } 82e810e48cSK. Y. Srinivasan } 83e810e48cSK. Y. Srinivasan 84619a4c8bSMichael Kelley #define hv_init_timer(timer, tick) \ 85619a4c8bSMichael Kelley wrmsrl(HV_X64_MSR_STIMER0_COUNT + (2*timer), tick) 86619a4c8bSMichael Kelley #define hv_init_timer_config(timer, val) \ 87619a4c8bSMichael Kelley wrmsrl(HV_X64_MSR_STIMER0_CONFIG + (2*timer), val) 88d5116b40SK. Y. Srinivasan 89155e4a2fSK. Y. Srinivasan #define hv_get_simp(val) rdmsrl(HV_X64_MSR_SIMP, val) 90155e4a2fSK. Y. Srinivasan #define hv_set_simp(val) wrmsrl(HV_X64_MSR_SIMP, val) 91155e4a2fSK. Y. Srinivasan 928e307bf8SK. Y. Srinivasan #define hv_get_siefp(val) rdmsrl(HV_X64_MSR_SIEFP, val) 938e307bf8SK. Y. Srinivasan #define hv_set_siefp(val) wrmsrl(HV_X64_MSR_SIEFP, val) 948e307bf8SK. Y. Srinivasan 9506d1d98aSK. Y. Srinivasan #define hv_get_synic_state(val) rdmsrl(HV_X64_MSR_SCONTROL, val) 9606d1d98aSK. Y. Srinivasan #define hv_set_synic_state(val) wrmsrl(HV_X64_MSR_SCONTROL, val) 9706d1d98aSK. Y. Srinivasan 987297ff0cSK. Y. Srinivasan #define hv_get_vp_index(index) rdmsrl(HV_X64_MSR_VP_INDEX, index) 997297ff0cSK. Y. Srinivasan 100619a4c8bSMichael Kelley #define hv_get_synint_state(int_num, val) \ 101619a4c8bSMichael Kelley rdmsrl(HV_X64_MSR_SINT0 + int_num, val) 102619a4c8bSMichael Kelley #define hv_set_synint_state(int_num, val) \ 103619a4c8bSMichael Kelley wrmsrl(HV_X64_MSR_SINT0 + int_num, val) 10437e11d5cSK. Y. Srinivasan 1059d9c9656SSunil Muthuswamy #define hv_get_crash_ctl(val) \ 1069d9c9656SSunil Muthuswamy rdmsrl(HV_X64_MSR_CRASH_CTL, val) 107cf910e83SSeiji Aguchi 108cf910e83SSeiji Aguchi void hyperv_callback_vector(void); 10993286261SVitaly Kuznetsov void hyperv_reenlightenment_vector(void); 110bc2b0331SK. Y. Srinivasan #ifdef CONFIG_TRACING 111bc2b0331SK. Y. Srinivasan #define trace_hyperv_callback_vector hyperv_callback_vector 112bc2b0331SK. Y. Srinivasan #endif 113a2a47c6cSKy Srinivasan void hyperv_vector_handler(struct pt_regs *regs); 11476d388cdSThomas Gleixner void hv_setup_vmbus_irq(void (*handler)(void)); 11576d388cdSThomas Gleixner void hv_remove_vmbus_irq(void); 116a2a47c6cSKy Srinivasan 1172517281dSVitaly Kuznetsov void hv_setup_kexec_handler(void (*handler)(void)); 1182517281dSVitaly Kuznetsov void hv_remove_kexec_handler(void); 119b4370df2SVitaly Kuznetsov void hv_setup_crash_handler(void (*handler)(struct pt_regs *regs)); 120b4370df2SVitaly Kuznetsov void hv_remove_crash_handler(void); 1218730046cSK. Y. Srinivasan 122248e742aSMichael Kelley /* 123248e742aSMichael Kelley * Routines for stimer0 Direct Mode handling. 124248e742aSMichael Kelley * On x86/x64, there are no percpu actions to take. 125248e742aSMichael Kelley */ 126248e742aSMichael Kelley void hv_stimer0_vector_handler(struct pt_regs *regs); 127248e742aSMichael Kelley void hv_stimer0_callback_vector(void); 128248e742aSMichael Kelley int hv_setup_stimer0_irq(int *irq, int *vector, void (*handler)(void)); 129248e742aSMichael Kelley void hv_remove_stimer0_irq(int irq); 130248e742aSMichael Kelley 131248e742aSMichael Kelley static inline void hv_enable_stimer0_percpu_irq(int irq) {} 132248e742aSMichael Kelley static inline void hv_disable_stimer0_percpu_irq(int irq) {} 133248e742aSMichael Kelley 134248e742aSMichael Kelley 1358730046cSK. Y. Srinivasan #if IS_ENABLED(CONFIG_HYPERV) 136dee863b5SVitaly Kuznetsov extern struct clocksource *hyperv_cs; 137fc53662fSVitaly Kuznetsov extern void *hv_hypercall_pg; 13868bb7bfbSK. Y. Srinivasan extern void __percpu **hyperv_pcpu_input_arg; 139fc53662fSVitaly Kuznetsov 140fc53662fSVitaly Kuznetsov static inline u64 hv_do_hypercall(u64 control, void *input, void *output) 141fc53662fSVitaly Kuznetsov { 142fc53662fSVitaly Kuznetsov u64 input_address = input ? virt_to_phys(input) : 0; 143fc53662fSVitaly Kuznetsov u64 output_address = output ? virt_to_phys(output) : 0; 144fc53662fSVitaly Kuznetsov u64 hv_status; 145fc53662fSVitaly Kuznetsov 146fc53662fSVitaly Kuznetsov #ifdef CONFIG_X86_64 147fc53662fSVitaly Kuznetsov if (!hv_hypercall_pg) 148fc53662fSVitaly Kuznetsov return U64_MAX; 149fc53662fSVitaly Kuznetsov 150fc53662fSVitaly Kuznetsov __asm__ __volatile__("mov %4, %%r8\n" 151e70e5892SDavid Woodhouse CALL_NOSPEC 152f5caf621SJosh Poimboeuf : "=a" (hv_status), ASM_CALL_CONSTRAINT, 153fc53662fSVitaly Kuznetsov "+c" (control), "+d" (input_address) 154e70e5892SDavid Woodhouse : "r" (output_address), 155e70e5892SDavid Woodhouse THUNK_TARGET(hv_hypercall_pg) 156fc53662fSVitaly Kuznetsov : "cc", "memory", "r8", "r9", "r10", "r11"); 157fc53662fSVitaly Kuznetsov #else 158fc53662fSVitaly Kuznetsov u32 input_address_hi = upper_32_bits(input_address); 159fc53662fSVitaly Kuznetsov u32 input_address_lo = lower_32_bits(input_address); 160fc53662fSVitaly Kuznetsov u32 output_address_hi = upper_32_bits(output_address); 161fc53662fSVitaly Kuznetsov u32 output_address_lo = lower_32_bits(output_address); 162fc53662fSVitaly Kuznetsov 163fc53662fSVitaly Kuznetsov if (!hv_hypercall_pg) 164fc53662fSVitaly Kuznetsov return U64_MAX; 165fc53662fSVitaly Kuznetsov 166e70e5892SDavid Woodhouse __asm__ __volatile__(CALL_NOSPEC 167fc53662fSVitaly Kuznetsov : "=A" (hv_status), 168f5caf621SJosh Poimboeuf "+c" (input_address_lo), ASM_CALL_CONSTRAINT 169fc53662fSVitaly Kuznetsov : "A" (control), 170fc53662fSVitaly Kuznetsov "b" (input_address_hi), 171fc53662fSVitaly Kuznetsov "D"(output_address_hi), "S"(output_address_lo), 172e70e5892SDavid Woodhouse THUNK_TARGET(hv_hypercall_pg) 173fc53662fSVitaly Kuznetsov : "cc", "memory"); 174fc53662fSVitaly Kuznetsov #endif /* !x86_64 */ 175fc53662fSVitaly Kuznetsov return hv_status; 176fc53662fSVitaly Kuznetsov } 177dee863b5SVitaly Kuznetsov 1786a8edbd0SVitaly Kuznetsov /* Fast hypercall with 8 bytes of input and no output */ 1796a8edbd0SVitaly Kuznetsov static inline u64 hv_do_fast_hypercall8(u16 code, u64 input1) 1806a8edbd0SVitaly Kuznetsov { 1816a8edbd0SVitaly Kuznetsov u64 hv_status, control = (u64)code | HV_HYPERCALL_FAST_BIT; 1826a8edbd0SVitaly Kuznetsov 1836a8edbd0SVitaly Kuznetsov #ifdef CONFIG_X86_64 1846a8edbd0SVitaly Kuznetsov { 185e70e5892SDavid Woodhouse __asm__ __volatile__(CALL_NOSPEC 186f5caf621SJosh Poimboeuf : "=a" (hv_status), ASM_CALL_CONSTRAINT, 1876a8edbd0SVitaly Kuznetsov "+c" (control), "+d" (input1) 188e70e5892SDavid Woodhouse : THUNK_TARGET(hv_hypercall_pg) 1896a8edbd0SVitaly Kuznetsov : "cc", "r8", "r9", "r10", "r11"); 1906a8edbd0SVitaly Kuznetsov } 1916a8edbd0SVitaly Kuznetsov #else 1926a8edbd0SVitaly Kuznetsov { 1936a8edbd0SVitaly Kuznetsov u32 input1_hi = upper_32_bits(input1); 1946a8edbd0SVitaly Kuznetsov u32 input1_lo = lower_32_bits(input1); 1956a8edbd0SVitaly Kuznetsov 196e70e5892SDavid Woodhouse __asm__ __volatile__ (CALL_NOSPEC 1976a8edbd0SVitaly Kuznetsov : "=A"(hv_status), 1986a8edbd0SVitaly Kuznetsov "+c"(input1_lo), 199f5caf621SJosh Poimboeuf ASM_CALL_CONSTRAINT 2006a8edbd0SVitaly Kuznetsov : "A" (control), 2016a8edbd0SVitaly Kuznetsov "b" (input1_hi), 202e70e5892SDavid Woodhouse THUNK_TARGET(hv_hypercall_pg) 2036a8edbd0SVitaly Kuznetsov : "cc", "edi", "esi"); 2046a8edbd0SVitaly Kuznetsov } 2056a8edbd0SVitaly Kuznetsov #endif 2066a8edbd0SVitaly Kuznetsov return hv_status; 2076a8edbd0SVitaly Kuznetsov } 2086a8edbd0SVitaly Kuznetsov 20953e52966SVitaly Kuznetsov /* Fast hypercall with 16 bytes of input */ 21053e52966SVitaly Kuznetsov static inline u64 hv_do_fast_hypercall16(u16 code, u64 input1, u64 input2) 21153e52966SVitaly Kuznetsov { 21253e52966SVitaly Kuznetsov u64 hv_status, control = (u64)code | HV_HYPERCALL_FAST_BIT; 21353e52966SVitaly Kuznetsov 21453e52966SVitaly Kuznetsov #ifdef CONFIG_X86_64 21553e52966SVitaly Kuznetsov { 21653e52966SVitaly Kuznetsov __asm__ __volatile__("mov %4, %%r8\n" 21753e52966SVitaly Kuznetsov CALL_NOSPEC 21853e52966SVitaly Kuznetsov : "=a" (hv_status), ASM_CALL_CONSTRAINT, 21953e52966SVitaly Kuznetsov "+c" (control), "+d" (input1) 22053e52966SVitaly Kuznetsov : "r" (input2), 22153e52966SVitaly Kuznetsov THUNK_TARGET(hv_hypercall_pg) 22253e52966SVitaly Kuznetsov : "cc", "r8", "r9", "r10", "r11"); 22353e52966SVitaly Kuznetsov } 22453e52966SVitaly Kuznetsov #else 22553e52966SVitaly Kuznetsov { 22653e52966SVitaly Kuznetsov u32 input1_hi = upper_32_bits(input1); 22753e52966SVitaly Kuznetsov u32 input1_lo = lower_32_bits(input1); 22853e52966SVitaly Kuznetsov u32 input2_hi = upper_32_bits(input2); 22953e52966SVitaly Kuznetsov u32 input2_lo = lower_32_bits(input2); 23053e52966SVitaly Kuznetsov 23153e52966SVitaly Kuznetsov __asm__ __volatile__ (CALL_NOSPEC 23253e52966SVitaly Kuznetsov : "=A"(hv_status), 23353e52966SVitaly Kuznetsov "+c"(input1_lo), ASM_CALL_CONSTRAINT 23453e52966SVitaly Kuznetsov : "A" (control), "b" (input1_hi), 23553e52966SVitaly Kuznetsov "D"(input2_hi), "S"(input2_lo), 23653e52966SVitaly Kuznetsov THUNK_TARGET(hv_hypercall_pg) 23753e52966SVitaly Kuznetsov : "cc"); 23853e52966SVitaly Kuznetsov } 23953e52966SVitaly Kuznetsov #endif 24053e52966SVitaly Kuznetsov return hv_status; 24153e52966SVitaly Kuznetsov } 24253e52966SVitaly Kuznetsov 243806c8927SVitaly Kuznetsov /* 244806c8927SVitaly Kuznetsov * Rep hypercalls. Callers of this functions are supposed to ensure that 245806c8927SVitaly Kuznetsov * rep_count and varhead_size comply with Hyper-V hypercall definition. 246806c8927SVitaly Kuznetsov */ 247806c8927SVitaly Kuznetsov static inline u64 hv_do_rep_hypercall(u16 code, u16 rep_count, u16 varhead_size, 248806c8927SVitaly Kuznetsov void *input, void *output) 249806c8927SVitaly Kuznetsov { 250806c8927SVitaly Kuznetsov u64 control = code; 251806c8927SVitaly Kuznetsov u64 status; 252806c8927SVitaly Kuznetsov u16 rep_comp; 253806c8927SVitaly Kuznetsov 254806c8927SVitaly Kuznetsov control |= (u64)varhead_size << HV_HYPERCALL_VARHEAD_OFFSET; 255806c8927SVitaly Kuznetsov control |= (u64)rep_count << HV_HYPERCALL_REP_COMP_OFFSET; 256806c8927SVitaly Kuznetsov 257806c8927SVitaly Kuznetsov do { 258806c8927SVitaly Kuznetsov status = hv_do_hypercall(control, input, output); 259806c8927SVitaly Kuznetsov if ((status & HV_HYPERCALL_RESULT_MASK) != HV_STATUS_SUCCESS) 260806c8927SVitaly Kuznetsov return status; 261806c8927SVitaly Kuznetsov 262806c8927SVitaly Kuznetsov /* Bits 32-43 of status have 'Reps completed' data. */ 263806c8927SVitaly Kuznetsov rep_comp = (status & HV_HYPERCALL_REP_COMP_MASK) >> 264806c8927SVitaly Kuznetsov HV_HYPERCALL_REP_COMP_OFFSET; 265806c8927SVitaly Kuznetsov 266806c8927SVitaly Kuznetsov control &= ~HV_HYPERCALL_REP_START_MASK; 267806c8927SVitaly Kuznetsov control |= (u64)rep_comp << HV_HYPERCALL_REP_START_OFFSET; 268806c8927SVitaly Kuznetsov 269806c8927SVitaly Kuznetsov touch_nmi_watchdog(); 270806c8927SVitaly Kuznetsov } while (rep_comp < rep_count); 271806c8927SVitaly Kuznetsov 272806c8927SVitaly Kuznetsov return status; 273806c8927SVitaly Kuznetsov } 274806c8927SVitaly Kuznetsov 2757415aea6SVitaly Kuznetsov /* 2767415aea6SVitaly Kuznetsov * Hypervisor's notion of virtual processor ID is different from 2777415aea6SVitaly Kuznetsov * Linux' notion of CPU ID. This information can only be retrieved 2787415aea6SVitaly Kuznetsov * in the context of the calling CPU. Setup a map for easy access 2797415aea6SVitaly Kuznetsov * to this information. 2807415aea6SVitaly Kuznetsov */ 2817415aea6SVitaly Kuznetsov extern u32 *hv_vp_index; 282a3b74243SVitaly Kuznetsov extern u32 hv_max_vp_index; 283a46d15ccSVitaly Kuznetsov extern struct hv_vp_assist_page **hv_vp_assist_page; 284a46d15ccSVitaly Kuznetsov 285a46d15ccSVitaly Kuznetsov static inline struct hv_vp_assist_page *hv_get_vp_assist_page(unsigned int cpu) 286a46d15ccSVitaly Kuznetsov { 287a46d15ccSVitaly Kuznetsov if (!hv_vp_assist_page) 288a46d15ccSVitaly Kuznetsov return NULL; 289a46d15ccSVitaly Kuznetsov 290a46d15ccSVitaly Kuznetsov return hv_vp_assist_page[cpu]; 291a46d15ccSVitaly Kuznetsov } 2927415aea6SVitaly Kuznetsov 2937415aea6SVitaly Kuznetsov /** 2947415aea6SVitaly Kuznetsov * hv_cpu_number_to_vp_number() - Map CPU to VP. 2957415aea6SVitaly Kuznetsov * @cpu_number: CPU number in Linux terms 2967415aea6SVitaly Kuznetsov * 2977415aea6SVitaly Kuznetsov * This function returns the mapping between the Linux processor 2987415aea6SVitaly Kuznetsov * number and the hypervisor's virtual processor number, useful 2997415aea6SVitaly Kuznetsov * in making hypercalls and such that talk about specific 3007415aea6SVitaly Kuznetsov * processors. 3017415aea6SVitaly Kuznetsov * 3027415aea6SVitaly Kuznetsov * Return: Virtual processor number in Hyper-V terms 3037415aea6SVitaly Kuznetsov */ 3047415aea6SVitaly Kuznetsov static inline int hv_cpu_number_to_vp_number(int cpu_number) 3057415aea6SVitaly Kuznetsov { 3067415aea6SVitaly Kuznetsov return hv_vp_index[cpu_number]; 3077415aea6SVitaly Kuznetsov } 30873638cddSK. Y. Srinivasan 309366f03b0SK. Y. Srinivasan static inline int cpumask_to_vpset(struct hv_vpset *vpset, 310366f03b0SK. Y. Srinivasan const struct cpumask *cpus) 311366f03b0SK. Y. Srinivasan { 312366f03b0SK. Y. Srinivasan int cpu, vcpu, vcpu_bank, vcpu_offset, nr_bank = 1; 313366f03b0SK. Y. Srinivasan 314366f03b0SK. Y. Srinivasan /* valid_bank_mask can represent up to 64 banks */ 315366f03b0SK. Y. Srinivasan if (hv_max_vp_index / 64 >= 64) 316366f03b0SK. Y. Srinivasan return 0; 317366f03b0SK. Y. Srinivasan 318366f03b0SK. Y. Srinivasan /* 319c9c92beeSVitaly Kuznetsov * Clear all banks up to the maximum possible bank as hv_tlb_flush_ex 320366f03b0SK. Y. Srinivasan * structs are not cleared between calls, we risk flushing unneeded 321366f03b0SK. Y. Srinivasan * vCPUs otherwise. 322366f03b0SK. Y. Srinivasan */ 323366f03b0SK. Y. Srinivasan for (vcpu_bank = 0; vcpu_bank <= hv_max_vp_index / 64; vcpu_bank++) 324366f03b0SK. Y. Srinivasan vpset->bank_contents[vcpu_bank] = 0; 325366f03b0SK. Y. Srinivasan 326366f03b0SK. Y. Srinivasan /* 327366f03b0SK. Y. Srinivasan * Some banks may end up being empty but this is acceptable. 328366f03b0SK. Y. Srinivasan */ 329366f03b0SK. Y. Srinivasan for_each_cpu(cpu, cpus) { 330366f03b0SK. Y. Srinivasan vcpu = hv_cpu_number_to_vp_number(cpu); 3311268ed0cSK. Y. Srinivasan if (vcpu == VP_INVAL) 3321268ed0cSK. Y. Srinivasan return -1; 333366f03b0SK. Y. Srinivasan vcpu_bank = vcpu / 64; 334366f03b0SK. Y. Srinivasan vcpu_offset = vcpu % 64; 335366f03b0SK. Y. Srinivasan __set_bit(vcpu_offset, (unsigned long *) 336366f03b0SK. Y. Srinivasan &vpset->bank_contents[vcpu_bank]); 337366f03b0SK. Y. Srinivasan if (vcpu_bank >= nr_bank) 338366f03b0SK. Y. Srinivasan nr_bank = vcpu_bank + 1; 339366f03b0SK. Y. Srinivasan } 340366f03b0SK. Y. Srinivasan vpset->valid_bank_mask = GENMASK_ULL(nr_bank - 1, 0); 341366f03b0SK. Y. Srinivasan return nr_bank; 342366f03b0SK. Y. Srinivasan } 343366f03b0SK. Y. Srinivasan 3446b48cb5fSK. Y. Srinivasan void __init hyperv_init(void); 3452ffd9e33SVitaly Kuznetsov void hyperv_setup_mmu_ops(void); 3467ed4325aSK. Y. Srinivasan void hyperv_report_panic(struct pt_regs *regs, long err); 34781b18bceSSunil Muthuswamy void hyperv_report_panic_msg(phys_addr_t pa, size_t size); 3484a5f3cdeSMichael Kelley bool hv_is_hyperv_initialized(void); 349d6f3609dSVitaly Kuznetsov void hyperv_cleanup(void); 35093286261SVitaly Kuznetsov 35193286261SVitaly Kuznetsov void hyperv_reenlightenment_intr(struct pt_regs *regs); 35293286261SVitaly Kuznetsov void set_hv_tscchange_cb(void (*cb)(void)); 35393286261SVitaly Kuznetsov void clear_hv_tscchange_cb(void); 35493286261SVitaly Kuznetsov void hyperv_stop_tsc_emulation(void); 355eb914cfeSTianyu Lan int hyperv_flush_guest_mapping(u64 as); 356*cc4edae4SLan Tianyu int hyperv_flush_guest_mapping_range(u64 as, 357*cc4edae4SLan Tianyu hyperv_fill_flush_list_func fill_func, void *data); 358*cc4edae4SLan Tianyu int hyperv_fill_flush_guest_mapping_list( 359*cc4edae4SLan Tianyu struct hv_guest_mapping_flush_list *flush, 360*cc4edae4SLan Tianyu u64 start_gfn, u64 end_gfn); 3612d2ccf24SThomas Gleixner 3622d2ccf24SThomas Gleixner #ifdef CONFIG_X86_64 3636b48cb5fSK. Y. Srinivasan void hv_apic_init(void); 3643a025de6SYi Sun void __init hv_init_spinlocks(void); 3653a025de6SYi Sun bool hv_vcpu_is_preempted(int vcpu); 3662d2ccf24SThomas Gleixner #else 3672d2ccf24SThomas Gleixner static inline void hv_apic_init(void) {} 3682d2ccf24SThomas Gleixner #endif 3692d2ccf24SThomas Gleixner 37079cadff2SVitaly Kuznetsov #else /* CONFIG_HYPERV */ 37179cadff2SVitaly Kuznetsov static inline void hyperv_init(void) {} 3724a5f3cdeSMichael Kelley static inline bool hv_is_hyperv_initialized(void) { return false; } 37379cadff2SVitaly Kuznetsov static inline void hyperv_cleanup(void) {} 3742ffd9e33SVitaly Kuznetsov static inline void hyperv_setup_mmu_ops(void) {} 37593286261SVitaly Kuznetsov static inline void set_hv_tscchange_cb(void (*cb)(void)) {} 37693286261SVitaly Kuznetsov static inline void clear_hv_tscchange_cb(void) {} 37793286261SVitaly Kuznetsov static inline void hyperv_stop_tsc_emulation(void) {}; 378a46d15ccSVitaly Kuznetsov static inline struct hv_vp_assist_page *hv_get_vp_assist_page(unsigned int cpu) 379a46d15ccSVitaly Kuznetsov { 380a46d15ccSVitaly Kuznetsov return NULL; 381a46d15ccSVitaly Kuznetsov } 382eb914cfeSTianyu Lan static inline int hyperv_flush_guest_mapping(u64 as) { return -1; } 383*cc4edae4SLan Tianyu static inline int hyperv_flush_guest_mapping_range(u64 as, 384*cc4edae4SLan Tianyu hyperv_fill_flush_list_func fill_func, void *data) 385*cc4edae4SLan Tianyu { 386*cc4edae4SLan Tianyu return -1; 387*cc4edae4SLan Tianyu } 38879cadff2SVitaly Kuznetsov #endif /* CONFIG_HYPERV */ 38979cadff2SVitaly Kuznetsov 390bd2a9adaSVitaly Kuznetsov #ifdef CONFIG_HYPERV_TSCPAGE 391bd2a9adaSVitaly Kuznetsov struct ms_hyperv_tsc_page *hv_get_tsc_page(void); 392e2768eaaSVitaly Kuznetsov static inline u64 hv_read_tsc_page_tsc(const struct ms_hyperv_tsc_page *tsc_pg, 393e2768eaaSVitaly Kuznetsov u64 *cur_tsc) 3940733379bSVitaly Kuznetsov { 395e2768eaaSVitaly Kuznetsov u64 scale, offset; 3960733379bSVitaly Kuznetsov u32 sequence; 3970733379bSVitaly Kuznetsov 3980733379bSVitaly Kuznetsov /* 3990733379bSVitaly Kuznetsov * The protocol for reading Hyper-V TSC page is specified in Hypervisor 4000733379bSVitaly Kuznetsov * Top-Level Functional Specification ver. 3.0 and above. To get the 4010733379bSVitaly Kuznetsov * reference time we must do the following: 4020733379bSVitaly Kuznetsov * - READ ReferenceTscSequence 4030733379bSVitaly Kuznetsov * A special '0' value indicates the time source is unreliable and we 4040733379bSVitaly Kuznetsov * need to use something else. The currently published specification 4050733379bSVitaly Kuznetsov * versions (up to 4.0b) contain a mistake and wrongly claim '-1' 4060733379bSVitaly Kuznetsov * instead of '0' as the special value, see commit c35b82ef0294. 4070733379bSVitaly Kuznetsov * - ReferenceTime = 4080733379bSVitaly Kuznetsov * ((RDTSC() * ReferenceTscScale) >> 64) + ReferenceTscOffset 4090733379bSVitaly Kuznetsov * - READ ReferenceTscSequence again. In case its value has changed 4100733379bSVitaly Kuznetsov * since our first reading we need to discard ReferenceTime and repeat 4110733379bSVitaly Kuznetsov * the whole sequence as the hypervisor was updating the page in 4120733379bSVitaly Kuznetsov * between. 4130733379bSVitaly Kuznetsov */ 4140733379bSVitaly Kuznetsov do { 4150733379bSVitaly Kuznetsov sequence = READ_ONCE(tsc_pg->tsc_sequence); 4160733379bSVitaly Kuznetsov if (!sequence) 4170733379bSVitaly Kuznetsov return U64_MAX; 4180733379bSVitaly Kuznetsov /* 4190733379bSVitaly Kuznetsov * Make sure we read sequence before we read other values from 4200733379bSVitaly Kuznetsov * TSC page. 4210733379bSVitaly Kuznetsov */ 4220733379bSVitaly Kuznetsov smp_rmb(); 4230733379bSVitaly Kuznetsov 4240733379bSVitaly Kuznetsov scale = READ_ONCE(tsc_pg->tsc_scale); 4250733379bSVitaly Kuznetsov offset = READ_ONCE(tsc_pg->tsc_offset); 426e2768eaaSVitaly Kuznetsov *cur_tsc = rdtsc_ordered(); 4270733379bSVitaly Kuznetsov 4280733379bSVitaly Kuznetsov /* 4290733379bSVitaly Kuznetsov * Make sure we read sequence after we read all other values 4300733379bSVitaly Kuznetsov * from TSC page. 4310733379bSVitaly Kuznetsov */ 4320733379bSVitaly Kuznetsov smp_rmb(); 4330733379bSVitaly Kuznetsov 4340733379bSVitaly Kuznetsov } while (READ_ONCE(tsc_pg->tsc_sequence) != sequence); 4350733379bSVitaly Kuznetsov 436e2768eaaSVitaly Kuznetsov return mul_u64_u64_shr(*cur_tsc, scale, 64) + offset; 437e2768eaaSVitaly Kuznetsov } 438e2768eaaSVitaly Kuznetsov 439e2768eaaSVitaly Kuznetsov static inline u64 hv_read_tsc_page(const struct ms_hyperv_tsc_page *tsc_pg) 440e2768eaaSVitaly Kuznetsov { 441e2768eaaSVitaly Kuznetsov u64 cur_tsc; 442e2768eaaSVitaly Kuznetsov 443e2768eaaSVitaly Kuznetsov return hv_read_tsc_page_tsc(tsc_pg, &cur_tsc); 4440733379bSVitaly Kuznetsov } 4450733379bSVitaly Kuznetsov 446bd2a9adaSVitaly Kuznetsov #else 447bd2a9adaSVitaly Kuznetsov static inline struct ms_hyperv_tsc_page *hv_get_tsc_page(void) 448bd2a9adaSVitaly Kuznetsov { 449bd2a9adaSVitaly Kuznetsov return NULL; 450bd2a9adaSVitaly Kuznetsov } 451e2768eaaSVitaly Kuznetsov 452e2768eaaSVitaly Kuznetsov static inline u64 hv_read_tsc_page_tsc(const struct ms_hyperv_tsc_page *tsc_pg, 453e2768eaaSVitaly Kuznetsov u64 *cur_tsc) 454e2768eaaSVitaly Kuznetsov { 455e2768eaaSVitaly Kuznetsov BUG(); 456e2768eaaSVitaly Kuznetsov return U64_MAX; 457e2768eaaSVitaly Kuznetsov } 458bd2a9adaSVitaly Kuznetsov #endif 459a2a47c6cSKy Srinivasan #endif 460