1 #include <linux/threads.h> 2 #include <linux/cpumask.h> 3 #include <linux/string.h> 4 #include <linux/kernel.h> 5 #include <linux/ctype.h> 6 #include <linux/dmar.h> 7 8 #include <asm/smp.h> 9 #include <asm/x2apic.h> 10 11 int x2apic_phys; 12 13 static struct apic apic_x2apic_phys; 14 15 static int set_x2apic_phys_mode(char *arg) 16 { 17 x2apic_phys = 1; 18 return 0; 19 } 20 early_param("x2apic_phys", set_x2apic_phys_mode); 21 22 static bool x2apic_fadt_phys(void) 23 { 24 #ifdef CONFIG_ACPI 25 if ((acpi_gbl_FADT.header.revision >= FADT2_REVISION_ID) && 26 (acpi_gbl_FADT.flags & ACPI_FADT_APIC_PHYSICAL)) { 27 printk(KERN_DEBUG "System requires x2apic physical mode\n"); 28 return true; 29 } 30 #endif 31 return false; 32 } 33 34 static int x2apic_acpi_madt_oem_check(char *oem_id, char *oem_table_id) 35 { 36 return x2apic_enabled() && (x2apic_phys || x2apic_fadt_phys()); 37 } 38 39 static void x2apic_send_IPI(int cpu, int vector) 40 { 41 u32 dest = per_cpu(x86_cpu_to_apicid, cpu); 42 43 x2apic_wrmsr_fence(); 44 __x2apic_send_IPI_dest(dest, vector, APIC_DEST_PHYSICAL); 45 } 46 47 static void 48 __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) 49 { 50 unsigned long query_cpu; 51 unsigned long this_cpu; 52 unsigned long flags; 53 54 x2apic_wrmsr_fence(); 55 56 local_irq_save(flags); 57 58 this_cpu = smp_processor_id(); 59 for_each_cpu(query_cpu, mask) { 60 if (apic_dest == APIC_DEST_ALLBUT && this_cpu == query_cpu) 61 continue; 62 __x2apic_send_IPI_dest(per_cpu(x86_cpu_to_apicid, query_cpu), 63 vector, APIC_DEST_PHYSICAL); 64 } 65 local_irq_restore(flags); 66 } 67 68 static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector) 69 { 70 __x2apic_send_IPI_mask(mask, vector, APIC_DEST_ALLINC); 71 } 72 73 static void 74 x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector) 75 { 76 __x2apic_send_IPI_mask(mask, vector, APIC_DEST_ALLBUT); 77 } 78 79 static void x2apic_send_IPI_allbutself(int vector) 80 { 81 __x2apic_send_IPI_mask(cpu_online_mask, vector, APIC_DEST_ALLBUT); 82 } 83 84 static void x2apic_send_IPI_all(int vector) 85 { 86 __x2apic_send_IPI_mask(cpu_online_mask, vector, APIC_DEST_ALLINC); 87 } 88 89 static void init_x2apic_ldr(void) 90 { 91 } 92 93 static int x2apic_phys_probe(void) 94 { 95 if (x2apic_mode && (x2apic_phys || x2apic_fadt_phys())) 96 return 1; 97 98 return apic == &apic_x2apic_phys; 99 } 100 101 static struct apic apic_x2apic_phys __ro_after_init = { 102 103 .name = "physical x2apic", 104 .probe = x2apic_phys_probe, 105 .acpi_madt_oem_check = x2apic_acpi_madt_oem_check, 106 .apic_id_valid = x2apic_apic_id_valid, 107 .apic_id_registered = x2apic_apic_id_registered, 108 109 .irq_delivery_mode = dest_Fixed, 110 .irq_dest_mode = 0, /* physical */ 111 112 .target_cpus = online_target_cpus, 113 .disable_esr = 0, 114 .dest_logical = 0, 115 .check_apicid_used = NULL, 116 117 .vector_allocation_domain = default_vector_allocation_domain, 118 .init_apic_ldr = init_x2apic_ldr, 119 120 .ioapic_phys_id_map = NULL, 121 .setup_apic_routing = NULL, 122 .cpu_present_to_apicid = default_cpu_present_to_apicid, 123 .apicid_to_cpu_present = NULL, 124 .check_phys_apicid_present = default_check_phys_apicid_present, 125 .phys_pkg_id = x2apic_phys_pkg_id, 126 127 .get_apic_id = x2apic_get_apic_id, 128 .set_apic_id = x2apic_set_apic_id, 129 130 .cpu_mask_to_apicid_and = default_cpu_mask_to_apicid_and, 131 132 .send_IPI = x2apic_send_IPI, 133 .send_IPI_mask = x2apic_send_IPI_mask, 134 .send_IPI_mask_allbutself = x2apic_send_IPI_mask_allbutself, 135 .send_IPI_allbutself = x2apic_send_IPI_allbutself, 136 .send_IPI_all = x2apic_send_IPI_all, 137 .send_IPI_self = x2apic_send_IPI_self, 138 139 .inquire_remote_apic = NULL, 140 141 .read = native_apic_msr_read, 142 .write = native_apic_msr_write, 143 .eoi_write = native_apic_msr_eoi_write, 144 .icr_read = native_x2apic_icr_read, 145 .icr_write = native_x2apic_icr_write, 146 .wait_icr_idle = native_x2apic_wait_icr_idle, 147 .safe_wait_icr_idle = native_safe_x2apic_wait_icr_idle, 148 }; 149 150 apic_driver(apic_x2apic_phys); 151