11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * SMP boot-related support 31da177e4SLinus Torvalds * 482975115SDavid Mosberger-Tang * Copyright (C) 1998-2003, 2005 Hewlett-Packard Co 51da177e4SLinus Torvalds * David Mosberger-Tang <davidm@hpl.hp.com> 6e927ecb0SSuresh Siddha * Copyright (C) 2001, 2004-2005 Intel Corp 7e927ecb0SSuresh Siddha * Rohit Seth <rohit.seth@intel.com> 8e927ecb0SSuresh Siddha * Suresh Siddha <suresh.b.siddha@intel.com> 9e927ecb0SSuresh Siddha * Gordon Jin <gordon.jin@intel.com> 10e927ecb0SSuresh Siddha * Ashok Raj <ashok.raj@intel.com> 111da177e4SLinus Torvalds * 121da177e4SLinus Torvalds * 01/05/16 Rohit Seth <rohit.seth@intel.com> Moved SMP booting functions from smp.c to here. 131da177e4SLinus Torvalds * 01/04/27 David Mosberger <davidm@hpl.hp.com> Added ITC synching code. 141da177e4SLinus Torvalds * 02/07/31 David Mosberger <davidm@hpl.hp.com> Switch over to hotplug-CPU boot-sequence. 151da177e4SLinus Torvalds * smp_boot_cpus()/smp_commence() is replaced by 161da177e4SLinus Torvalds * smp_prepare_cpus()/__cpu_up()/smp_cpus_done(). 17b8d8b883SAshok Raj * 04/06/21 Ashok Raj <ashok.raj@intel.com> Added CPU Hotplug Support 18e927ecb0SSuresh Siddha * 04/12/26 Jin Gordon <gordon.jin@intel.com> 19e927ecb0SSuresh Siddha * 04/12/26 Rohit Seth <rohit.seth@intel.com> 20e927ecb0SSuresh Siddha * Add multi-threading and multi-core detection 21e927ecb0SSuresh Siddha * 05/01/30 Suresh Siddha <suresh.b.siddha@intel.com> 22e927ecb0SSuresh Siddha * Setup cpu_sibling_map and cpu_core_map 231da177e4SLinus Torvalds */ 241da177e4SLinus Torvalds 251da177e4SLinus Torvalds #include <linux/module.h> 261da177e4SLinus Torvalds #include <linux/acpi.h> 271da177e4SLinus Torvalds #include <linux/bootmem.h> 281da177e4SLinus Torvalds #include <linux/cpu.h> 291da177e4SLinus Torvalds #include <linux/delay.h> 301da177e4SLinus Torvalds #include <linux/init.h> 311da177e4SLinus Torvalds #include <linux/interrupt.h> 321da177e4SLinus Torvalds #include <linux/irq.h> 331da177e4SLinus Torvalds #include <linux/kernel.h> 341da177e4SLinus Torvalds #include <linux/kernel_stat.h> 351da177e4SLinus Torvalds #include <linux/mm.h> 361da177e4SLinus Torvalds #include <linux/notifier.h> 371da177e4SLinus Torvalds #include <linux/smp.h> 381da177e4SLinus Torvalds #include <linux/spinlock.h> 391da177e4SLinus Torvalds #include <linux/efi.h> 401da177e4SLinus Torvalds #include <linux/percpu.h> 411da177e4SLinus Torvalds #include <linux/bitops.h> 421da177e4SLinus Torvalds 431da177e4SLinus Torvalds #include <asm/atomic.h> 441da177e4SLinus Torvalds #include <asm/cache.h> 451da177e4SLinus Torvalds #include <asm/current.h> 461da177e4SLinus Torvalds #include <asm/delay.h> 471da177e4SLinus Torvalds #include <asm/ia32.h> 481da177e4SLinus Torvalds #include <asm/io.h> 491da177e4SLinus Torvalds #include <asm/irq.h> 501da177e4SLinus Torvalds #include <asm/machvec.h> 511da177e4SLinus Torvalds #include <asm/mca.h> 521da177e4SLinus Torvalds #include <asm/page.h> 531da177e4SLinus Torvalds #include <asm/pgalloc.h> 541da177e4SLinus Torvalds #include <asm/pgtable.h> 551da177e4SLinus Torvalds #include <asm/processor.h> 561da177e4SLinus Torvalds #include <asm/ptrace.h> 571da177e4SLinus Torvalds #include <asm/sal.h> 581da177e4SLinus Torvalds #include <asm/system.h> 591da177e4SLinus Torvalds #include <asm/tlbflush.h> 601da177e4SLinus Torvalds #include <asm/unistd.h> 611da177e4SLinus Torvalds 621da177e4SLinus Torvalds #define SMP_DEBUG 0 631da177e4SLinus Torvalds 641da177e4SLinus Torvalds #if SMP_DEBUG 651da177e4SLinus Torvalds #define Dprintk(x...) printk(x) 661da177e4SLinus Torvalds #else 671da177e4SLinus Torvalds #define Dprintk(x...) 681da177e4SLinus Torvalds #endif 691da177e4SLinus Torvalds 70b8d8b883SAshok Raj #ifdef CONFIG_HOTPLUG_CPU 71ff741906SAshok Raj #ifdef CONFIG_PERMIT_BSP_REMOVE 72ff741906SAshok Raj #define bsp_remove_ok 1 73ff741906SAshok Raj #else 74ff741906SAshok Raj #define bsp_remove_ok 0 75ff741906SAshok Raj #endif 76ff741906SAshok Raj 77b8d8b883SAshok Raj /* 78b8d8b883SAshok Raj * Store all idle threads, this can be reused instead of creating 79b8d8b883SAshok Raj * a new thread. Also avoids complicated thread destroy functionality 80b8d8b883SAshok Raj * for idle threads. 81b8d8b883SAshok Raj */ 82b8d8b883SAshok Raj struct task_struct *idle_thread_array[NR_CPUS]; 83b8d8b883SAshok Raj 84b8d8b883SAshok Raj /* 85b8d8b883SAshok Raj * Global array allocated for NR_CPUS at boot time 86b8d8b883SAshok Raj */ 87b8d8b883SAshok Raj struct sal_to_os_boot sal_boot_rendez_state[NR_CPUS]; 88b8d8b883SAshok Raj 89b8d8b883SAshok Raj /* 90b8d8b883SAshok Raj * start_ap in head.S uses this to store current booting cpu 91b8d8b883SAshok Raj * info. 92b8d8b883SAshok Raj */ 93b8d8b883SAshok Raj struct sal_to_os_boot *sal_state_for_booting_cpu = &sal_boot_rendez_state[0]; 94b8d8b883SAshok Raj 95b8d8b883SAshok Raj #define set_brendez_area(x) (sal_state_for_booting_cpu = &sal_boot_rendez_state[(x)]); 96b8d8b883SAshok Raj 97b8d8b883SAshok Raj #define get_idle_for_cpu(x) (idle_thread_array[(x)]) 98b8d8b883SAshok Raj #define set_idle_for_cpu(x,p) (idle_thread_array[(x)] = (p)) 99b8d8b883SAshok Raj 100b8d8b883SAshok Raj #else 101b8d8b883SAshok Raj 102b8d8b883SAshok Raj #define get_idle_for_cpu(x) (NULL) 103b8d8b883SAshok Raj #define set_idle_for_cpu(x,p) 104b8d8b883SAshok Raj #define set_brendez_area(x) 105b8d8b883SAshok Raj #endif 106b8d8b883SAshok Raj 1071da177e4SLinus Torvalds 1081da177e4SLinus Torvalds /* 1091da177e4SLinus Torvalds * ITC synchronization related stuff: 1101da177e4SLinus Torvalds */ 111ff741906SAshok Raj #define MASTER (0) 1121da177e4SLinus Torvalds #define SLAVE (SMP_CACHE_BYTES/8) 1131da177e4SLinus Torvalds 1141da177e4SLinus Torvalds #define NUM_ROUNDS 64 /* magic value */ 1151da177e4SLinus Torvalds #define NUM_ITERS 5 /* likewise */ 1161da177e4SLinus Torvalds 1171da177e4SLinus Torvalds static DEFINE_SPINLOCK(itc_sync_lock); 1181da177e4SLinus Torvalds static volatile unsigned long go[SLAVE + 1]; 1191da177e4SLinus Torvalds 1201da177e4SLinus Torvalds #define DEBUG_ITC_SYNC 0 1211da177e4SLinus Torvalds 1221da177e4SLinus Torvalds extern void __devinit calibrate_delay (void); 1231da177e4SLinus Torvalds extern void start_ap (void); 1241da177e4SLinus Torvalds extern unsigned long ia64_iobase; 1251da177e4SLinus Torvalds 12636c8b586SIngo Molnar struct task_struct *task_for_booting_cpu; 1271da177e4SLinus Torvalds 1281da177e4SLinus Torvalds /* 1291da177e4SLinus Torvalds * State for each CPU 1301da177e4SLinus Torvalds */ 1311da177e4SLinus Torvalds DEFINE_PER_CPU(int, cpu_state); 1321da177e4SLinus Torvalds 1331da177e4SLinus Torvalds /* Bitmasks of currently online, and possible CPUs */ 1341da177e4SLinus Torvalds cpumask_t cpu_online_map; 1351da177e4SLinus Torvalds EXPORT_SYMBOL(cpu_online_map); 13669aa234bSAshok Raj cpumask_t cpu_possible_map = CPU_MASK_NONE; 1371da177e4SLinus Torvalds EXPORT_SYMBOL(cpu_possible_map); 1381da177e4SLinus Torvalds 139e927ecb0SSuresh Siddha cpumask_t cpu_core_map[NR_CPUS] __cacheline_aligned; 140e927ecb0SSuresh Siddha cpumask_t cpu_sibling_map[NR_CPUS] __cacheline_aligned; 141e927ecb0SSuresh Siddha int smp_num_siblings = 1; 142e927ecb0SSuresh Siddha int smp_num_cpucores = 1; 143e927ecb0SSuresh Siddha 1441da177e4SLinus Torvalds /* which logical CPU number maps to which CPU (physical APIC ID) */ 1451da177e4SLinus Torvalds volatile int ia64_cpu_to_sapicid[NR_CPUS]; 1461da177e4SLinus Torvalds EXPORT_SYMBOL(ia64_cpu_to_sapicid); 1471da177e4SLinus Torvalds 1481da177e4SLinus Torvalds static volatile cpumask_t cpu_callin_map; 1491da177e4SLinus Torvalds 1501da177e4SLinus Torvalds struct smp_boot_data smp_boot_data __initdata; 1511da177e4SLinus Torvalds 1521da177e4SLinus Torvalds unsigned long ap_wakeup_vector = -1; /* External Int use to wakeup APs */ 1531da177e4SLinus Torvalds 1541da177e4SLinus Torvalds char __initdata no_int_routing; 1551da177e4SLinus Torvalds 1561da177e4SLinus Torvalds unsigned char smp_int_redirect; /* are INT and IPI redirectable by the chipset? */ 1571da177e4SLinus Torvalds 158ff741906SAshok Raj #ifdef CONFIG_FORCE_CPEI_RETARGET 159ff741906SAshok Raj #define CPEI_OVERRIDE_DEFAULT (1) 160ff741906SAshok Raj #else 161ff741906SAshok Raj #define CPEI_OVERRIDE_DEFAULT (0) 162ff741906SAshok Raj #endif 163ff741906SAshok Raj 164ff741906SAshok Raj unsigned int force_cpei_retarget = CPEI_OVERRIDE_DEFAULT; 165ff741906SAshok Raj 166ff741906SAshok Raj static int __init 167ff741906SAshok Raj cmdl_force_cpei(char *str) 168ff741906SAshok Raj { 169ff741906SAshok Raj int value=0; 170ff741906SAshok Raj 171ff741906SAshok Raj get_option (&str, &value); 172ff741906SAshok Raj force_cpei_retarget = value; 173ff741906SAshok Raj 174ff741906SAshok Raj return 1; 175ff741906SAshok Raj } 176ff741906SAshok Raj 177ff741906SAshok Raj __setup("force_cpei=", cmdl_force_cpei); 178ff741906SAshok Raj 1791da177e4SLinus Torvalds static int __init 1801da177e4SLinus Torvalds nointroute (char *str) 1811da177e4SLinus Torvalds { 1821da177e4SLinus Torvalds no_int_routing = 1; 1831da177e4SLinus Torvalds printk ("no_int_routing on\n"); 1841da177e4SLinus Torvalds return 1; 1851da177e4SLinus Torvalds } 1861da177e4SLinus Torvalds 1871da177e4SLinus Torvalds __setup("nointroute", nointroute); 1881da177e4SLinus Torvalds 189ff741906SAshok Raj static void fix_b0_for_bsp(void) 190ff741906SAshok Raj { 191ff741906SAshok Raj #ifdef CONFIG_HOTPLUG_CPU 192ff741906SAshok Raj int cpuid; 193ff741906SAshok Raj static int fix_bsp_b0 = 1; 194ff741906SAshok Raj 195ff741906SAshok Raj cpuid = smp_processor_id(); 196ff741906SAshok Raj 197ff741906SAshok Raj /* 198ff741906SAshok Raj * Cache the b0 value on the first AP that comes up 199ff741906SAshok Raj */ 200ff741906SAshok Raj if (!(fix_bsp_b0 && cpuid)) 201ff741906SAshok Raj return; 202ff741906SAshok Raj 203ff741906SAshok Raj sal_boot_rendez_state[0].br[0] = sal_boot_rendez_state[cpuid].br[0]; 204ff741906SAshok Raj printk ("Fixed BSP b0 value from CPU %d\n", cpuid); 205ff741906SAshok Raj 206ff741906SAshok Raj fix_bsp_b0 = 0; 207ff741906SAshok Raj #endif 208ff741906SAshok Raj } 209ff741906SAshok Raj 2101da177e4SLinus Torvalds void 2111da177e4SLinus Torvalds sync_master (void *arg) 2121da177e4SLinus Torvalds { 2131da177e4SLinus Torvalds unsigned long flags, i; 2141da177e4SLinus Torvalds 2151da177e4SLinus Torvalds go[MASTER] = 0; 2161da177e4SLinus Torvalds 2171da177e4SLinus Torvalds local_irq_save(flags); 2181da177e4SLinus Torvalds { 2191da177e4SLinus Torvalds for (i = 0; i < NUM_ROUNDS*NUM_ITERS; ++i) { 22082975115SDavid Mosberger-Tang while (!go[MASTER]) 22182975115SDavid Mosberger-Tang cpu_relax(); 2221da177e4SLinus Torvalds go[MASTER] = 0; 2231da177e4SLinus Torvalds go[SLAVE] = ia64_get_itc(); 2241da177e4SLinus Torvalds } 2251da177e4SLinus Torvalds } 2261da177e4SLinus Torvalds local_irq_restore(flags); 2271da177e4SLinus Torvalds } 2281da177e4SLinus Torvalds 2291da177e4SLinus Torvalds /* 2301da177e4SLinus Torvalds * Return the number of cycles by which our itc differs from the itc on the master 2311da177e4SLinus Torvalds * (time-keeper) CPU. A positive number indicates our itc is ahead of the master, 2321da177e4SLinus Torvalds * negative that it is behind. 2331da177e4SLinus Torvalds */ 2341da177e4SLinus Torvalds static inline long 2351da177e4SLinus Torvalds get_delta (long *rt, long *master) 2361da177e4SLinus Torvalds { 2371da177e4SLinus Torvalds unsigned long best_t0 = 0, best_t1 = ~0UL, best_tm = 0; 2381da177e4SLinus Torvalds unsigned long tcenter, t0, t1, tm; 2391da177e4SLinus Torvalds long i; 2401da177e4SLinus Torvalds 2411da177e4SLinus Torvalds for (i = 0; i < NUM_ITERS; ++i) { 2421da177e4SLinus Torvalds t0 = ia64_get_itc(); 2431da177e4SLinus Torvalds go[MASTER] = 1; 24482975115SDavid Mosberger-Tang while (!(tm = go[SLAVE])) 24582975115SDavid Mosberger-Tang cpu_relax(); 2461da177e4SLinus Torvalds go[SLAVE] = 0; 2471da177e4SLinus Torvalds t1 = ia64_get_itc(); 2481da177e4SLinus Torvalds 2491da177e4SLinus Torvalds if (t1 - t0 < best_t1 - best_t0) 2501da177e4SLinus Torvalds best_t0 = t0, best_t1 = t1, best_tm = tm; 2511da177e4SLinus Torvalds } 2521da177e4SLinus Torvalds 2531da177e4SLinus Torvalds *rt = best_t1 - best_t0; 2541da177e4SLinus Torvalds *master = best_tm - best_t0; 2551da177e4SLinus Torvalds 2561da177e4SLinus Torvalds /* average best_t0 and best_t1 without overflow: */ 2571da177e4SLinus Torvalds tcenter = (best_t0/2 + best_t1/2); 2581da177e4SLinus Torvalds if (best_t0 % 2 + best_t1 % 2 == 2) 2591da177e4SLinus Torvalds ++tcenter; 2601da177e4SLinus Torvalds return tcenter - best_tm; 2611da177e4SLinus Torvalds } 2621da177e4SLinus Torvalds 2631da177e4SLinus Torvalds /* 2641da177e4SLinus Torvalds * Synchronize ar.itc of the current (slave) CPU with the ar.itc of the MASTER CPU 2651da177e4SLinus Torvalds * (normally the time-keeper CPU). We use a closed loop to eliminate the possibility of 2661da177e4SLinus Torvalds * unaccounted-for errors (such as getting a machine check in the middle of a calibration 2671da177e4SLinus Torvalds * step). The basic idea is for the slave to ask the master what itc value it has and to 2681da177e4SLinus Torvalds * read its own itc before and after the master responds. Each iteration gives us three 2691da177e4SLinus Torvalds * timestamps: 2701da177e4SLinus Torvalds * 2711da177e4SLinus Torvalds * slave master 2721da177e4SLinus Torvalds * 2731da177e4SLinus Torvalds * t0 ---\ 2741da177e4SLinus Torvalds * ---\ 2751da177e4SLinus Torvalds * ---> 2761da177e4SLinus Torvalds * tm 2771da177e4SLinus Torvalds * /--- 2781da177e4SLinus Torvalds * /--- 2791da177e4SLinus Torvalds * t1 <--- 2801da177e4SLinus Torvalds * 2811da177e4SLinus Torvalds * 2821da177e4SLinus Torvalds * The goal is to adjust the slave's ar.itc such that tm falls exactly half-way between t0 2831da177e4SLinus Torvalds * and t1. If we achieve this, the clocks are synchronized provided the interconnect 2841da177e4SLinus Torvalds * between the slave and the master is symmetric. Even if the interconnect were 2851da177e4SLinus Torvalds * asymmetric, we would still know that the synchronization error is smaller than the 2861da177e4SLinus Torvalds * roundtrip latency (t0 - t1). 2871da177e4SLinus Torvalds * 2881da177e4SLinus Torvalds * When the interconnect is quiet and symmetric, this lets us synchronize the itc to 2891da177e4SLinus Torvalds * within one or two cycles. However, we can only *guarantee* that the synchronization is 2901da177e4SLinus Torvalds * accurate to within a round-trip time, which is typically in the range of several 2911da177e4SLinus Torvalds * hundred cycles (e.g., ~500 cycles). In practice, this means that the itc's are usually 2921da177e4SLinus Torvalds * almost perfectly synchronized, but we shouldn't assume that the accuracy is much better 2931da177e4SLinus Torvalds * than half a micro second or so. 2941da177e4SLinus Torvalds */ 2951da177e4SLinus Torvalds void 2961da177e4SLinus Torvalds ia64_sync_itc (unsigned int master) 2971da177e4SLinus Torvalds { 2981da177e4SLinus Torvalds long i, delta, adj, adjust_latency = 0, done = 0; 2991da177e4SLinus Torvalds unsigned long flags, rt, master_time_stamp, bound; 3001da177e4SLinus Torvalds #if DEBUG_ITC_SYNC 3011da177e4SLinus Torvalds struct { 3021da177e4SLinus Torvalds long rt; /* roundtrip time */ 3031da177e4SLinus Torvalds long master; /* master's timestamp */ 3041da177e4SLinus Torvalds long diff; /* difference between midpoint and master's timestamp */ 3051da177e4SLinus Torvalds long lat; /* estimate of itc adjustment latency */ 3061da177e4SLinus Torvalds } t[NUM_ROUNDS]; 3071da177e4SLinus Torvalds #endif 3081da177e4SLinus Torvalds 3091da177e4SLinus Torvalds /* 3101da177e4SLinus Torvalds * Make sure local timer ticks are disabled while we sync. If 3111da177e4SLinus Torvalds * they were enabled, we'd have to worry about nasty issues 3121da177e4SLinus Torvalds * like setting the ITC ahead of (or a long time before) the 3131da177e4SLinus Torvalds * next scheduled tick. 3141da177e4SLinus Torvalds */ 3151da177e4SLinus Torvalds BUG_ON((ia64_get_itv() & (1 << 16)) == 0); 3161da177e4SLinus Torvalds 3171da177e4SLinus Torvalds go[MASTER] = 1; 3181da177e4SLinus Torvalds 3191da177e4SLinus Torvalds if (smp_call_function_single(master, sync_master, NULL, 1, 0) < 0) { 3201da177e4SLinus Torvalds printk(KERN_ERR "sync_itc: failed to get attention of CPU %u!\n", master); 3211da177e4SLinus Torvalds return; 3221da177e4SLinus Torvalds } 3231da177e4SLinus Torvalds 32482975115SDavid Mosberger-Tang while (go[MASTER]) 32582975115SDavid Mosberger-Tang cpu_relax(); /* wait for master to be ready */ 3261da177e4SLinus Torvalds 3271da177e4SLinus Torvalds spin_lock_irqsave(&itc_sync_lock, flags); 3281da177e4SLinus Torvalds { 3291da177e4SLinus Torvalds for (i = 0; i < NUM_ROUNDS; ++i) { 3301da177e4SLinus Torvalds delta = get_delta(&rt, &master_time_stamp); 3311da177e4SLinus Torvalds if (delta == 0) { 3321da177e4SLinus Torvalds done = 1; /* let's lock on to this... */ 3331da177e4SLinus Torvalds bound = rt; 3341da177e4SLinus Torvalds } 3351da177e4SLinus Torvalds 3361da177e4SLinus Torvalds if (!done) { 3371da177e4SLinus Torvalds if (i > 0) { 3381da177e4SLinus Torvalds adjust_latency += -delta; 3391da177e4SLinus Torvalds adj = -delta + adjust_latency/4; 3401da177e4SLinus Torvalds } else 3411da177e4SLinus Torvalds adj = -delta; 3421da177e4SLinus Torvalds 3431da177e4SLinus Torvalds ia64_set_itc(ia64_get_itc() + adj); 3441da177e4SLinus Torvalds } 3451da177e4SLinus Torvalds #if DEBUG_ITC_SYNC 3461da177e4SLinus Torvalds t[i].rt = rt; 3471da177e4SLinus Torvalds t[i].master = master_time_stamp; 3481da177e4SLinus Torvalds t[i].diff = delta; 3491da177e4SLinus Torvalds t[i].lat = adjust_latency/4; 3501da177e4SLinus Torvalds #endif 3511da177e4SLinus Torvalds } 3521da177e4SLinus Torvalds } 3531da177e4SLinus Torvalds spin_unlock_irqrestore(&itc_sync_lock, flags); 3541da177e4SLinus Torvalds 3551da177e4SLinus Torvalds #if DEBUG_ITC_SYNC 3561da177e4SLinus Torvalds for (i = 0; i < NUM_ROUNDS; ++i) 3571da177e4SLinus Torvalds printk("rt=%5ld master=%5ld diff=%5ld adjlat=%5ld\n", 3581da177e4SLinus Torvalds t[i].rt, t[i].master, t[i].diff, t[i].lat); 3591da177e4SLinus Torvalds #endif 3601da177e4SLinus Torvalds 3611da177e4SLinus Torvalds printk(KERN_INFO "CPU %d: synchronized ITC with CPU %u (last diff %ld cycles, " 3621da177e4SLinus Torvalds "maxerr %lu cycles)\n", smp_processor_id(), master, delta, rt); 3631da177e4SLinus Torvalds } 3641da177e4SLinus Torvalds 3651da177e4SLinus Torvalds /* 3661da177e4SLinus Torvalds * Ideally sets up per-cpu profiling hooks. Doesn't do much now... 3671da177e4SLinus Torvalds */ 3681da177e4SLinus Torvalds static inline void __devinit 3691da177e4SLinus Torvalds smp_setup_percpu_timer (void) 3701da177e4SLinus Torvalds { 3711da177e4SLinus Torvalds } 3721da177e4SLinus Torvalds 373d86ebd14STony Luck static void __cpuinit 3741da177e4SLinus Torvalds smp_callin (void) 3751da177e4SLinus Torvalds { 376ff741906SAshok Raj int cpuid, phys_id, itc_master; 377ead6caaeSJack Steiner struct cpuinfo_ia64 *last_cpuinfo, *this_cpuinfo; 3781da177e4SLinus Torvalds extern void ia64_init_itm(void); 379ff741906SAshok Raj extern volatile int time_keeper_id; 3801da177e4SLinus Torvalds 3811da177e4SLinus Torvalds #ifdef CONFIG_PERFMON 3821da177e4SLinus Torvalds extern void pfm_init_percpu(void); 3831da177e4SLinus Torvalds #endif 3841da177e4SLinus Torvalds 3851da177e4SLinus Torvalds cpuid = smp_processor_id(); 3861da177e4SLinus Torvalds phys_id = hard_smp_processor_id(); 387ff741906SAshok Raj itc_master = time_keeper_id; 3881da177e4SLinus Torvalds 3891da177e4SLinus Torvalds if (cpu_online(cpuid)) { 3901da177e4SLinus Torvalds printk(KERN_ERR "huh, phys CPU#0x%x, CPU#0x%x already present??\n", 3911da177e4SLinus Torvalds phys_id, cpuid); 3921da177e4SLinus Torvalds BUG(); 3931da177e4SLinus Torvalds } 3941da177e4SLinus Torvalds 395ff741906SAshok Raj fix_b0_for_bsp(); 396ff741906SAshok Raj 3971da177e4SLinus Torvalds lock_ipi_calllock(); 398e1b30a39SYasuaki Ishimatsu spin_lock(&vector_lock); 399e1b30a39SYasuaki Ishimatsu /* Setup the per cpu irq handling data structures */ 400e1b30a39SYasuaki Ishimatsu __setup_vector_irq(cpuid); 4011da177e4SLinus Torvalds cpu_set(cpuid, cpu_online_map); 4021da177e4SLinus Torvalds unlock_ipi_calllock(); 403a9fa06c2SShaohua Li per_cpu(cpu_state, cpuid) = CPU_ONLINE; 404e1b30a39SYasuaki Ishimatsu spin_unlock(&vector_lock); 4051da177e4SLinus Torvalds 4061da177e4SLinus Torvalds smp_setup_percpu_timer(); 4071da177e4SLinus Torvalds 4081da177e4SLinus Torvalds ia64_mca_cmc_vector_setup(); /* Setup vector on AP */ 4091da177e4SLinus Torvalds 4101da177e4SLinus Torvalds #ifdef CONFIG_PERFMON 4111da177e4SLinus Torvalds pfm_init_percpu(); 4121da177e4SLinus Torvalds #endif 4131da177e4SLinus Torvalds 4141da177e4SLinus Torvalds local_irq_enable(); 4151da177e4SLinus Torvalds 4161da177e4SLinus Torvalds if (!(sal_platform_features & IA64_SAL_PLATFORM_FEATURE_ITC_DRIFT)) { 4171da177e4SLinus Torvalds /* 4181da177e4SLinus Torvalds * Synchronize the ITC with the BP. Need to do this after irqs are 4191da177e4SLinus Torvalds * enabled because ia64_sync_itc() calls smp_call_function_single(), which 4201da177e4SLinus Torvalds * calls spin_unlock_bh(), which calls spin_unlock_bh(), which calls 4211da177e4SLinus Torvalds * local_bh_enable(), which bugs out if irqs are not enabled... 4221da177e4SLinus Torvalds */ 423ff741906SAshok Raj Dprintk("Going to syncup ITC with ITC Master.\n"); 424ff741906SAshok Raj ia64_sync_itc(itc_master); 4251da177e4SLinus Torvalds } 4261da177e4SLinus Torvalds 4271da177e4SLinus Torvalds /* 4281da177e4SLinus Torvalds * Get our bogomips. 4291da177e4SLinus Torvalds */ 4301da177e4SLinus Torvalds ia64_init_itm(); 431ead6caaeSJack Steiner 432ead6caaeSJack Steiner /* 433ead6caaeSJack Steiner * Delay calibration can be skipped if new processor is identical to the 434ead6caaeSJack Steiner * previous processor. 435ead6caaeSJack Steiner */ 436ead6caaeSJack Steiner last_cpuinfo = cpu_data(cpuid - 1); 437ead6caaeSJack Steiner this_cpuinfo = local_cpu_data; 438ead6caaeSJack Steiner if (last_cpuinfo->itc_freq != this_cpuinfo->itc_freq || 439ead6caaeSJack Steiner last_cpuinfo->proc_freq != this_cpuinfo->proc_freq || 440ead6caaeSJack Steiner last_cpuinfo->features != this_cpuinfo->features || 441ead6caaeSJack Steiner last_cpuinfo->revision != this_cpuinfo->revision || 442ead6caaeSJack Steiner last_cpuinfo->family != this_cpuinfo->family || 443ead6caaeSJack Steiner last_cpuinfo->archrev != this_cpuinfo->archrev || 444ead6caaeSJack Steiner last_cpuinfo->model != this_cpuinfo->model) 4451da177e4SLinus Torvalds calibrate_delay(); 4461da177e4SLinus Torvalds local_cpu_data->loops_per_jiffy = loops_per_jiffy; 4471da177e4SLinus Torvalds 4481da177e4SLinus Torvalds #ifdef CONFIG_IA32_SUPPORT 4491da177e4SLinus Torvalds ia32_gdt_init(); 4501da177e4SLinus Torvalds #endif 4511da177e4SLinus Torvalds 4521da177e4SLinus Torvalds /* 4531da177e4SLinus Torvalds * Allow the master to continue. 4541da177e4SLinus Torvalds */ 4551da177e4SLinus Torvalds cpu_set(cpuid, cpu_callin_map); 4561da177e4SLinus Torvalds Dprintk("Stack on CPU %d at about %p\n",cpuid, &cpuid); 4571da177e4SLinus Torvalds } 4581da177e4SLinus Torvalds 4591da177e4SLinus Torvalds 4601da177e4SLinus Torvalds /* 4611da177e4SLinus Torvalds * Activate a secondary processor. head.S calls this. 4621da177e4SLinus Torvalds */ 463d86ebd14STony Luck int __cpuinit 4641da177e4SLinus Torvalds start_secondary (void *unused) 4651da177e4SLinus Torvalds { 4661da177e4SLinus Torvalds /* Early console may use I/O ports */ 4671da177e4SLinus Torvalds ia64_set_kr(IA64_KR_IO_BASE, __pa(ia64_iobase)); 4681da177e4SLinus Torvalds Dprintk("start_secondary: starting CPU 0x%x\n", hard_smp_processor_id()); 4691da177e4SLinus Torvalds efi_map_pal_code(); 4701da177e4SLinus Torvalds cpu_init(); 4715bfb5d69SNick Piggin preempt_disable(); 4721da177e4SLinus Torvalds smp_callin(); 4731da177e4SLinus Torvalds 4741da177e4SLinus Torvalds cpu_idle(); 4751da177e4SLinus Torvalds return 0; 4761da177e4SLinus Torvalds } 4771da177e4SLinus Torvalds 4781da177e4SLinus Torvalds struct pt_regs * __devinit idle_regs(struct pt_regs *regs) 4791da177e4SLinus Torvalds { 4801da177e4SLinus Torvalds return NULL; 4811da177e4SLinus Torvalds } 4821da177e4SLinus Torvalds 4831da177e4SLinus Torvalds struct create_idle { 4846d5aefb8SDavid Howells struct work_struct work; 4851da177e4SLinus Torvalds struct task_struct *idle; 4861da177e4SLinus Torvalds struct completion done; 4871da177e4SLinus Torvalds int cpu; 4881da177e4SLinus Torvalds }; 4891da177e4SLinus Torvalds 4909d6f40b8STony Luck void __cpuinit 4916d5aefb8SDavid Howells do_fork_idle(struct work_struct *work) 4921da177e4SLinus Torvalds { 4936d5aefb8SDavid Howells struct create_idle *c_idle = 4946d5aefb8SDavid Howells container_of(work, struct create_idle, work); 4951da177e4SLinus Torvalds 4961da177e4SLinus Torvalds c_idle->idle = fork_idle(c_idle->cpu); 4971da177e4SLinus Torvalds complete(&c_idle->done); 4981da177e4SLinus Torvalds } 4991da177e4SLinus Torvalds 5009d6f40b8STony Luck static int __cpuinit 5011da177e4SLinus Torvalds do_boot_cpu (int sapicid, int cpu) 5021da177e4SLinus Torvalds { 5031da177e4SLinus Torvalds int timeout; 5041da177e4SLinus Torvalds struct create_idle c_idle = { 5056d5aefb8SDavid Howells .work = __WORK_INITIALIZER(c_idle.work, do_fork_idle), 5061da177e4SLinus Torvalds .cpu = cpu, 5071da177e4SLinus Torvalds .done = COMPLETION_INITIALIZER(c_idle.done), 5081da177e4SLinus Torvalds }; 509b8d8b883SAshok Raj 510b8d8b883SAshok Raj c_idle.idle = get_idle_for_cpu(cpu); 511b8d8b883SAshok Raj if (c_idle.idle) { 512b8d8b883SAshok Raj init_idle(c_idle.idle, cpu); 513b8d8b883SAshok Raj goto do_rest; 514b8d8b883SAshok Raj } 515b8d8b883SAshok Raj 5161da177e4SLinus Torvalds /* 5171da177e4SLinus Torvalds * We can't use kernel_thread since we must avoid to reschedule the child. 5181da177e4SLinus Torvalds */ 5191da177e4SLinus Torvalds if (!keventd_up() || current_is_keventd()) 5206d5aefb8SDavid Howells c_idle.work.func(&c_idle.work); 5211da177e4SLinus Torvalds else { 5226d5aefb8SDavid Howells schedule_work(&c_idle.work); 5231da177e4SLinus Torvalds wait_for_completion(&c_idle.done); 5241da177e4SLinus Torvalds } 5251da177e4SLinus Torvalds 5261da177e4SLinus Torvalds if (IS_ERR(c_idle.idle)) 5271da177e4SLinus Torvalds panic("failed fork for CPU %d", cpu); 528b8d8b883SAshok Raj 529b8d8b883SAshok Raj set_idle_for_cpu(cpu, c_idle.idle); 530b8d8b883SAshok Raj 531b8d8b883SAshok Raj do_rest: 5321da177e4SLinus Torvalds task_for_booting_cpu = c_idle.idle; 5331da177e4SLinus Torvalds 5341da177e4SLinus Torvalds Dprintk("Sending wakeup vector %lu to AP 0x%x/0x%x.\n", ap_wakeup_vector, cpu, sapicid); 5351da177e4SLinus Torvalds 536b8d8b883SAshok Raj set_brendez_area(cpu); 5371da177e4SLinus Torvalds platform_send_ipi(cpu, ap_wakeup_vector, IA64_IPI_DM_INT, 0); 5381da177e4SLinus Torvalds 5391da177e4SLinus Torvalds /* 5401da177e4SLinus Torvalds * Wait 10s total for the AP to start 5411da177e4SLinus Torvalds */ 5421da177e4SLinus Torvalds Dprintk("Waiting on callin_map ..."); 5431da177e4SLinus Torvalds for (timeout = 0; timeout < 100000; timeout++) { 5441da177e4SLinus Torvalds if (cpu_isset(cpu, cpu_callin_map)) 5451da177e4SLinus Torvalds break; /* It has booted */ 5461da177e4SLinus Torvalds udelay(100); 5471da177e4SLinus Torvalds } 5481da177e4SLinus Torvalds Dprintk("\n"); 5491da177e4SLinus Torvalds 5501da177e4SLinus Torvalds if (!cpu_isset(cpu, cpu_callin_map)) { 5511da177e4SLinus Torvalds printk(KERN_ERR "Processor 0x%x/0x%x is stuck.\n", cpu, sapicid); 5521da177e4SLinus Torvalds ia64_cpu_to_sapicid[cpu] = -1; 5531da177e4SLinus Torvalds cpu_clear(cpu, cpu_online_map); /* was set in smp_callin() */ 5541da177e4SLinus Torvalds return -EINVAL; 5551da177e4SLinus Torvalds } 5561da177e4SLinus Torvalds return 0; 5571da177e4SLinus Torvalds } 5581da177e4SLinus Torvalds 5591da177e4SLinus Torvalds static int __init 5601da177e4SLinus Torvalds decay (char *str) 5611da177e4SLinus Torvalds { 5621da177e4SLinus Torvalds int ticks; 5631da177e4SLinus Torvalds get_option (&str, &ticks); 5641da177e4SLinus Torvalds return 1; 5651da177e4SLinus Torvalds } 5661da177e4SLinus Torvalds 5671da177e4SLinus Torvalds __setup("decay=", decay); 5681da177e4SLinus Torvalds 5691da177e4SLinus Torvalds /* 5701da177e4SLinus Torvalds * Initialize the logical CPU number to SAPICID mapping 5711da177e4SLinus Torvalds */ 5721da177e4SLinus Torvalds void __init 5731da177e4SLinus Torvalds smp_build_cpu_map (void) 5741da177e4SLinus Torvalds { 5751da177e4SLinus Torvalds int sapicid, cpu, i; 5761da177e4SLinus Torvalds int boot_cpu_id = hard_smp_processor_id(); 5771da177e4SLinus Torvalds 5781da177e4SLinus Torvalds for (cpu = 0; cpu < NR_CPUS; cpu++) { 5791da177e4SLinus Torvalds ia64_cpu_to_sapicid[cpu] = -1; 5801da177e4SLinus Torvalds } 5811da177e4SLinus Torvalds 5821da177e4SLinus Torvalds ia64_cpu_to_sapicid[0] = boot_cpu_id; 5831da177e4SLinus Torvalds cpus_clear(cpu_present_map); 5841da177e4SLinus Torvalds cpu_set(0, cpu_present_map); 5851da177e4SLinus Torvalds cpu_set(0, cpu_possible_map); 5861da177e4SLinus Torvalds for (cpu = 1, i = 0; i < smp_boot_data.cpu_count; i++) { 5871da177e4SLinus Torvalds sapicid = smp_boot_data.cpu_phys_id[i]; 5881da177e4SLinus Torvalds if (sapicid == boot_cpu_id) 5891da177e4SLinus Torvalds continue; 5901da177e4SLinus Torvalds cpu_set(cpu, cpu_present_map); 5911da177e4SLinus Torvalds cpu_set(cpu, cpu_possible_map); 5921da177e4SLinus Torvalds ia64_cpu_to_sapicid[cpu] = sapicid; 5931da177e4SLinus Torvalds cpu++; 5941da177e4SLinus Torvalds } 5951da177e4SLinus Torvalds } 5961da177e4SLinus Torvalds 5971da177e4SLinus Torvalds /* 5981da177e4SLinus Torvalds * Cycle through the APs sending Wakeup IPIs to boot each. 5991da177e4SLinus Torvalds */ 6001da177e4SLinus Torvalds void __init 6011da177e4SLinus Torvalds smp_prepare_cpus (unsigned int max_cpus) 6021da177e4SLinus Torvalds { 6031da177e4SLinus Torvalds int boot_cpu_id = hard_smp_processor_id(); 6041da177e4SLinus Torvalds 6051da177e4SLinus Torvalds /* 6061da177e4SLinus Torvalds * Initialize the per-CPU profiling counter/multiplier 6071da177e4SLinus Torvalds */ 6081da177e4SLinus Torvalds 6091da177e4SLinus Torvalds smp_setup_percpu_timer(); 6101da177e4SLinus Torvalds 6111da177e4SLinus Torvalds /* 6121da177e4SLinus Torvalds * We have the boot CPU online for sure. 6131da177e4SLinus Torvalds */ 6141da177e4SLinus Torvalds cpu_set(0, cpu_online_map); 6151da177e4SLinus Torvalds cpu_set(0, cpu_callin_map); 6161da177e4SLinus Torvalds 6171da177e4SLinus Torvalds local_cpu_data->loops_per_jiffy = loops_per_jiffy; 6181da177e4SLinus Torvalds ia64_cpu_to_sapicid[0] = boot_cpu_id; 6191da177e4SLinus Torvalds 6201da177e4SLinus Torvalds printk(KERN_INFO "Boot processor id 0x%x/0x%x\n", 0, boot_cpu_id); 6211da177e4SLinus Torvalds 6221da177e4SLinus Torvalds current_thread_info()->cpu = 0; 6231da177e4SLinus Torvalds 6241da177e4SLinus Torvalds /* 6251da177e4SLinus Torvalds * If SMP should be disabled, then really disable it! 6261da177e4SLinus Torvalds */ 6271da177e4SLinus Torvalds if (!max_cpus) { 6281da177e4SLinus Torvalds printk(KERN_INFO "SMP mode deactivated.\n"); 6291da177e4SLinus Torvalds cpus_clear(cpu_online_map); 6301da177e4SLinus Torvalds cpus_clear(cpu_present_map); 6311da177e4SLinus Torvalds cpus_clear(cpu_possible_map); 6321da177e4SLinus Torvalds cpu_set(0, cpu_online_map); 6331da177e4SLinus Torvalds cpu_set(0, cpu_present_map); 6341da177e4SLinus Torvalds cpu_set(0, cpu_possible_map); 6351da177e4SLinus Torvalds return; 6361da177e4SLinus Torvalds } 6371da177e4SLinus Torvalds } 6381da177e4SLinus Torvalds 6391da177e4SLinus Torvalds void __devinit smp_prepare_boot_cpu(void) 6401da177e4SLinus Torvalds { 6411da177e4SLinus Torvalds cpu_set(smp_processor_id(), cpu_online_map); 6421da177e4SLinus Torvalds cpu_set(smp_processor_id(), cpu_callin_map); 643a9fa06c2SShaohua Li per_cpu(cpu_state, smp_processor_id()) = CPU_ONLINE; 6441da177e4SLinus Torvalds } 6451da177e4SLinus Torvalds 6461da177e4SLinus Torvalds #ifdef CONFIG_HOTPLUG_CPU 647e927ecb0SSuresh Siddha static inline void 648e927ecb0SSuresh Siddha clear_cpu_sibling_map(int cpu) 649e927ecb0SSuresh Siddha { 650e927ecb0SSuresh Siddha int i; 651e927ecb0SSuresh Siddha 652e927ecb0SSuresh Siddha for_each_cpu_mask(i, cpu_sibling_map[cpu]) 653e927ecb0SSuresh Siddha cpu_clear(cpu, cpu_sibling_map[i]); 654e927ecb0SSuresh Siddha for_each_cpu_mask(i, cpu_core_map[cpu]) 655e927ecb0SSuresh Siddha cpu_clear(cpu, cpu_core_map[i]); 656e927ecb0SSuresh Siddha 657e927ecb0SSuresh Siddha cpu_sibling_map[cpu] = cpu_core_map[cpu] = CPU_MASK_NONE; 658e927ecb0SSuresh Siddha } 659e927ecb0SSuresh Siddha 660e927ecb0SSuresh Siddha static void 661e927ecb0SSuresh Siddha remove_siblinginfo(int cpu) 662e927ecb0SSuresh Siddha { 663e927ecb0SSuresh Siddha int last = 0; 664e927ecb0SSuresh Siddha 665e927ecb0SSuresh Siddha if (cpu_data(cpu)->threads_per_core == 1 && 666e927ecb0SSuresh Siddha cpu_data(cpu)->cores_per_socket == 1) { 667e927ecb0SSuresh Siddha cpu_clear(cpu, cpu_core_map[cpu]); 668e927ecb0SSuresh Siddha cpu_clear(cpu, cpu_sibling_map[cpu]); 669e927ecb0SSuresh Siddha return; 670e927ecb0SSuresh Siddha } 671e927ecb0SSuresh Siddha 672e927ecb0SSuresh Siddha last = (cpus_weight(cpu_core_map[cpu]) == 1 ? 1 : 0); 673e927ecb0SSuresh Siddha 674e927ecb0SSuresh Siddha /* remove it from all sibling map's */ 675e927ecb0SSuresh Siddha clear_cpu_sibling_map(cpu); 676e927ecb0SSuresh Siddha } 677e927ecb0SSuresh Siddha 6781da177e4SLinus Torvalds extern void fixup_irqs(void); 679ff741906SAshok Raj 680ff741906SAshok Raj int migrate_platform_irqs(unsigned int cpu) 681ff741906SAshok Raj { 682ff741906SAshok Raj int new_cpei_cpu; 683ff741906SAshok Raj irq_desc_t *desc = NULL; 684ff741906SAshok Raj cpumask_t mask; 685ff741906SAshok Raj int retval = 0; 686ff741906SAshok Raj 687ff741906SAshok Raj /* 688ff741906SAshok Raj * dont permit CPEI target to removed. 689ff741906SAshok Raj */ 690ff741906SAshok Raj if (cpe_vector > 0 && is_cpu_cpei_target(cpu)) { 691ff741906SAshok Raj printk ("CPU (%d) is CPEI Target\n", cpu); 692ff741906SAshok Raj if (can_cpei_retarget()) { 693ff741906SAshok Raj /* 694ff741906SAshok Raj * Now re-target the CPEI to a different processor 695ff741906SAshok Raj */ 696ff741906SAshok Raj new_cpei_cpu = any_online_cpu(cpu_online_map); 697ff741906SAshok Raj mask = cpumask_of_cpu(new_cpei_cpu); 698ff741906SAshok Raj set_cpei_target_cpu(new_cpei_cpu); 699a8553acdSIngo Molnar desc = irq_desc + ia64_cpe_irq; 700ff741906SAshok Raj /* 70172fdbdceSSimon Arlott * Switch for now, immediately, we need to do fake intr 702ff741906SAshok Raj * as other interrupts, but need to study CPEI behaviour with 703ff741906SAshok Raj * polling before making changes. 704ff741906SAshok Raj */ 705ff741906SAshok Raj if (desc) { 706d1bef4edSIngo Molnar desc->chip->disable(ia64_cpe_irq); 707d1bef4edSIngo Molnar desc->chip->set_affinity(ia64_cpe_irq, mask); 708d1bef4edSIngo Molnar desc->chip->enable(ia64_cpe_irq); 709ff741906SAshok Raj printk ("Re-targetting CPEI to cpu %d\n", new_cpei_cpu); 710ff741906SAshok Raj } 711ff741906SAshok Raj } 712ff741906SAshok Raj if (!desc) { 713ff741906SAshok Raj printk ("Unable to retarget CPEI, offline cpu [%d] failed\n", cpu); 714ff741906SAshok Raj retval = -EBUSY; 715ff741906SAshok Raj } 716ff741906SAshok Raj } 717ff741906SAshok Raj return retval; 718ff741906SAshok Raj } 719ff741906SAshok Raj 7201da177e4SLinus Torvalds /* must be called with cpucontrol mutex held */ 7211da177e4SLinus Torvalds int __cpu_disable(void) 7221da177e4SLinus Torvalds { 7231da177e4SLinus Torvalds int cpu = smp_processor_id(); 7241da177e4SLinus Torvalds 7251da177e4SLinus Torvalds /* 7261da177e4SLinus Torvalds * dont permit boot processor for now 7271da177e4SLinus Torvalds */ 728ff741906SAshok Raj if (cpu == 0 && !bsp_remove_ok) { 729ff741906SAshok Raj printk ("Your platform does not support removal of BSP\n"); 730ff741906SAshok Raj return (-EBUSY); 731ff741906SAshok Raj } 732ff741906SAshok Raj 733ff741906SAshok Raj cpu_clear(cpu, cpu_online_map); 734ff741906SAshok Raj 735ff741906SAshok Raj if (migrate_platform_irqs(cpu)) { 736ff741906SAshok Raj cpu_set(cpu, cpu_online_map); 737ff741906SAshok Raj return (-EBUSY); 738ff741906SAshok Raj } 7391da177e4SLinus Torvalds 740e927ecb0SSuresh Siddha remove_siblinginfo(cpu); 741f3705136SZwane Mwaikambo cpu_clear(cpu, cpu_online_map); 7421da177e4SLinus Torvalds fixup_irqs(); 7431da177e4SLinus Torvalds local_flush_tlb_all(); 744b8d8b883SAshok Raj cpu_clear(cpu, cpu_callin_map); 7451da177e4SLinus Torvalds return 0; 7461da177e4SLinus Torvalds } 7471da177e4SLinus Torvalds 7481da177e4SLinus Torvalds void __cpu_die(unsigned int cpu) 7491da177e4SLinus Torvalds { 7501da177e4SLinus Torvalds unsigned int i; 7511da177e4SLinus Torvalds 7521da177e4SLinus Torvalds for (i = 0; i < 100; i++) { 7531da177e4SLinus Torvalds /* They ack this in play_dead by setting CPU_DEAD */ 7541da177e4SLinus Torvalds if (per_cpu(cpu_state, cpu) == CPU_DEAD) 7551da177e4SLinus Torvalds { 756b8d8b883SAshok Raj printk ("CPU %d is now offline\n", cpu); 7571da177e4SLinus Torvalds return; 7581da177e4SLinus Torvalds } 7591da177e4SLinus Torvalds msleep(100); 7601da177e4SLinus Torvalds } 7611da177e4SLinus Torvalds printk(KERN_ERR "CPU %u didn't die...\n", cpu); 7621da177e4SLinus Torvalds } 7631da177e4SLinus Torvalds #else /* !CONFIG_HOTPLUG_CPU */ 7641da177e4SLinus Torvalds int __cpu_disable(void) 7651da177e4SLinus Torvalds { 7661da177e4SLinus Torvalds return -ENOSYS; 7671da177e4SLinus Torvalds } 7681da177e4SLinus Torvalds 7691da177e4SLinus Torvalds void __cpu_die(unsigned int cpu) 7701da177e4SLinus Torvalds { 7711da177e4SLinus Torvalds /* We said "no" in __cpu_disable */ 7721da177e4SLinus Torvalds BUG(); 7731da177e4SLinus Torvalds } 7741da177e4SLinus Torvalds #endif /* CONFIG_HOTPLUG_CPU */ 7751da177e4SLinus Torvalds 7761da177e4SLinus Torvalds void 7771da177e4SLinus Torvalds smp_cpus_done (unsigned int dummy) 7781da177e4SLinus Torvalds { 7791da177e4SLinus Torvalds int cpu; 7801da177e4SLinus Torvalds unsigned long bogosum = 0; 7811da177e4SLinus Torvalds 7821da177e4SLinus Torvalds /* 7831da177e4SLinus Torvalds * Allow the user to impress friends. 7841da177e4SLinus Torvalds */ 7851da177e4SLinus Torvalds 786dc565b52Shawkes@sgi.com for_each_online_cpu(cpu) { 7871da177e4SLinus Torvalds bogosum += cpu_data(cpu)->loops_per_jiffy; 788dc565b52Shawkes@sgi.com } 7891da177e4SLinus Torvalds 7901da177e4SLinus Torvalds printk(KERN_INFO "Total of %d processors activated (%lu.%02lu BogoMIPS).\n", 7911da177e4SLinus Torvalds (int)num_online_cpus(), bogosum/(500000/HZ), (bogosum/(5000/HZ))%100); 7921da177e4SLinus Torvalds } 7931da177e4SLinus Torvalds 794e927ecb0SSuresh Siddha static inline void __devinit 795e927ecb0SSuresh Siddha set_cpu_sibling_map(int cpu) 796e927ecb0SSuresh Siddha { 797e927ecb0SSuresh Siddha int i; 798e927ecb0SSuresh Siddha 799e927ecb0SSuresh Siddha for_each_online_cpu(i) { 800e927ecb0SSuresh Siddha if ((cpu_data(cpu)->socket_id == cpu_data(i)->socket_id)) { 801e927ecb0SSuresh Siddha cpu_set(i, cpu_core_map[cpu]); 802e927ecb0SSuresh Siddha cpu_set(cpu, cpu_core_map[i]); 803e927ecb0SSuresh Siddha if (cpu_data(cpu)->core_id == cpu_data(i)->core_id) { 804e927ecb0SSuresh Siddha cpu_set(i, cpu_sibling_map[cpu]); 805e927ecb0SSuresh Siddha cpu_set(cpu, cpu_sibling_map[i]); 806e927ecb0SSuresh Siddha } 807e927ecb0SSuresh Siddha } 808e927ecb0SSuresh Siddha } 809e927ecb0SSuresh Siddha } 810e927ecb0SSuresh Siddha 8119d6f40b8STony Luck int __cpuinit 8121da177e4SLinus Torvalds __cpu_up (unsigned int cpu) 8131da177e4SLinus Torvalds { 8141da177e4SLinus Torvalds int ret; 8151da177e4SLinus Torvalds int sapicid; 8161da177e4SLinus Torvalds 8171da177e4SLinus Torvalds sapicid = ia64_cpu_to_sapicid[cpu]; 8181da177e4SLinus Torvalds if (sapicid == -1) 8191da177e4SLinus Torvalds return -EINVAL; 8201da177e4SLinus Torvalds 8211da177e4SLinus Torvalds /* 822b8d8b883SAshok Raj * Already booted cpu? not valid anymore since we dont 823b8d8b883SAshok Raj * do idle loop tightspin anymore. 8241da177e4SLinus Torvalds */ 8251da177e4SLinus Torvalds if (cpu_isset(cpu, cpu_callin_map)) 826b8d8b883SAshok Raj return -EINVAL; 827b8d8b883SAshok Raj 828a9fa06c2SShaohua Li per_cpu(cpu_state, cpu) = CPU_UP_PREPARE; 8291da177e4SLinus Torvalds /* Processor goes to start_secondary(), sets online flag */ 8301da177e4SLinus Torvalds ret = do_boot_cpu(sapicid, cpu); 8311da177e4SLinus Torvalds if (ret < 0) 8321da177e4SLinus Torvalds return ret; 8331da177e4SLinus Torvalds 834e927ecb0SSuresh Siddha if (cpu_data(cpu)->threads_per_core == 1 && 835e927ecb0SSuresh Siddha cpu_data(cpu)->cores_per_socket == 1) { 836e927ecb0SSuresh Siddha cpu_set(cpu, cpu_sibling_map[cpu]); 837e927ecb0SSuresh Siddha cpu_set(cpu, cpu_core_map[cpu]); 838e927ecb0SSuresh Siddha return 0; 839e927ecb0SSuresh Siddha } 840e927ecb0SSuresh Siddha 841e927ecb0SSuresh Siddha set_cpu_sibling_map(cpu); 842e927ecb0SSuresh Siddha 8431da177e4SLinus Torvalds return 0; 8441da177e4SLinus Torvalds } 8451da177e4SLinus Torvalds 8461da177e4SLinus Torvalds /* 84772fdbdceSSimon Arlott * Assume that CPUs have been discovered by some platform-dependent interface. For 8481da177e4SLinus Torvalds * SoftSDV/Lion, that would be ACPI. 8491da177e4SLinus Torvalds * 8501da177e4SLinus Torvalds * Setup of the IPI irq handler is done in irq.c:init_IRQ_SMP(). 8511da177e4SLinus Torvalds */ 8521da177e4SLinus Torvalds void __init 8531da177e4SLinus Torvalds init_smp_config(void) 8541da177e4SLinus Torvalds { 8551da177e4SLinus Torvalds struct fptr { 8561da177e4SLinus Torvalds unsigned long fp; 8571da177e4SLinus Torvalds unsigned long gp; 8581da177e4SLinus Torvalds } *ap_startup; 8591da177e4SLinus Torvalds long sal_ret; 8601da177e4SLinus Torvalds 86172fdbdceSSimon Arlott /* Tell SAL where to drop the APs. */ 8621da177e4SLinus Torvalds ap_startup = (struct fptr *) start_ap; 8631da177e4SLinus Torvalds sal_ret = ia64_sal_set_vectors(SAL_VECTOR_OS_BOOT_RENDEZ, 8641da177e4SLinus Torvalds ia64_tpa(ap_startup->fp), ia64_tpa(ap_startup->gp), 0, 0, 0, 0); 8651da177e4SLinus Torvalds if (sal_ret < 0) 8661da177e4SLinus Torvalds printk(KERN_ERR "SMP: Can't set SAL AP Boot Rendezvous: %s\n", 8671da177e4SLinus Torvalds ia64_sal_strerror(sal_ret)); 8681da177e4SLinus Torvalds } 8691da177e4SLinus Torvalds 870e927ecb0SSuresh Siddha /* 871e927ecb0SSuresh Siddha * identify_siblings(cpu) gets called from identify_cpu. This populates the 872e927ecb0SSuresh Siddha * information related to logical execution units in per_cpu_data structure. 873e927ecb0SSuresh Siddha */ 874e927ecb0SSuresh Siddha void __devinit 875e927ecb0SSuresh Siddha identify_siblings(struct cpuinfo_ia64 *c) 876e927ecb0SSuresh Siddha { 877e927ecb0SSuresh Siddha s64 status; 878e927ecb0SSuresh Siddha u16 pltid; 879e927ecb0SSuresh Siddha pal_logical_to_physical_t info; 880e927ecb0SSuresh Siddha 881e927ecb0SSuresh Siddha if (smp_num_cpucores == 1 && smp_num_siblings == 1) 882e927ecb0SSuresh Siddha return; 883e927ecb0SSuresh Siddha 8844129a953SFenghua Yu if ((status = ia64_pal_logical_to_phys(-1, &info)) != PAL_STATUS_SUCCESS) { 885e927ecb0SSuresh Siddha printk(KERN_ERR "ia64_pal_logical_to_phys failed with %ld\n", 886e927ecb0SSuresh Siddha status); 887e927ecb0SSuresh Siddha return; 888e927ecb0SSuresh Siddha } 889e927ecb0SSuresh Siddha if ((status = ia64_sal_physical_id_info(&pltid)) != PAL_STATUS_SUCCESS) { 890e927ecb0SSuresh Siddha printk(KERN_ERR "ia64_sal_pltid failed with %ld\n", status); 891e927ecb0SSuresh Siddha return; 892e927ecb0SSuresh Siddha } 893e927ecb0SSuresh Siddha 894e927ecb0SSuresh Siddha c->socket_id = (pltid << 8) | info.overview_ppid; 895e927ecb0SSuresh Siddha c->cores_per_socket = info.overview_cpp; 896e927ecb0SSuresh Siddha c->threads_per_core = info.overview_tpc; 8974129a953SFenghua Yu c->num_log = info.overview_num_log; 898e927ecb0SSuresh Siddha 899e927ecb0SSuresh Siddha c->core_id = info.log1_cid; 900e927ecb0SSuresh Siddha c->thread_id = info.log1_tid; 901e927ecb0SSuresh Siddha } 902dd562c05SStephane Eranian 903dd562c05SStephane Eranian /* 904dd562c05SStephane Eranian * returns non zero, if multi-threading is enabled 905dd562c05SStephane Eranian * on at least one physical package. Due to hotplug cpu 906dd562c05SStephane Eranian * and (maxcpus=), all threads may not necessarily be enabled 907dd562c05SStephane Eranian * even though the processor supports multi-threading. 908dd562c05SStephane Eranian */ 909dd562c05SStephane Eranian int is_multithreading_enabled(void) 910dd562c05SStephane Eranian { 911dd562c05SStephane Eranian int i, j; 912dd562c05SStephane Eranian 913dd562c05SStephane Eranian for_each_present_cpu(i) { 914dd562c05SStephane Eranian for_each_present_cpu(j) { 915dd562c05SStephane Eranian if (j == i) 916dd562c05SStephane Eranian continue; 917dd562c05SStephane Eranian if ((cpu_data(j)->socket_id == cpu_data(i)->socket_id)) { 918dd562c05SStephane Eranian if (cpu_data(j)->core_id == cpu_data(i)->core_id) 919dd562c05SStephane Eranian return 1; 920dd562c05SStephane Eranian } 921dd562c05SStephane Eranian } 922dd562c05SStephane Eranian } 923dd562c05SStephane Eranian return 0; 924dd562c05SStephane Eranian } 925dd562c05SStephane Eranian EXPORT_SYMBOL_GPL(is_multithreading_enabled); 926