1*2ac6795fSAnup Patel // SPDX-License-Identifier: GPL-2.0 2*2ac6795fSAnup Patel /* 3*2ac6795fSAnup Patel * Copyright (C) 2020 Western Digital Corporation or its affiliates. 4*2ac6795fSAnup Patel * 5*2ac6795fSAnup Patel * Most of the M-mode (i.e. NoMMU) RISC-V systems usually have a 6*2ac6795fSAnup Patel * CLINT MMIO timer device. 7*2ac6795fSAnup Patel */ 8*2ac6795fSAnup Patel 9*2ac6795fSAnup Patel #define pr_fmt(fmt) "clint: " fmt 10*2ac6795fSAnup Patel #include <linux/bitops.h> 11*2ac6795fSAnup Patel #include <linux/clocksource.h> 12*2ac6795fSAnup Patel #include <linux/clockchips.h> 13*2ac6795fSAnup Patel #include <linux/cpu.h> 14*2ac6795fSAnup Patel #include <linux/delay.h> 15*2ac6795fSAnup Patel #include <linux/module.h> 16*2ac6795fSAnup Patel #include <linux/of_address.h> 17*2ac6795fSAnup Patel #include <linux/sched_clock.h> 18*2ac6795fSAnup Patel #include <linux/io-64-nonatomic-lo-hi.h> 19*2ac6795fSAnup Patel #include <linux/interrupt.h> 20*2ac6795fSAnup Patel #include <linux/of_irq.h> 21*2ac6795fSAnup Patel #include <linux/smp.h> 22*2ac6795fSAnup Patel 23*2ac6795fSAnup Patel #define CLINT_IPI_OFF 0 24*2ac6795fSAnup Patel #define CLINT_TIMER_CMP_OFF 0x4000 25*2ac6795fSAnup Patel #define CLINT_TIMER_VAL_OFF 0xbff8 26*2ac6795fSAnup Patel 27*2ac6795fSAnup Patel /* CLINT manages IPI and Timer for RISC-V M-mode */ 28*2ac6795fSAnup Patel static u32 __iomem *clint_ipi_base; 29*2ac6795fSAnup Patel static u64 __iomem *clint_timer_cmp; 30*2ac6795fSAnup Patel static u64 __iomem *clint_timer_val; 31*2ac6795fSAnup Patel static unsigned long clint_timer_freq; 32*2ac6795fSAnup Patel static unsigned int clint_timer_irq; 33*2ac6795fSAnup Patel 34*2ac6795fSAnup Patel static void clint_send_ipi(const struct cpumask *target) 35*2ac6795fSAnup Patel { 36*2ac6795fSAnup Patel unsigned int cpu; 37*2ac6795fSAnup Patel 38*2ac6795fSAnup Patel for_each_cpu(cpu, target) 39*2ac6795fSAnup Patel writel(1, clint_ipi_base + cpuid_to_hartid_map(cpu)); 40*2ac6795fSAnup Patel } 41*2ac6795fSAnup Patel 42*2ac6795fSAnup Patel static void clint_clear_ipi(void) 43*2ac6795fSAnup Patel { 44*2ac6795fSAnup Patel writel(0, clint_ipi_base + cpuid_to_hartid_map(smp_processor_id())); 45*2ac6795fSAnup Patel } 46*2ac6795fSAnup Patel 47*2ac6795fSAnup Patel static struct riscv_ipi_ops clint_ipi_ops = { 48*2ac6795fSAnup Patel .ipi_inject = clint_send_ipi, 49*2ac6795fSAnup Patel .ipi_clear = clint_clear_ipi, 50*2ac6795fSAnup Patel }; 51*2ac6795fSAnup Patel 52*2ac6795fSAnup Patel #ifdef CONFIG_64BIT 53*2ac6795fSAnup Patel #define clint_get_cycles() readq_relaxed(clint_timer_val) 54*2ac6795fSAnup Patel #else 55*2ac6795fSAnup Patel #define clint_get_cycles() readl_relaxed(clint_timer_val) 56*2ac6795fSAnup Patel #define clint_get_cycles_hi() readl_relaxed(((u32 *)clint_timer_val) + 1) 57*2ac6795fSAnup Patel #endif 58*2ac6795fSAnup Patel 59*2ac6795fSAnup Patel #ifdef CONFIG_64BIT 60*2ac6795fSAnup Patel static u64 notrace clint_get_cycles64(void) 61*2ac6795fSAnup Patel { 62*2ac6795fSAnup Patel return clint_get_cycles(); 63*2ac6795fSAnup Patel } 64*2ac6795fSAnup Patel #else /* CONFIG_64BIT */ 65*2ac6795fSAnup Patel static u64 notrace clint_get_cycles64(void) 66*2ac6795fSAnup Patel { 67*2ac6795fSAnup Patel u32 hi, lo; 68*2ac6795fSAnup Patel 69*2ac6795fSAnup Patel do { 70*2ac6795fSAnup Patel hi = clint_get_cycles_hi(); 71*2ac6795fSAnup Patel lo = clint_get_cycles(); 72*2ac6795fSAnup Patel } while (hi != clint_get_cycles_hi()); 73*2ac6795fSAnup Patel 74*2ac6795fSAnup Patel return ((u64)hi << 32) | lo; 75*2ac6795fSAnup Patel } 76*2ac6795fSAnup Patel #endif /* CONFIG_64BIT */ 77*2ac6795fSAnup Patel 78*2ac6795fSAnup Patel static u64 clint_rdtime(struct clocksource *cs) 79*2ac6795fSAnup Patel { 80*2ac6795fSAnup Patel return clint_get_cycles64(); 81*2ac6795fSAnup Patel } 82*2ac6795fSAnup Patel 83*2ac6795fSAnup Patel static struct clocksource clint_clocksource = { 84*2ac6795fSAnup Patel .name = "clint_clocksource", 85*2ac6795fSAnup Patel .rating = 300, 86*2ac6795fSAnup Patel .mask = CLOCKSOURCE_MASK(64), 87*2ac6795fSAnup Patel .flags = CLOCK_SOURCE_IS_CONTINUOUS, 88*2ac6795fSAnup Patel .read = clint_rdtime, 89*2ac6795fSAnup Patel }; 90*2ac6795fSAnup Patel 91*2ac6795fSAnup Patel static int clint_clock_next_event(unsigned long delta, 92*2ac6795fSAnup Patel struct clock_event_device *ce) 93*2ac6795fSAnup Patel { 94*2ac6795fSAnup Patel void __iomem *r = clint_timer_cmp + 95*2ac6795fSAnup Patel cpuid_to_hartid_map(smp_processor_id()); 96*2ac6795fSAnup Patel 97*2ac6795fSAnup Patel csr_set(CSR_IE, IE_TIE); 98*2ac6795fSAnup Patel writeq_relaxed(clint_get_cycles64() + delta, r); 99*2ac6795fSAnup Patel return 0; 100*2ac6795fSAnup Patel } 101*2ac6795fSAnup Patel 102*2ac6795fSAnup Patel static DEFINE_PER_CPU(struct clock_event_device, clint_clock_event) = { 103*2ac6795fSAnup Patel .name = "clint_clockevent", 104*2ac6795fSAnup Patel .features = CLOCK_EVT_FEAT_ONESHOT, 105*2ac6795fSAnup Patel .rating = 100, 106*2ac6795fSAnup Patel .set_next_event = clint_clock_next_event, 107*2ac6795fSAnup Patel }; 108*2ac6795fSAnup Patel 109*2ac6795fSAnup Patel static int clint_timer_starting_cpu(unsigned int cpu) 110*2ac6795fSAnup Patel { 111*2ac6795fSAnup Patel struct clock_event_device *ce = per_cpu_ptr(&clint_clock_event, cpu); 112*2ac6795fSAnup Patel 113*2ac6795fSAnup Patel ce->cpumask = cpumask_of(cpu); 114*2ac6795fSAnup Patel clockevents_config_and_register(ce, clint_timer_freq, 100, 0x7fffffff); 115*2ac6795fSAnup Patel 116*2ac6795fSAnup Patel enable_percpu_irq(clint_timer_irq, 117*2ac6795fSAnup Patel irq_get_trigger_type(clint_timer_irq)); 118*2ac6795fSAnup Patel return 0; 119*2ac6795fSAnup Patel } 120*2ac6795fSAnup Patel 121*2ac6795fSAnup Patel static int clint_timer_dying_cpu(unsigned int cpu) 122*2ac6795fSAnup Patel { 123*2ac6795fSAnup Patel disable_percpu_irq(clint_timer_irq); 124*2ac6795fSAnup Patel return 0; 125*2ac6795fSAnup Patel } 126*2ac6795fSAnup Patel 127*2ac6795fSAnup Patel static irqreturn_t clint_timer_interrupt(int irq, void *dev_id) 128*2ac6795fSAnup Patel { 129*2ac6795fSAnup Patel struct clock_event_device *evdev = this_cpu_ptr(&clint_clock_event); 130*2ac6795fSAnup Patel 131*2ac6795fSAnup Patel csr_clear(CSR_IE, IE_TIE); 132*2ac6795fSAnup Patel evdev->event_handler(evdev); 133*2ac6795fSAnup Patel 134*2ac6795fSAnup Patel return IRQ_HANDLED; 135*2ac6795fSAnup Patel } 136*2ac6795fSAnup Patel 137*2ac6795fSAnup Patel static int __init clint_timer_init_dt(struct device_node *np) 138*2ac6795fSAnup Patel { 139*2ac6795fSAnup Patel int rc; 140*2ac6795fSAnup Patel u32 i, nr_irqs; 141*2ac6795fSAnup Patel void __iomem *base; 142*2ac6795fSAnup Patel struct of_phandle_args oirq; 143*2ac6795fSAnup Patel 144*2ac6795fSAnup Patel /* 145*2ac6795fSAnup Patel * Ensure that CLINT device interrupts are either RV_IRQ_TIMER or 146*2ac6795fSAnup Patel * RV_IRQ_SOFT. If it's anything else then we ignore the device. 147*2ac6795fSAnup Patel */ 148*2ac6795fSAnup Patel nr_irqs = of_irq_count(np); 149*2ac6795fSAnup Patel for (i = 0; i < nr_irqs; i++) { 150*2ac6795fSAnup Patel if (of_irq_parse_one(np, i, &oirq)) { 151*2ac6795fSAnup Patel pr_err("%pOFP: failed to parse irq %d.\n", np, i); 152*2ac6795fSAnup Patel continue; 153*2ac6795fSAnup Patel } 154*2ac6795fSAnup Patel 155*2ac6795fSAnup Patel if ((oirq.args_count != 1) || 156*2ac6795fSAnup Patel (oirq.args[0] != RV_IRQ_TIMER && 157*2ac6795fSAnup Patel oirq.args[0] != RV_IRQ_SOFT)) { 158*2ac6795fSAnup Patel pr_err("%pOFP: invalid irq %d (hwirq %d)\n", 159*2ac6795fSAnup Patel np, i, oirq.args[0]); 160*2ac6795fSAnup Patel return -ENODEV; 161*2ac6795fSAnup Patel } 162*2ac6795fSAnup Patel 163*2ac6795fSAnup Patel /* Find parent irq domain and map timer irq */ 164*2ac6795fSAnup Patel if (!clint_timer_irq && 165*2ac6795fSAnup Patel oirq.args[0] == RV_IRQ_TIMER && 166*2ac6795fSAnup Patel irq_find_host(oirq.np)) 167*2ac6795fSAnup Patel clint_timer_irq = irq_of_parse_and_map(np, i); 168*2ac6795fSAnup Patel } 169*2ac6795fSAnup Patel 170*2ac6795fSAnup Patel /* If CLINT timer irq not found then fail */ 171*2ac6795fSAnup Patel if (!clint_timer_irq) { 172*2ac6795fSAnup Patel pr_err("%pOFP: timer irq not found\n", np); 173*2ac6795fSAnup Patel return -ENODEV; 174*2ac6795fSAnup Patel } 175*2ac6795fSAnup Patel 176*2ac6795fSAnup Patel base = of_iomap(np, 0); 177*2ac6795fSAnup Patel if (!base) { 178*2ac6795fSAnup Patel pr_err("%pOFP: could not map registers\n", np); 179*2ac6795fSAnup Patel return -ENODEV; 180*2ac6795fSAnup Patel } 181*2ac6795fSAnup Patel 182*2ac6795fSAnup Patel clint_ipi_base = base + CLINT_IPI_OFF; 183*2ac6795fSAnup Patel clint_timer_cmp = base + CLINT_TIMER_CMP_OFF; 184*2ac6795fSAnup Patel clint_timer_val = base + CLINT_TIMER_VAL_OFF; 185*2ac6795fSAnup Patel clint_timer_freq = riscv_timebase; 186*2ac6795fSAnup Patel 187*2ac6795fSAnup Patel pr_info("%pOFP: timer running at %ld Hz\n", np, clint_timer_freq); 188*2ac6795fSAnup Patel 189*2ac6795fSAnup Patel rc = clocksource_register_hz(&clint_clocksource, clint_timer_freq); 190*2ac6795fSAnup Patel if (rc) { 191*2ac6795fSAnup Patel pr_err("%pOFP: clocksource register failed [%d]\n", np, rc); 192*2ac6795fSAnup Patel goto fail_iounmap; 193*2ac6795fSAnup Patel } 194*2ac6795fSAnup Patel 195*2ac6795fSAnup Patel sched_clock_register(clint_get_cycles64, 64, clint_timer_freq); 196*2ac6795fSAnup Patel 197*2ac6795fSAnup Patel rc = request_percpu_irq(clint_timer_irq, clint_timer_interrupt, 198*2ac6795fSAnup Patel "clint-timer", &clint_clock_event); 199*2ac6795fSAnup Patel if (rc) { 200*2ac6795fSAnup Patel pr_err("registering percpu irq failed [%d]\n", rc); 201*2ac6795fSAnup Patel goto fail_iounmap; 202*2ac6795fSAnup Patel } 203*2ac6795fSAnup Patel 204*2ac6795fSAnup Patel rc = cpuhp_setup_state(CPUHP_AP_CLINT_TIMER_STARTING, 205*2ac6795fSAnup Patel "clockevents/clint/timer:starting", 206*2ac6795fSAnup Patel clint_timer_starting_cpu, 207*2ac6795fSAnup Patel clint_timer_dying_cpu); 208*2ac6795fSAnup Patel if (rc) { 209*2ac6795fSAnup Patel pr_err("%pOFP: cpuhp setup state failed [%d]\n", np, rc); 210*2ac6795fSAnup Patel goto fail_free_irq; 211*2ac6795fSAnup Patel } 212*2ac6795fSAnup Patel 213*2ac6795fSAnup Patel riscv_set_ipi_ops(&clint_ipi_ops); 214*2ac6795fSAnup Patel clint_clear_ipi(); 215*2ac6795fSAnup Patel 216*2ac6795fSAnup Patel return 0; 217*2ac6795fSAnup Patel 218*2ac6795fSAnup Patel fail_free_irq: 219*2ac6795fSAnup Patel free_irq(clint_timer_irq, &clint_clock_event); 220*2ac6795fSAnup Patel fail_iounmap: 221*2ac6795fSAnup Patel iounmap(base); 222*2ac6795fSAnup Patel return rc; 223*2ac6795fSAnup Patel } 224*2ac6795fSAnup Patel 225*2ac6795fSAnup Patel TIMER_OF_DECLARE(clint_timer, "riscv,clint0", clint_timer_init_dt); 226*2ac6795fSAnup Patel TIMER_OF_DECLARE(clint_timer1, "sifive,clint0", clint_timer_init_dt); 227