| 1 | // SPDX-License-Identifier: GPL-2.0 |
| 2 | /* |
| 3 | * Timer present on EcoNet EN75xx MIPS based SoCs. |
| 4 | * |
| 5 | * Copyright (C) 2025 by Caleb James DeLisle <cjd@cjdns.fr> |
| 6 | */ |
| 7 | |
| 8 | #include <linux/io.h> |
| 9 | #include <linux/cpumask.h> |
| 10 | #include <linux/interrupt.h> |
| 11 | #include <linux/clockchips.h> |
| 12 | #include <linux/sched_clock.h> |
| 13 | #include <linux/of.h> |
| 14 | #include <linux/of_irq.h> |
| 15 | #include <linux/of_address.h> |
| 16 | #include <linux/cpuhotplug.h> |
| 17 | #include <linux/clk.h> |
| 18 | |
| 19 | #define ECONET_BITS 32 |
| 20 | #define ECONET_MIN_DELTA 0x00001000 |
| 21 | #define ECONET_MAX_DELTA GENMASK(ECONET_BITS - 2, 0) |
| 22 | /* 34Kc hardware has 1 block and 1004Kc has 2. */ |
| 23 | #define ECONET_NUM_BLOCKS DIV_ROUND_UP(NR_CPUS, 2) |
| 24 | |
| 25 | static struct { |
| 26 | void __iomem *membase[ECONET_NUM_BLOCKS]; |
| 27 | u32 freq_hz; |
| 28 | } econet_timer __ro_after_init; |
| 29 | |
| 30 | static DEFINE_PER_CPU(struct clock_event_device, econet_timer_pcpu); |
| 31 | |
| 32 | /* Each memory block has 2 timers, the order of registers is: |
| 33 | * CTL, CMR0, CNT0, CMR1, CNT1 |
| 34 | */ |
| 35 | static inline void __iomem *reg_ctl(u32 timer_n) |
| 36 | { |
| 37 | return econet_timer.membase[timer_n >> 1]; |
| 38 | } |
| 39 | |
| 40 | static inline void __iomem *reg_compare(u32 timer_n) |
| 41 | { |
| 42 | return econet_timer.membase[timer_n >> 1] + (timer_n & 1) * 0x08 + 0x04; |
| 43 | } |
| 44 | |
| 45 | static inline void __iomem *reg_count(u32 timer_n) |
| 46 | { |
| 47 | return econet_timer.membase[timer_n >> 1] + (timer_n & 1) * 0x08 + 0x08; |
| 48 | } |
| 49 | |
| 50 | static inline u32 ctl_bit_enabled(u32 timer_n) |
| 51 | { |
| 52 | return 1U << (timer_n & 1); |
| 53 | } |
| 54 | |
| 55 | static inline u32 ctl_bit_pending(u32 timer_n) |
| 56 | { |
| 57 | return 1U << ((timer_n & 1) + 16); |
| 58 | } |
| 59 | |
| 60 | static bool cevt_is_pending(int cpu_id) |
| 61 | { |
| 62 | return ioread32(reg_ctl(timer_n: cpu_id)) & ctl_bit_pending(timer_n: cpu_id); |
| 63 | } |
| 64 | |
| 65 | static irqreturn_t cevt_interrupt(int irq, void *dev_id) |
| 66 | { |
| 67 | struct clock_event_device *dev = this_cpu_ptr(&econet_timer_pcpu); |
| 68 | int cpu = cpumask_first(srcp: dev->cpumask); |
| 69 | |
| 70 | /* Each VPE has its own events, |
| 71 | * so this will only happen on spurious interrupt. |
| 72 | */ |
| 73 | if (!cevt_is_pending(cpu_id: cpu)) |
| 74 | return IRQ_NONE; |
| 75 | |
| 76 | iowrite32(ioread32(reg_count(timer_n: cpu)), reg_compare(timer_n: cpu)); |
| 77 | dev->event_handler(dev); |
| 78 | return IRQ_HANDLED; |
| 79 | } |
| 80 | |
| 81 | static int cevt_set_next_event(ulong delta, struct clock_event_device *dev) |
| 82 | { |
| 83 | u32 next; |
| 84 | int cpu; |
| 85 | |
| 86 | cpu = cpumask_first(srcp: dev->cpumask); |
| 87 | next = ioread32(reg_count(timer_n: cpu)) + delta; |
| 88 | iowrite32(next, reg_compare(timer_n: cpu)); |
| 89 | |
| 90 | if ((s32)(next - ioread32(reg_count(timer_n: cpu))) < ECONET_MIN_DELTA / 2) |
| 91 | return -ETIME; |
| 92 | |
| 93 | return 0; |
| 94 | } |
| 95 | |
| 96 | static int cevt_init_cpu(uint cpu) |
| 97 | { |
| 98 | struct clock_event_device *cd = &per_cpu(econet_timer_pcpu, cpu); |
| 99 | u32 reg; |
| 100 | |
| 101 | pr_debug("%s: Setting up clockevent for CPU %d\n" , cd->name, cpu); |
| 102 | |
| 103 | reg = ioread32(reg_ctl(timer_n: cpu)) | ctl_bit_enabled(timer_n: cpu); |
| 104 | iowrite32(reg, reg_ctl(timer_n: cpu)); |
| 105 | |
| 106 | enable_percpu_irq(irq: cd->irq, type: IRQ_TYPE_NONE); |
| 107 | |
| 108 | /* Do this last because it synchronously configures the timer */ |
| 109 | clockevents_config_and_register(dev: cd, freq: econet_timer.freq_hz, |
| 110 | ECONET_MIN_DELTA, ECONET_MAX_DELTA); |
| 111 | |
| 112 | return 0; |
| 113 | } |
| 114 | |
| 115 | static u64 notrace sched_clock_read(void) |
| 116 | { |
| 117 | /* Always read from clock zero no matter the CPU */ |
| 118 | return (u64)ioread32(reg_count(timer_n: 0)); |
| 119 | } |
| 120 | |
| 121 | /* Init */ |
| 122 | |
| 123 | static void __init cevt_dev_init(uint cpu) |
| 124 | { |
| 125 | iowrite32(0, reg_count(timer_n: cpu)); |
| 126 | iowrite32(U32_MAX, reg_compare(timer_n: cpu)); |
| 127 | } |
| 128 | |
| 129 | static int __init cevt_init(struct device_node *np) |
| 130 | { |
| 131 | int i, irq, ret; |
| 132 | |
| 133 | irq = irq_of_parse_and_map(node: np, index: 0); |
| 134 | if (irq <= 0) { |
| 135 | pr_err("%pOFn: irq_of_parse_and_map failed" , np); |
| 136 | return -EINVAL; |
| 137 | } |
| 138 | |
| 139 | ret = request_percpu_irq(irq, handler: cevt_interrupt, devname: np->name, percpu_dev_id: &econet_timer_pcpu); |
| 140 | |
| 141 | if (ret < 0) { |
| 142 | pr_err("%pOFn: IRQ %d setup failed (%d)\n" , np, irq, ret); |
| 143 | goto err_unmap_irq; |
| 144 | } |
| 145 | |
| 146 | for_each_possible_cpu(i) { |
| 147 | struct clock_event_device *cd = &per_cpu(econet_timer_pcpu, i); |
| 148 | |
| 149 | cd->rating = 310; |
| 150 | cd->features = CLOCK_EVT_FEAT_ONESHOT | |
| 151 | CLOCK_EVT_FEAT_C3STOP | |
| 152 | CLOCK_EVT_FEAT_PERCPU; |
| 153 | cd->set_next_event = cevt_set_next_event; |
| 154 | cd->irq = irq; |
| 155 | cd->cpumask = cpumask_of(i); |
| 156 | cd->name = np->name; |
| 157 | |
| 158 | cevt_dev_init(cpu: i); |
| 159 | } |
| 160 | |
| 161 | cpuhp_setup_state(state: CPUHP_AP_ONLINE_DYN, |
| 162 | name: "clockevents/econet/timer:starting" , |
| 163 | startup: cevt_init_cpu, NULL); |
| 164 | return 0; |
| 165 | |
| 166 | err_unmap_irq: |
| 167 | irq_dispose_mapping(virq: irq); |
| 168 | return ret; |
| 169 | } |
| 170 | |
| 171 | static int __init timer_init(struct device_node *np) |
| 172 | { |
| 173 | int num_blocks = DIV_ROUND_UP(num_possible_cpus(), 2); |
| 174 | struct clk *clk; |
| 175 | int ret; |
| 176 | |
| 177 | clk = of_clk_get(np, index: 0); |
| 178 | if (IS_ERR(ptr: clk)) { |
| 179 | pr_err("%pOFn: Failed to get CPU clock from DT %ld\n" , np, PTR_ERR(clk)); |
| 180 | return PTR_ERR(ptr: clk); |
| 181 | } |
| 182 | |
| 183 | econet_timer.freq_hz = clk_get_rate(clk); |
| 184 | |
| 185 | for (int i = 0; i < num_blocks; i++) { |
| 186 | econet_timer.membase[i] = of_iomap(node: np, index: i); |
| 187 | if (!econet_timer.membase[i]) { |
| 188 | pr_err("%pOFn: failed to map register [%d]\n" , np, i); |
| 189 | return -ENXIO; |
| 190 | } |
| 191 | } |
| 192 | |
| 193 | /* For clocksource purposes always read clock zero, whatever the CPU */ |
| 194 | ret = clocksource_mmio_init(reg_count(timer_n: 0), np->name, |
| 195 | econet_timer.freq_hz, 301, ECONET_BITS, |
| 196 | clocksource_mmio_readl_up); |
| 197 | if (ret) { |
| 198 | pr_err("%pOFn: clocksource_mmio_init failed: %d" , np, ret); |
| 199 | return ret; |
| 200 | } |
| 201 | |
| 202 | ret = cevt_init(np); |
| 203 | if (ret < 0) |
| 204 | return ret; |
| 205 | |
| 206 | sched_clock_register(read: sched_clock_read, ECONET_BITS, |
| 207 | rate: econet_timer.freq_hz); |
| 208 | |
| 209 | pr_info("%pOFn: using %u.%03u MHz high precision timer\n" , np, |
| 210 | econet_timer.freq_hz / 1000000, |
| 211 | (econet_timer.freq_hz / 1000) % 1000); |
| 212 | |
| 213 | return 0; |
| 214 | } |
| 215 | |
| 216 | TIMER_OF_DECLARE(econet_timer_hpt, "econet,en751221-timer" , timer_init); |
| 217 | |