1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * IRQ subsystem internal functions and variables: 4 * 5 * Do not ever include this file from anything else than 6 * kernel/irq/. Do not even think about using any information outside 7 * of this file for your non core code. 8 */ 9 #include <linux/irqdesc.h> 10 #include <linux/kernel_stat.h> 11 #include <linux/pm_runtime.h> 12 #include <linux/sched/clock.h> 13 14 #ifdef CONFIG_SPARSE_IRQ 15 # define IRQ_BITMAP_BITS (NR_IRQS + 8196) 16 #else 17 # define IRQ_BITMAP_BITS NR_IRQS 18 #endif 19 20 #define istate core_internal_state__do_not_mess_with_it 21 22 extern bool noirqdebug; 23 24 extern struct irqaction chained_action; 25 26 /* 27 * Bits used by threaded handlers: 28 * IRQTF_RUNTHREAD - signals that the interrupt handler thread should run 29 * IRQTF_WARNED - warning "IRQ_WAKE_THREAD w/o thread_fn" has been printed 30 * IRQTF_AFFINITY - irq thread is requested to adjust affinity 31 * IRQTF_FORCED_THREAD - irq action is force threaded 32 */ 33 enum { 34 IRQTF_RUNTHREAD, 35 IRQTF_WARNED, 36 IRQTF_AFFINITY, 37 IRQTF_FORCED_THREAD, 38 }; 39 40 /* 41 * Bit masks for desc->core_internal_state__do_not_mess_with_it 42 * 43 * IRQS_AUTODETECT - autodetection in progress 44 * IRQS_SPURIOUS_DISABLED - was disabled due to spurious interrupt 45 * detection 46 * IRQS_POLL_INPROGRESS - polling in progress 47 * IRQS_ONESHOT - irq is not unmasked in primary handler 48 * IRQS_REPLAY - irq is replayed 49 * IRQS_WAITING - irq is waiting 50 * IRQS_PENDING - irq is pending and replayed later 51 * IRQS_SUSPENDED - irq is suspended 52 * IRQS_NMI - irq line is used to deliver NMIs 53 */ 54 enum { 55 IRQS_AUTODETECT = 0x00000001, 56 IRQS_SPURIOUS_DISABLED = 0x00000002, 57 IRQS_POLL_INPROGRESS = 0x00000008, 58 IRQS_ONESHOT = 0x00000020, 59 IRQS_REPLAY = 0x00000040, 60 IRQS_WAITING = 0x00000080, 61 IRQS_PENDING = 0x00000200, 62 IRQS_SUSPENDED = 0x00000800, 63 IRQS_TIMINGS = 0x00001000, 64 IRQS_NMI = 0x00002000, 65 }; 66 67 #include "debug.h" 68 #include "settings.h" 69 70 extern int __irq_set_trigger(struct irq_desc *desc, unsigned long flags); 71 extern void __disable_irq(struct irq_desc *desc); 72 extern void __enable_irq(struct irq_desc *desc); 73 74 #define IRQ_RESEND true 75 #define IRQ_NORESEND false 76 77 #define IRQ_START_FORCE true 78 #define IRQ_START_COND false 79 80 extern int irq_activate(struct irq_desc *desc); 81 extern int irq_activate_and_startup(struct irq_desc *desc, bool resend); 82 extern int irq_startup(struct irq_desc *desc, bool resend, bool force); 83 84 extern void irq_shutdown(struct irq_desc *desc); 85 extern void irq_enable(struct irq_desc *desc); 86 extern void irq_disable(struct irq_desc *desc); 87 extern void irq_percpu_enable(struct irq_desc *desc, unsigned int cpu); 88 extern void irq_percpu_disable(struct irq_desc *desc, unsigned int cpu); 89 extern void mask_irq(struct irq_desc *desc); 90 extern void unmask_irq(struct irq_desc *desc); 91 extern void unmask_threaded_irq(struct irq_desc *desc); 92 93 #ifdef CONFIG_SPARSE_IRQ 94 static inline void irq_mark_irq(unsigned int irq) { } 95 #else 96 extern void irq_mark_irq(unsigned int irq); 97 #endif 98 99 extern void init_kstat_irqs(struct irq_desc *desc, int node, int nr); 100 101 irqreturn_t __handle_irq_event_percpu(struct irq_desc *desc, unsigned int *flags); 102 irqreturn_t handle_irq_event_percpu(struct irq_desc *desc); 103 irqreturn_t handle_irq_event(struct irq_desc *desc); 104 105 /* Resending of interrupts :*/ 106 void check_irq_resend(struct irq_desc *desc); 107 bool irq_wait_for_poll(struct irq_desc *desc); 108 void __irq_wake_thread(struct irq_desc *desc, struct irqaction *action); 109 110 #ifdef CONFIG_PROC_FS 111 extern void register_irq_proc(unsigned int irq, struct irq_desc *desc); 112 extern void unregister_irq_proc(unsigned int irq, struct irq_desc *desc); 113 extern void register_handler_proc(unsigned int irq, struct irqaction *action); 114 extern void unregister_handler_proc(unsigned int irq, struct irqaction *action); 115 #else 116 static inline void register_irq_proc(unsigned int irq, struct irq_desc *desc) { } 117 static inline void unregister_irq_proc(unsigned int irq, struct irq_desc *desc) { } 118 static inline void register_handler_proc(unsigned int irq, 119 struct irqaction *action) { } 120 static inline void unregister_handler_proc(unsigned int irq, 121 struct irqaction *action) { } 122 #endif 123 124 extern bool irq_can_set_affinity_usr(unsigned int irq); 125 126 extern int irq_select_affinity_usr(unsigned int irq); 127 128 extern void irq_set_thread_affinity(struct irq_desc *desc); 129 130 extern int irq_do_set_affinity(struct irq_data *data, 131 const struct cpumask *dest, bool force); 132 133 #ifdef CONFIG_SMP 134 extern int irq_setup_affinity(struct irq_desc *desc); 135 #else 136 static inline int irq_setup_affinity(struct irq_desc *desc) { return 0; } 137 #endif 138 139 /* Inline functions for support of irq chips on slow busses */ 140 static inline void chip_bus_lock(struct irq_desc *desc) 141 { 142 if (unlikely(desc->irq_data.chip->irq_bus_lock)) 143 desc->irq_data.chip->irq_bus_lock(&desc->irq_data); 144 } 145 146 static inline void chip_bus_sync_unlock(struct irq_desc *desc) 147 { 148 if (unlikely(desc->irq_data.chip->irq_bus_sync_unlock)) 149 desc->irq_data.chip->irq_bus_sync_unlock(&desc->irq_data); 150 } 151 152 #define _IRQ_DESC_CHECK (1 << 0) 153 #define _IRQ_DESC_PERCPU (1 << 1) 154 155 #define IRQ_GET_DESC_CHECK_GLOBAL (_IRQ_DESC_CHECK) 156 #define IRQ_GET_DESC_CHECK_PERCPU (_IRQ_DESC_CHECK | _IRQ_DESC_PERCPU) 157 158 #define for_each_action_of_desc(desc, act) \ 159 for (act = desc->action; act; act = act->next) 160 161 struct irq_desc * 162 __irq_get_desc_lock(unsigned int irq, unsigned long *flags, bool bus, 163 unsigned int check); 164 void __irq_put_desc_unlock(struct irq_desc *desc, unsigned long flags, bool bus); 165 166 static inline struct irq_desc * 167 irq_get_desc_buslock(unsigned int irq, unsigned long *flags, unsigned int check) 168 { 169 return __irq_get_desc_lock(irq, flags, true, check); 170 } 171 172 static inline void 173 irq_put_desc_busunlock(struct irq_desc *desc, unsigned long flags) 174 { 175 __irq_put_desc_unlock(desc, flags, true); 176 } 177 178 static inline struct irq_desc * 179 irq_get_desc_lock(unsigned int irq, unsigned long *flags, unsigned int check) 180 { 181 return __irq_get_desc_lock(irq, flags, false, check); 182 } 183 184 static inline void 185 irq_put_desc_unlock(struct irq_desc *desc, unsigned long flags) 186 { 187 __irq_put_desc_unlock(desc, flags, false); 188 } 189 190 #define __irqd_to_state(d) ACCESS_PRIVATE((d)->common, state_use_accessors) 191 192 static inline unsigned int irqd_get(struct irq_data *d) 193 { 194 return __irqd_to_state(d); 195 } 196 197 /* 198 * Manipulation functions for irq_data.state 199 */ 200 static inline void irqd_set_move_pending(struct irq_data *d) 201 { 202 __irqd_to_state(d) |= IRQD_SETAFFINITY_PENDING; 203 } 204 205 static inline void irqd_clr_move_pending(struct irq_data *d) 206 { 207 __irqd_to_state(d) &= ~IRQD_SETAFFINITY_PENDING; 208 } 209 210 static inline void irqd_set_managed_shutdown(struct irq_data *d) 211 { 212 __irqd_to_state(d) |= IRQD_MANAGED_SHUTDOWN; 213 } 214 215 static inline void irqd_clr_managed_shutdown(struct irq_data *d) 216 { 217 __irqd_to_state(d) &= ~IRQD_MANAGED_SHUTDOWN; 218 } 219 220 static inline void irqd_clear(struct irq_data *d, unsigned int mask) 221 { 222 __irqd_to_state(d) &= ~mask; 223 } 224 225 static inline void irqd_set(struct irq_data *d, unsigned int mask) 226 { 227 __irqd_to_state(d) |= mask; 228 } 229 230 static inline bool irqd_has_set(struct irq_data *d, unsigned int mask) 231 { 232 return __irqd_to_state(d) & mask; 233 } 234 235 static inline void irq_state_set_disabled(struct irq_desc *desc) 236 { 237 irqd_set(&desc->irq_data, IRQD_IRQ_DISABLED); 238 } 239 240 static inline void irq_state_set_masked(struct irq_desc *desc) 241 { 242 irqd_set(&desc->irq_data, IRQD_IRQ_MASKED); 243 } 244 245 #undef __irqd_to_state 246 247 static inline void __kstat_incr_irqs_this_cpu(struct irq_desc *desc) 248 { 249 __this_cpu_inc(*desc->kstat_irqs); 250 __this_cpu_inc(kstat.irqs_sum); 251 } 252 253 static inline void kstat_incr_irqs_this_cpu(struct irq_desc *desc) 254 { 255 __kstat_incr_irqs_this_cpu(desc); 256 desc->tot_count++; 257 } 258 259 static inline int irq_desc_get_node(struct irq_desc *desc) 260 { 261 return irq_common_data_get_node(&desc->irq_common_data); 262 } 263 264 static inline int irq_desc_is_chained(struct irq_desc *desc) 265 { 266 return (desc->action && desc->action == &chained_action); 267 } 268 269 #ifdef CONFIG_PM_SLEEP 270 bool irq_pm_check_wakeup(struct irq_desc *desc); 271 void irq_pm_install_action(struct irq_desc *desc, struct irqaction *action); 272 void irq_pm_remove_action(struct irq_desc *desc, struct irqaction *action); 273 #else 274 static inline bool irq_pm_check_wakeup(struct irq_desc *desc) { return false; } 275 static inline void 276 irq_pm_install_action(struct irq_desc *desc, struct irqaction *action) { } 277 static inline void 278 irq_pm_remove_action(struct irq_desc *desc, struct irqaction *action) { } 279 #endif 280 281 #ifdef CONFIG_IRQ_TIMINGS 282 283 #define IRQ_TIMINGS_SHIFT 5 284 #define IRQ_TIMINGS_SIZE (1 << IRQ_TIMINGS_SHIFT) 285 #define IRQ_TIMINGS_MASK (IRQ_TIMINGS_SIZE - 1) 286 287 /** 288 * struct irq_timings - irq timings storing structure 289 * @values: a circular buffer of u64 encoded <timestamp,irq> values 290 * @count: the number of elements in the array 291 */ 292 struct irq_timings { 293 u64 values[IRQ_TIMINGS_SIZE]; 294 int count; 295 }; 296 297 DECLARE_PER_CPU(struct irq_timings, irq_timings); 298 299 extern void irq_timings_free(int irq); 300 extern int irq_timings_alloc(int irq); 301 302 static inline void irq_remove_timings(struct irq_desc *desc) 303 { 304 desc->istate &= ~IRQS_TIMINGS; 305 306 irq_timings_free(irq_desc_get_irq(desc)); 307 } 308 309 static inline void irq_setup_timings(struct irq_desc *desc, struct irqaction *act) 310 { 311 int irq = irq_desc_get_irq(desc); 312 int ret; 313 314 /* 315 * We don't need the measurement because the idle code already 316 * knows the next expiry event. 317 */ 318 if (act->flags & __IRQF_TIMER) 319 return; 320 321 /* 322 * In case the timing allocation fails, we just want to warn, 323 * not fail, so letting the system boot anyway. 324 */ 325 ret = irq_timings_alloc(irq); 326 if (ret) { 327 pr_warn("Failed to allocate irq timing stats for irq%d (%d)", 328 irq, ret); 329 return; 330 } 331 332 desc->istate |= IRQS_TIMINGS; 333 } 334 335 extern void irq_timings_enable(void); 336 extern void irq_timings_disable(void); 337 338 DECLARE_STATIC_KEY_FALSE(irq_timing_enabled); 339 340 /* 341 * The interrupt number and the timestamp are encoded into a single 342 * u64 variable to optimize the size. 343 * 48 bit time stamp and 16 bit IRQ number is way sufficient. 344 * Who cares an IRQ after 78 hours of idle time? 345 */ 346 static inline u64 irq_timing_encode(u64 timestamp, int irq) 347 { 348 return (timestamp << 16) | irq; 349 } 350 351 static inline int irq_timing_decode(u64 value, u64 *timestamp) 352 { 353 *timestamp = value >> 16; 354 return value & U16_MAX; 355 } 356 357 /* 358 * The function record_irq_time is only called in one place in the 359 * interrupts handler. We want this function always inline so the code 360 * inside is embedded in the function and the static key branching 361 * code can act at the higher level. Without the explicit 362 * __always_inline we can end up with a function call and a small 363 * overhead in the hotpath for nothing. 364 */ 365 static __always_inline void record_irq_time(struct irq_desc *desc) 366 { 367 if (!static_branch_likely(&irq_timing_enabled)) 368 return; 369 370 if (desc->istate & IRQS_TIMINGS) { 371 struct irq_timings *timings = this_cpu_ptr(&irq_timings); 372 373 timings->values[timings->count & IRQ_TIMINGS_MASK] = 374 irq_timing_encode(local_clock(), 375 irq_desc_get_irq(desc)); 376 377 timings->count++; 378 } 379 } 380 #else 381 static inline void irq_remove_timings(struct irq_desc *desc) {} 382 static inline void irq_setup_timings(struct irq_desc *desc, 383 struct irqaction *act) {}; 384 static inline void record_irq_time(struct irq_desc *desc) {} 385 #endif /* CONFIG_IRQ_TIMINGS */ 386 387 388 #ifdef CONFIG_GENERIC_IRQ_CHIP 389 void irq_init_generic_chip(struct irq_chip_generic *gc, const char *name, 390 int num_ct, unsigned int irq_base, 391 void __iomem *reg_base, irq_flow_handler_t handler); 392 #else 393 static inline void 394 irq_init_generic_chip(struct irq_chip_generic *gc, const char *name, 395 int num_ct, unsigned int irq_base, 396 void __iomem *reg_base, irq_flow_handler_t handler) { } 397 #endif /* CONFIG_GENERIC_IRQ_CHIP */ 398 399 #ifdef CONFIG_GENERIC_PENDING_IRQ 400 static inline bool irq_can_move_pcntxt(struct irq_data *data) 401 { 402 return irqd_can_move_in_process_context(data); 403 } 404 static inline bool irq_move_pending(struct irq_data *data) 405 { 406 return irqd_is_setaffinity_pending(data); 407 } 408 static inline void 409 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) 410 { 411 cpumask_copy(desc->pending_mask, mask); 412 } 413 static inline void 414 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) 415 { 416 cpumask_copy(mask, desc->pending_mask); 417 } 418 static inline struct cpumask *irq_desc_get_pending_mask(struct irq_desc *desc) 419 { 420 return desc->pending_mask; 421 } 422 bool irq_fixup_move_pending(struct irq_desc *desc, bool force_clear); 423 #else /* CONFIG_GENERIC_PENDING_IRQ */ 424 static inline bool irq_can_move_pcntxt(struct irq_data *data) 425 { 426 return true; 427 } 428 static inline bool irq_move_pending(struct irq_data *data) 429 { 430 return false; 431 } 432 static inline void 433 irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask) 434 { 435 } 436 static inline void 437 irq_get_pending(struct cpumask *mask, struct irq_desc *desc) 438 { 439 } 440 static inline struct cpumask *irq_desc_get_pending_mask(struct irq_desc *desc) 441 { 442 return NULL; 443 } 444 static inline bool irq_fixup_move_pending(struct irq_desc *desc, bool fclear) 445 { 446 return false; 447 } 448 #endif /* !CONFIG_GENERIC_PENDING_IRQ */ 449 450 #if !defined(CONFIG_IRQ_DOMAIN) || !defined(CONFIG_IRQ_DOMAIN_HIERARCHY) 451 static inline int irq_domain_activate_irq(struct irq_data *data, bool reserve) 452 { 453 irqd_set_activated(data); 454 return 0; 455 } 456 static inline void irq_domain_deactivate_irq(struct irq_data *data) 457 { 458 irqd_clr_activated(data); 459 } 460 #endif 461 462 #ifdef CONFIG_GENERIC_IRQ_DEBUGFS 463 #include <linux/debugfs.h> 464 465 void irq_add_debugfs_entry(unsigned int irq, struct irq_desc *desc); 466 static inline void irq_remove_debugfs_entry(struct irq_desc *desc) 467 { 468 debugfs_remove(desc->debugfs_file); 469 kfree(desc->dev_name); 470 } 471 void irq_debugfs_copy_devname(int irq, struct device *dev); 472 # ifdef CONFIG_IRQ_DOMAIN 473 void irq_domain_debugfs_init(struct dentry *root); 474 # else 475 static inline void irq_domain_debugfs_init(struct dentry *root) 476 { 477 } 478 # endif 479 #else /* CONFIG_GENERIC_IRQ_DEBUGFS */ 480 static inline void irq_add_debugfs_entry(unsigned int irq, struct irq_desc *d) 481 { 482 } 483 static inline void irq_remove_debugfs_entry(struct irq_desc *d) 484 { 485 } 486 static inline void irq_debugfs_copy_devname(int irq, struct device *dev) 487 { 488 } 489 #endif /* CONFIG_GENERIC_IRQ_DEBUGFS */ 490