xref: /openbmc/linux/arch/sh/kernel/unwinder.c (revision 5933f6d2)
15933f6d2SKuninori Morimoto // SPDX-License-Identifier: GPL-2.0
2bf61ad1fSMatt Fleming /*
3bf61ad1fSMatt Fleming  * Copyright (C) 2009  Matt Fleming
4bf61ad1fSMatt Fleming  *
5bf61ad1fSMatt Fleming  * Based, in part, on kernel/time/clocksource.c.
6bf61ad1fSMatt Fleming  *
7bf61ad1fSMatt Fleming  * This file provides arbitration code for stack unwinders.
8bf61ad1fSMatt Fleming  *
9bf61ad1fSMatt Fleming  * Multiple stack unwinders can be available on a system, usually with
10bf61ad1fSMatt Fleming  * the most accurate unwinder being the currently active one.
11bf61ad1fSMatt Fleming  */
12bf61ad1fSMatt Fleming #include <linux/errno.h>
13bf61ad1fSMatt Fleming #include <linux/list.h>
14bf61ad1fSMatt Fleming #include <linux/spinlock.h>
154ab8f241SPaul Mundt #include <linux/module.h>
16bf61ad1fSMatt Fleming #include <asm/unwinder.h>
1760063497SArun Sharma #include <linux/atomic.h>
18bf61ad1fSMatt Fleming 
19bf61ad1fSMatt Fleming /*
20bf61ad1fSMatt Fleming  * This is the most basic stack unwinder an architecture can
21bf61ad1fSMatt Fleming  * provide. For architectures without reliable frame pointers, e.g.
22bf61ad1fSMatt Fleming  * RISC CPUs, it can be implemented by looking through the stack for
23bf61ad1fSMatt Fleming  * addresses that lie within the kernel text section.
24bf61ad1fSMatt Fleming  *
25bf61ad1fSMatt Fleming  * Other CPUs, e.g. x86, can use their frame pointer register to
26bf61ad1fSMatt Fleming  * construct more accurate stack traces.
27bf61ad1fSMatt Fleming  */
28bf61ad1fSMatt Fleming static struct list_head unwinder_list;
29bf61ad1fSMatt Fleming static struct unwinder stack_reader = {
30bf61ad1fSMatt Fleming 	.name = "stack-reader",
31bf61ad1fSMatt Fleming 	.dump = stack_reader_dump,
32bf61ad1fSMatt Fleming 	.rating = 50,
33bf61ad1fSMatt Fleming 	.list = {
34bf61ad1fSMatt Fleming 		.next = &unwinder_list,
35bf61ad1fSMatt Fleming 		.prev = &unwinder_list,
36bf61ad1fSMatt Fleming 	},
37bf61ad1fSMatt Fleming };
38bf61ad1fSMatt Fleming 
39bf61ad1fSMatt Fleming /*
40bf61ad1fSMatt Fleming  * "curr_unwinder" points to the stack unwinder currently in use. This
41bf61ad1fSMatt Fleming  * is the unwinder with the highest rating.
42bf61ad1fSMatt Fleming  *
43bf61ad1fSMatt Fleming  * "unwinder_list" is a linked-list of all available unwinders, sorted
44bf61ad1fSMatt Fleming  * by rating.
45bf61ad1fSMatt Fleming  *
46bf61ad1fSMatt Fleming  * All modifications of "curr_unwinder" and "unwinder_list" must be
47bf61ad1fSMatt Fleming  * performed whilst holding "unwinder_lock".
48bf61ad1fSMatt Fleming  */
49bf61ad1fSMatt Fleming static struct unwinder *curr_unwinder = &stack_reader;
50bf61ad1fSMatt Fleming 
51bf61ad1fSMatt Fleming static struct list_head unwinder_list = {
52bf61ad1fSMatt Fleming 	.next = &stack_reader.list,
53bf61ad1fSMatt Fleming 	.prev = &stack_reader.list,
54bf61ad1fSMatt Fleming };
55bf61ad1fSMatt Fleming 
56bf61ad1fSMatt Fleming static DEFINE_SPINLOCK(unwinder_lock);
57bf61ad1fSMatt Fleming 
58bf61ad1fSMatt Fleming /**
59bf61ad1fSMatt Fleming  * select_unwinder - Select the best registered stack unwinder.
60bf61ad1fSMatt Fleming  *
61bf61ad1fSMatt Fleming  * Private function. Must hold unwinder_lock when called.
62bf61ad1fSMatt Fleming  *
63bf61ad1fSMatt Fleming  * Select the stack unwinder with the best rating. This is useful for
64bf61ad1fSMatt Fleming  * setting up curr_unwinder.
65bf61ad1fSMatt Fleming  */
select_unwinder(void)66bf61ad1fSMatt Fleming static struct unwinder *select_unwinder(void)
67bf61ad1fSMatt Fleming {
68bf61ad1fSMatt Fleming 	struct unwinder *best;
69bf61ad1fSMatt Fleming 
70bf61ad1fSMatt Fleming 	if (list_empty(&unwinder_list))
71bf61ad1fSMatt Fleming 		return NULL;
72bf61ad1fSMatt Fleming 
73bf61ad1fSMatt Fleming 	best = list_entry(unwinder_list.next, struct unwinder, list);
74bf61ad1fSMatt Fleming 	if (best == curr_unwinder)
75bf61ad1fSMatt Fleming 		return NULL;
76bf61ad1fSMatt Fleming 
77bf61ad1fSMatt Fleming 	return best;
78bf61ad1fSMatt Fleming }
79bf61ad1fSMatt Fleming 
80bf61ad1fSMatt Fleming /*
81bf61ad1fSMatt Fleming  * Enqueue the stack unwinder sorted by rating.
82bf61ad1fSMatt Fleming  */
unwinder_enqueue(struct unwinder * ops)83bf61ad1fSMatt Fleming static int unwinder_enqueue(struct unwinder *ops)
84bf61ad1fSMatt Fleming {
85bf61ad1fSMatt Fleming 	struct list_head *tmp, *entry = &unwinder_list;
86bf61ad1fSMatt Fleming 
87bf61ad1fSMatt Fleming 	list_for_each(tmp, &unwinder_list) {
88bf61ad1fSMatt Fleming 		struct unwinder *o;
89bf61ad1fSMatt Fleming 
90bf61ad1fSMatt Fleming 		o = list_entry(tmp, struct unwinder, list);
91bf61ad1fSMatt Fleming 		if (o == ops)
92bf61ad1fSMatt Fleming 			return -EBUSY;
93bf61ad1fSMatt Fleming 		/* Keep track of the place, where to insert */
94bf61ad1fSMatt Fleming 		if (o->rating >= ops->rating)
95bf61ad1fSMatt Fleming 			entry = tmp;
96bf61ad1fSMatt Fleming 	}
97bf61ad1fSMatt Fleming 	list_add(&ops->list, entry);
98bf61ad1fSMatt Fleming 
99bf61ad1fSMatt Fleming 	return 0;
100bf61ad1fSMatt Fleming }
101bf61ad1fSMatt Fleming 
102bf61ad1fSMatt Fleming /**
103bf61ad1fSMatt Fleming  * unwinder_register - Used to install new stack unwinder
104bf61ad1fSMatt Fleming  * @u: unwinder to be registered
105bf61ad1fSMatt Fleming  *
106bf61ad1fSMatt Fleming  * Install the new stack unwinder on the unwinder list, which is sorted
107bf61ad1fSMatt Fleming  * by rating.
108bf61ad1fSMatt Fleming  *
109bf61ad1fSMatt Fleming  * Returns -EBUSY if registration fails, zero otherwise.
110bf61ad1fSMatt Fleming  */
unwinder_register(struct unwinder * u)111bf61ad1fSMatt Fleming int unwinder_register(struct unwinder *u)
112bf61ad1fSMatt Fleming {
113bf61ad1fSMatt Fleming 	unsigned long flags;
114bf61ad1fSMatt Fleming 	int ret;
115bf61ad1fSMatt Fleming 
116bf61ad1fSMatt Fleming 	spin_lock_irqsave(&unwinder_lock, flags);
117bf61ad1fSMatt Fleming 	ret = unwinder_enqueue(u);
118bf61ad1fSMatt Fleming 	if (!ret)
119bf61ad1fSMatt Fleming 		curr_unwinder = select_unwinder();
120bf61ad1fSMatt Fleming 	spin_unlock_irqrestore(&unwinder_lock, flags);
121bf61ad1fSMatt Fleming 
122bf61ad1fSMatt Fleming 	return ret;
123bf61ad1fSMatt Fleming }
124bf61ad1fSMatt Fleming 
125b344e24aSMatt Fleming int unwinder_faulted = 0;
126b344e24aSMatt Fleming 
127bf61ad1fSMatt Fleming /*
128bf61ad1fSMatt Fleming  * Unwind the call stack and pass information to the stacktrace_ops
129bf61ad1fSMatt Fleming  * functions. Also handle the case where we need to switch to a new
130bf61ad1fSMatt Fleming  * stack dumper because the current one faulted unexpectedly.
131bf61ad1fSMatt Fleming  */
unwind_stack(struct task_struct * task,struct pt_regs * regs,unsigned long * sp,const struct stacktrace_ops * ops,void * data)132bf61ad1fSMatt Fleming void unwind_stack(struct task_struct *task, struct pt_regs *regs,
133bf61ad1fSMatt Fleming 		  unsigned long *sp, const struct stacktrace_ops *ops,
134bf61ad1fSMatt Fleming 		  void *data)
135bf61ad1fSMatt Fleming {
136bf61ad1fSMatt Fleming 	unsigned long flags;
137bf61ad1fSMatt Fleming 
138bf61ad1fSMatt Fleming 	/*
139bf61ad1fSMatt Fleming 	 * The problem with unwinders with high ratings is that they are
140bf61ad1fSMatt Fleming 	 * inherently more complicated than the simple ones with lower
141bf61ad1fSMatt Fleming 	 * ratings. We are therefore more likely to fault in the
142bf61ad1fSMatt Fleming 	 * complicated ones, e.g. hitting BUG()s. If we fault in the
143bf61ad1fSMatt Fleming 	 * code for the current stack unwinder we try to downgrade to
144bf61ad1fSMatt Fleming 	 * one with a lower rating.
145bf61ad1fSMatt Fleming 	 *
146bf61ad1fSMatt Fleming 	 * Hopefully this will give us a semi-reliable stacktrace so we
147bf61ad1fSMatt Fleming 	 * can diagnose why curr_unwinder->dump() faulted.
148bf61ad1fSMatt Fleming 	 */
149b344e24aSMatt Fleming 	if (unwinder_faulted) {
150bf61ad1fSMatt Fleming 		spin_lock_irqsave(&unwinder_lock, flags);
151bf61ad1fSMatt Fleming 
152b344e24aSMatt Fleming 		/* Make sure no one beat us to changing the unwinder */
153b344e24aSMatt Fleming 		if (unwinder_faulted && !list_is_singular(&unwinder_list)) {
154bf61ad1fSMatt Fleming 			list_del(&curr_unwinder->list);
155bf61ad1fSMatt Fleming 			curr_unwinder = select_unwinder();
156b344e24aSMatt Fleming 
157b344e24aSMatt Fleming 			unwinder_faulted = 0;
158bf61ad1fSMatt Fleming 		}
159bf61ad1fSMatt Fleming 
160bf61ad1fSMatt Fleming 		spin_unlock_irqrestore(&unwinder_lock, flags);
161bf61ad1fSMatt Fleming 	}
162bf61ad1fSMatt Fleming 
163bf61ad1fSMatt Fleming 	curr_unwinder->dump(task, regs, sp, ops, data);
164b344e24aSMatt Fleming }
1654ab8f241SPaul Mundt EXPORT_SYMBOL_GPL(unwind_stack);
166