xref: /openbmc/linux/lib/ref_tracker.c (revision 86e281fc)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 
3 #define pr_fmt(fmt) "ref_tracker: " fmt
4 
5 #include <linux/export.h>
6 #include <linux/list_sort.h>
7 #include <linux/ref_tracker.h>
8 #include <linux/slab.h>
9 #include <linux/stacktrace.h>
10 #include <linux/stackdepot.h>
11 
12 #define REF_TRACKER_STACK_ENTRIES 16
13 #define STACK_BUF_SIZE 1024
14 
15 struct ref_tracker {
16 	struct list_head	head;   /* anchor into dir->list or dir->quarantine */
17 	bool			dead;
18 	depot_stack_handle_t	alloc_stack_handle;
19 	depot_stack_handle_t	free_stack_handle;
20 };
21 
22 struct ref_tracker_dir_stats {
23 	int total;
24 	int count;
25 	struct {
26 		depot_stack_handle_t stack_handle;
27 		unsigned int count;
28 	} stacks[];
29 };
30 
31 static struct ref_tracker_dir_stats *
32 ref_tracker_get_stats(struct ref_tracker_dir *dir, unsigned int limit)
33 {
34 	struct ref_tracker_dir_stats *stats;
35 	struct ref_tracker *tracker;
36 
37 	stats = kmalloc(struct_size(stats, stacks, limit),
38 			GFP_NOWAIT | __GFP_NOWARN);
39 	if (!stats)
40 		return ERR_PTR(-ENOMEM);
41 	stats->total = 0;
42 	stats->count = 0;
43 
44 	list_for_each_entry(tracker, &dir->list, head) {
45 		depot_stack_handle_t stack = tracker->alloc_stack_handle;
46 		int i;
47 
48 		++stats->total;
49 		for (i = 0; i < stats->count; ++i)
50 			if (stats->stacks[i].stack_handle == stack)
51 				break;
52 		if (i >= limit)
53 			continue;
54 		if (i >= stats->count) {
55 			stats->stacks[i].stack_handle = stack;
56 			stats->stacks[i].count = 0;
57 			++stats->count;
58 		}
59 		++stats->stacks[i].count;
60 	}
61 
62 	return stats;
63 }
64 
65 struct ostream {
66 	char *buf;
67 	int size, used;
68 };
69 
70 #define pr_ostream(stream, fmt, args...) \
71 ({ \
72 	struct ostream *_s = (stream); \
73 \
74 	if (!_s->buf) { \
75 		pr_err(fmt, ##args); \
76 	} else { \
77 		int ret, len = _s->size - _s->used; \
78 		ret = snprintf(_s->buf + _s->used, len, pr_fmt(fmt), ##args); \
79 		_s->used += min(ret, len); \
80 	} \
81 })
82 
83 static void
84 __ref_tracker_dir_pr_ostream(struct ref_tracker_dir *dir,
85 			     unsigned int display_limit, struct ostream *s)
86 {
87 	struct ref_tracker_dir_stats *stats;
88 	unsigned int i = 0, skipped;
89 	depot_stack_handle_t stack;
90 	char *sbuf;
91 
92 	lockdep_assert_held(&dir->lock);
93 
94 	if (list_empty(&dir->list))
95 		return;
96 
97 	stats = ref_tracker_get_stats(dir, display_limit);
98 	if (IS_ERR(stats)) {
99 		pr_ostream(s, "%s@%pK: couldn't get stats, error %pe\n",
100 			   dir->name, dir, stats);
101 		return;
102 	}
103 
104 	sbuf = kmalloc(STACK_BUF_SIZE, GFP_NOWAIT | __GFP_NOWARN);
105 
106 	for (i = 0, skipped = stats->total; i < stats->count; ++i) {
107 		stack = stats->stacks[i].stack_handle;
108 		if (sbuf && !stack_depot_snprint(stack, sbuf, STACK_BUF_SIZE, 4))
109 			sbuf[0] = 0;
110 		pr_ostream(s, "%s@%pK has %d/%d users at\n%s\n", dir->name, dir,
111 			   stats->stacks[i].count, stats->total, sbuf);
112 		skipped -= stats->stacks[i].count;
113 	}
114 
115 	if (skipped)
116 		pr_ostream(s, "%s@%pK skipped reports about %d/%d users.\n",
117 			   dir->name, dir, skipped, stats->total);
118 
119 	kfree(sbuf);
120 
121 	kfree(stats);
122 }
123 
124 void ref_tracker_dir_print_locked(struct ref_tracker_dir *dir,
125 				  unsigned int display_limit)
126 {
127 	struct ostream os = {};
128 
129 	__ref_tracker_dir_pr_ostream(dir, display_limit, &os);
130 }
131 EXPORT_SYMBOL(ref_tracker_dir_print_locked);
132 
133 void ref_tracker_dir_print(struct ref_tracker_dir *dir,
134 			   unsigned int display_limit)
135 {
136 	unsigned long flags;
137 
138 	spin_lock_irqsave(&dir->lock, flags);
139 	ref_tracker_dir_print_locked(dir, display_limit);
140 	spin_unlock_irqrestore(&dir->lock, flags);
141 }
142 EXPORT_SYMBOL(ref_tracker_dir_print);
143 
144 int ref_tracker_dir_snprint(struct ref_tracker_dir *dir, char *buf, size_t size)
145 {
146 	struct ostream os = { .buf = buf, .size = size };
147 	unsigned long flags;
148 
149 	spin_lock_irqsave(&dir->lock, flags);
150 	__ref_tracker_dir_pr_ostream(dir, 16, &os);
151 	spin_unlock_irqrestore(&dir->lock, flags);
152 
153 	return os.used;
154 }
155 EXPORT_SYMBOL(ref_tracker_dir_snprint);
156 
157 void ref_tracker_dir_exit(struct ref_tracker_dir *dir)
158 {
159 	struct ref_tracker *tracker, *n;
160 	unsigned long flags;
161 	bool leak = false;
162 
163 	dir->dead = true;
164 	spin_lock_irqsave(&dir->lock, flags);
165 	list_for_each_entry_safe(tracker, n, &dir->quarantine, head) {
166 		list_del(&tracker->head);
167 		kfree(tracker);
168 		dir->quarantine_avail++;
169 	}
170 	if (!list_empty(&dir->list)) {
171 		ref_tracker_dir_print_locked(dir, 16);
172 		leak = true;
173 		list_for_each_entry_safe(tracker, n, &dir->list, head) {
174 			list_del(&tracker->head);
175 			kfree(tracker);
176 		}
177 	}
178 	spin_unlock_irqrestore(&dir->lock, flags);
179 	WARN_ON_ONCE(leak);
180 	WARN_ON_ONCE(refcount_read(&dir->untracked) != 1);
181 	WARN_ON_ONCE(refcount_read(&dir->no_tracker) != 1);
182 }
183 EXPORT_SYMBOL(ref_tracker_dir_exit);
184 
185 int ref_tracker_alloc(struct ref_tracker_dir *dir,
186 		      struct ref_tracker **trackerp,
187 		      gfp_t gfp)
188 {
189 	unsigned long entries[REF_TRACKER_STACK_ENTRIES];
190 	struct ref_tracker *tracker;
191 	unsigned int nr_entries;
192 	gfp_t gfp_mask = gfp | __GFP_NOWARN;
193 	unsigned long flags;
194 
195 	WARN_ON_ONCE(dir->dead);
196 
197 	if (!trackerp) {
198 		refcount_inc(&dir->no_tracker);
199 		return 0;
200 	}
201 	if (gfp & __GFP_DIRECT_RECLAIM)
202 		gfp_mask |= __GFP_NOFAIL;
203 	*trackerp = tracker = kzalloc(sizeof(*tracker), gfp_mask);
204 	if (unlikely(!tracker)) {
205 		pr_err_once("memory allocation failure, unreliable refcount tracker.\n");
206 		refcount_inc(&dir->untracked);
207 		return -ENOMEM;
208 	}
209 	nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 1);
210 	tracker->alloc_stack_handle = stack_depot_save(entries, nr_entries, gfp);
211 
212 	spin_lock_irqsave(&dir->lock, flags);
213 	list_add(&tracker->head, &dir->list);
214 	spin_unlock_irqrestore(&dir->lock, flags);
215 	return 0;
216 }
217 EXPORT_SYMBOL_GPL(ref_tracker_alloc);
218 
219 int ref_tracker_free(struct ref_tracker_dir *dir,
220 		     struct ref_tracker **trackerp)
221 {
222 	unsigned long entries[REF_TRACKER_STACK_ENTRIES];
223 	depot_stack_handle_t stack_handle;
224 	struct ref_tracker *tracker;
225 	unsigned int nr_entries;
226 	unsigned long flags;
227 
228 	WARN_ON_ONCE(dir->dead);
229 
230 	if (!trackerp) {
231 		refcount_dec(&dir->no_tracker);
232 		return 0;
233 	}
234 	tracker = *trackerp;
235 	if (!tracker) {
236 		refcount_dec(&dir->untracked);
237 		return -EEXIST;
238 	}
239 	nr_entries = stack_trace_save(entries, ARRAY_SIZE(entries), 1);
240 	stack_handle = stack_depot_save(entries, nr_entries,
241 					GFP_NOWAIT | __GFP_NOWARN);
242 
243 	spin_lock_irqsave(&dir->lock, flags);
244 	if (tracker->dead) {
245 		pr_err("reference already released.\n");
246 		if (tracker->alloc_stack_handle) {
247 			pr_err("allocated in:\n");
248 			stack_depot_print(tracker->alloc_stack_handle);
249 		}
250 		if (tracker->free_stack_handle) {
251 			pr_err("freed in:\n");
252 			stack_depot_print(tracker->free_stack_handle);
253 		}
254 		spin_unlock_irqrestore(&dir->lock, flags);
255 		WARN_ON_ONCE(1);
256 		return -EINVAL;
257 	}
258 	tracker->dead = true;
259 
260 	tracker->free_stack_handle = stack_handle;
261 
262 	list_move_tail(&tracker->head, &dir->quarantine);
263 	if (!dir->quarantine_avail) {
264 		tracker = list_first_entry(&dir->quarantine, struct ref_tracker, head);
265 		list_del(&tracker->head);
266 	} else {
267 		dir->quarantine_avail--;
268 		tracker = NULL;
269 	}
270 	spin_unlock_irqrestore(&dir->lock, flags);
271 
272 	kfree(tracker);
273 	return 0;
274 }
275 EXPORT_SYMBOL_GPL(ref_tracker_free);
276