1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (c) 2017 The Linux Foundation. All rights reserved.
3  */
4 
5 #include "msm_gem.h"
6 #include "a5xx_gpu.h"
7 
8 /*
9  * Try to transition the preemption state from old to new. Return
10  * true on success or false if the original state wasn't 'old'
11  */
12 static inline bool try_preempt_state(struct a5xx_gpu *a5xx_gpu,
13 		enum preempt_state old, enum preempt_state new)
14 {
15 	enum preempt_state cur = atomic_cmpxchg(&a5xx_gpu->preempt_state,
16 		old, new);
17 
18 	return (cur == old);
19 }
20 
21 /*
22  * Force the preemption state to the specified state.  This is used in cases
23  * where the current state is known and won't change
24  */
25 static inline void set_preempt_state(struct a5xx_gpu *gpu,
26 		enum preempt_state new)
27 {
28 	/*
29 	 * preempt_state may be read by other cores trying to trigger a
30 	 * preemption or in the interrupt handler so barriers are needed
31 	 * before...
32 	 */
33 	smp_mb__before_atomic();
34 	atomic_set(&gpu->preempt_state, new);
35 	/* ... and after*/
36 	smp_mb__after_atomic();
37 }
38 
39 /* Write the most recent wptr for the given ring into the hardware */
40 static inline void update_wptr(struct msm_gpu *gpu, struct msm_ringbuffer *ring)
41 {
42 	unsigned long flags;
43 	uint32_t wptr;
44 
45 	if (!ring)
46 		return;
47 
48 	spin_lock_irqsave(&ring->lock, flags);
49 	wptr = get_wptr(ring);
50 	spin_unlock_irqrestore(&ring->lock, flags);
51 
52 	gpu_write(gpu, REG_A5XX_CP_RB_WPTR, wptr);
53 }
54 
55 /* Return the highest priority ringbuffer with something in it */
56 static struct msm_ringbuffer *get_next_ring(struct msm_gpu *gpu)
57 {
58 	unsigned long flags;
59 	int i;
60 
61 	for (i = 0; i < gpu->nr_rings; i++) {
62 		bool empty;
63 		struct msm_ringbuffer *ring = gpu->rb[i];
64 
65 		spin_lock_irqsave(&ring->lock, flags);
66 		empty = (get_wptr(ring) == ring->memptrs->rptr);
67 		spin_unlock_irqrestore(&ring->lock, flags);
68 
69 		if (!empty)
70 			return ring;
71 	}
72 
73 	return NULL;
74 }
75 
76 static void a5xx_preempt_timer(struct timer_list *t)
77 {
78 	struct a5xx_gpu *a5xx_gpu = from_timer(a5xx_gpu, t, preempt_timer);
79 	struct msm_gpu *gpu = &a5xx_gpu->base.base;
80 	struct drm_device *dev = gpu->dev;
81 	struct msm_drm_private *priv = dev->dev_private;
82 
83 	if (!try_preempt_state(a5xx_gpu, PREEMPT_TRIGGERED, PREEMPT_FAULTED))
84 		return;
85 
86 	DRM_DEV_ERROR(dev->dev, "%s: preemption timed out\n", gpu->name);
87 	queue_work(priv->wq, &gpu->recover_work);
88 }
89 
90 /* Try to trigger a preemption switch */
91 void a5xx_preempt_trigger(struct msm_gpu *gpu)
92 {
93 	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
94 	struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
95 	unsigned long flags;
96 	struct msm_ringbuffer *ring;
97 
98 	if (gpu->nr_rings == 1)
99 		return;
100 
101 	/*
102 	 * Try to start preemption by moving from NONE to START. If
103 	 * unsuccessful, a preemption is already in flight
104 	 */
105 	if (!try_preempt_state(a5xx_gpu, PREEMPT_NONE, PREEMPT_START))
106 		return;
107 
108 	/* Get the next ring to preempt to */
109 	ring = get_next_ring(gpu);
110 
111 	/*
112 	 * If no ring is populated or the highest priority ring is the current
113 	 * one do nothing except to update the wptr to the latest and greatest
114 	 */
115 	if (!ring || (a5xx_gpu->cur_ring == ring)) {
116 		/*
117 		 * Its possible that while a preemption request is in progress
118 		 * from an irq context, a user context trying to submit might
119 		 * fail to update the write pointer, because it determines
120 		 * that the preempt state is not PREEMPT_NONE.
121 		 *
122 		 * Close the race by introducing an intermediate
123 		 * state PREEMPT_ABORT to let the submit path
124 		 * know that the ringbuffer is not going to change
125 		 * and can safely update the write pointer.
126 		 */
127 
128 		set_preempt_state(a5xx_gpu, PREEMPT_ABORT);
129 		update_wptr(gpu, a5xx_gpu->cur_ring);
130 		set_preempt_state(a5xx_gpu, PREEMPT_NONE);
131 		return;
132 	}
133 
134 	/* Make sure the wptr doesn't update while we're in motion */
135 	spin_lock_irqsave(&ring->lock, flags);
136 	a5xx_gpu->preempt[ring->id]->wptr = get_wptr(ring);
137 	spin_unlock_irqrestore(&ring->lock, flags);
138 
139 	/* Set the address of the incoming preemption record */
140 	gpu_write64(gpu, REG_A5XX_CP_CONTEXT_SWITCH_RESTORE_ADDR_LO,
141 		REG_A5XX_CP_CONTEXT_SWITCH_RESTORE_ADDR_HI,
142 		a5xx_gpu->preempt_iova[ring->id]);
143 
144 	a5xx_gpu->next_ring = ring;
145 
146 	/* Start a timer to catch a stuck preemption */
147 	mod_timer(&a5xx_gpu->preempt_timer, jiffies + msecs_to_jiffies(10000));
148 
149 	/* Set the preemption state to triggered */
150 	set_preempt_state(a5xx_gpu, PREEMPT_TRIGGERED);
151 
152 	/* Make sure everything is written before hitting the button */
153 	wmb();
154 
155 	/* And actually start the preemption */
156 	gpu_write(gpu, REG_A5XX_CP_CONTEXT_SWITCH_CNTL, 1);
157 }
158 
159 void a5xx_preempt_irq(struct msm_gpu *gpu)
160 {
161 	uint32_t status;
162 	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
163 	struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
164 	struct drm_device *dev = gpu->dev;
165 	struct msm_drm_private *priv = dev->dev_private;
166 
167 	if (!try_preempt_state(a5xx_gpu, PREEMPT_TRIGGERED, PREEMPT_PENDING))
168 		return;
169 
170 	/* Delete the preemption watchdog timer */
171 	del_timer(&a5xx_gpu->preempt_timer);
172 
173 	/*
174 	 * The hardware should be setting CP_CONTEXT_SWITCH_CNTL to zero before
175 	 * firing the interrupt, but there is a non zero chance of a hardware
176 	 * condition or a software race that could set it again before we have a
177 	 * chance to finish. If that happens, log and go for recovery
178 	 */
179 	status = gpu_read(gpu, REG_A5XX_CP_CONTEXT_SWITCH_CNTL);
180 	if (unlikely(status)) {
181 		set_preempt_state(a5xx_gpu, PREEMPT_FAULTED);
182 		DRM_DEV_ERROR(dev->dev, "%s: Preemption failed to complete\n",
183 			gpu->name);
184 		queue_work(priv->wq, &gpu->recover_work);
185 		return;
186 	}
187 
188 	a5xx_gpu->cur_ring = a5xx_gpu->next_ring;
189 	a5xx_gpu->next_ring = NULL;
190 
191 	update_wptr(gpu, a5xx_gpu->cur_ring);
192 
193 	set_preempt_state(a5xx_gpu, PREEMPT_NONE);
194 }
195 
196 void a5xx_preempt_hw_init(struct msm_gpu *gpu)
197 {
198 	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
199 	struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
200 	int i;
201 
202 	/* Always come up on rb 0 */
203 	a5xx_gpu->cur_ring = gpu->rb[0];
204 
205 	/* No preemption if we only have one ring */
206 	if (gpu->nr_rings == 1)
207 		return;
208 
209 	for (i = 0; i < gpu->nr_rings; i++) {
210 		a5xx_gpu->preempt[i]->wptr = 0;
211 		a5xx_gpu->preempt[i]->rptr = 0;
212 		a5xx_gpu->preempt[i]->rbase = gpu->rb[i]->iova;
213 	}
214 
215 	/* Write a 0 to signal that we aren't switching pagetables */
216 	gpu_write64(gpu, REG_A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_LO,
217 		REG_A5XX_CP_CONTEXT_SWITCH_SMMU_INFO_HI, 0);
218 
219 	/* Reset the preemption state */
220 	set_preempt_state(a5xx_gpu, PREEMPT_NONE);
221 }
222 
223 static int preempt_init_ring(struct a5xx_gpu *a5xx_gpu,
224 		struct msm_ringbuffer *ring)
225 {
226 	struct adreno_gpu *adreno_gpu = &a5xx_gpu->base;
227 	struct msm_gpu *gpu = &adreno_gpu->base;
228 	struct a5xx_preempt_record *ptr;
229 	void *counters;
230 	struct drm_gem_object *bo = NULL, *counters_bo = NULL;
231 	u64 iova = 0, counters_iova = 0;
232 
233 	ptr = msm_gem_kernel_new(gpu->dev,
234 		A5XX_PREEMPT_RECORD_SIZE + A5XX_PREEMPT_COUNTER_SIZE,
235 		MSM_BO_UNCACHED | MSM_BO_MAP_PRIV, gpu->aspace, &bo, &iova);
236 
237 	if (IS_ERR(ptr))
238 		return PTR_ERR(ptr);
239 
240 	/* The buffer to store counters needs to be unprivileged */
241 	counters = msm_gem_kernel_new(gpu->dev,
242 		A5XX_PREEMPT_COUNTER_SIZE,
243 		MSM_BO_UNCACHED, gpu->aspace, &counters_bo, &counters_iova);
244 	if (IS_ERR(counters)) {
245 		msm_gem_kernel_put(bo, gpu->aspace, true);
246 		return PTR_ERR(counters);
247 	}
248 
249 	msm_gem_object_set_name(bo, "preempt");
250 	msm_gem_object_set_name(counters_bo, "preempt_counters");
251 
252 	a5xx_gpu->preempt_bo[ring->id] = bo;
253 	a5xx_gpu->preempt_counters_bo[ring->id] = counters_bo;
254 	a5xx_gpu->preempt_iova[ring->id] = iova;
255 	a5xx_gpu->preempt[ring->id] = ptr;
256 
257 	/* Set up the defaults on the preemption record */
258 
259 	ptr->magic = A5XX_PREEMPT_RECORD_MAGIC;
260 	ptr->info = 0;
261 	ptr->data = 0;
262 	ptr->cntl = MSM_GPU_RB_CNTL_DEFAULT | AXXX_CP_RB_CNTL_NO_UPDATE;
263 
264 	ptr->rptr_addr = shadowptr(a5xx_gpu, ring);
265 	ptr->counter = counters_iova;
266 
267 	return 0;
268 }
269 
270 void a5xx_preempt_fini(struct msm_gpu *gpu)
271 {
272 	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
273 	struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
274 	int i;
275 
276 	for (i = 0; i < gpu->nr_rings; i++) {
277 		msm_gem_kernel_put(a5xx_gpu->preempt_bo[i], gpu->aspace, true);
278 		msm_gem_kernel_put(a5xx_gpu->preempt_counters_bo[i],
279 			gpu->aspace, true);
280 	}
281 }
282 
283 void a5xx_preempt_init(struct msm_gpu *gpu)
284 {
285 	struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu);
286 	struct a5xx_gpu *a5xx_gpu = to_a5xx_gpu(adreno_gpu);
287 	int i;
288 
289 	/* No preemption if we only have one ring */
290 	if (gpu->nr_rings <= 1)
291 		return;
292 
293 	for (i = 0; i < gpu->nr_rings; i++) {
294 		if (preempt_init_ring(a5xx_gpu, gpu->rb[i])) {
295 			/*
296 			 * On any failure our adventure is over. Clean up and
297 			 * set nr_rings to 1 to force preemption off
298 			 */
299 			a5xx_preempt_fini(gpu);
300 			gpu->nr_rings = 1;
301 
302 			return;
303 		}
304 	}
305 
306 	timer_setup(&a5xx_gpu->preempt_timer, a5xx_preempt_timer, 0);
307 }
308