xref: /openbmc/linux/drivers/gpu/drm/gma500/psb_irq.c (revision 06ba8020)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /**************************************************************************
3  * Copyright (c) 2007, Intel Corporation.
4  * All Rights Reserved.
5  *
6  * Intel funded Tungsten Graphics (http://www.tungstengraphics.com) to
7  * develop this driver.
8  *
9  **************************************************************************/
10 
11 #include <drm/drm_drv.h>
12 #include <drm/drm_vblank.h>
13 
14 #include "power.h"
15 #include "psb_drv.h"
16 #include "psb_intel_reg.h"
17 #include "psb_irq.h"
18 #include "psb_reg.h"
19 
20 /*
21  * inline functions
22  */
23 
24 static inline u32 gma_pipestat(int pipe)
25 {
26 	if (pipe == 0)
27 		return PIPEASTAT;
28 	if (pipe == 1)
29 		return PIPEBSTAT;
30 	if (pipe == 2)
31 		return PIPECSTAT;
32 	BUG();
33 }
34 
35 static inline u32 gma_pipeconf(int pipe)
36 {
37 	if (pipe == 0)
38 		return PIPEACONF;
39 	if (pipe == 1)
40 		return PIPEBCONF;
41 	if (pipe == 2)
42 		return PIPECCONF;
43 	BUG();
44 }
45 
46 void gma_enable_pipestat(struct drm_psb_private *dev_priv, int pipe, u32 mask)
47 {
48 	if ((dev_priv->pipestat[pipe] & mask) != mask) {
49 		u32 reg = gma_pipestat(pipe);
50 		dev_priv->pipestat[pipe] |= mask;
51 		/* Enable the interrupt, clear any pending status */
52 		if (gma_power_begin(&dev_priv->dev, false)) {
53 			u32 writeVal = PSB_RVDC32(reg);
54 			writeVal |= (mask | (mask >> 16));
55 			PSB_WVDC32(writeVal, reg);
56 			(void) PSB_RVDC32(reg);
57 			gma_power_end(&dev_priv->dev);
58 		}
59 	}
60 }
61 
62 void gma_disable_pipestat(struct drm_psb_private *dev_priv, int pipe, u32 mask)
63 {
64 	if ((dev_priv->pipestat[pipe] & mask) != 0) {
65 		u32 reg = gma_pipestat(pipe);
66 		dev_priv->pipestat[pipe] &= ~mask;
67 		if (gma_power_begin(&dev_priv->dev, false)) {
68 			u32 writeVal = PSB_RVDC32(reg);
69 			writeVal &= ~mask;
70 			PSB_WVDC32(writeVal, reg);
71 			(void) PSB_RVDC32(reg);
72 			gma_power_end(&dev_priv->dev);
73 		}
74 	}
75 }
76 
77 /*
78  * Display controller interrupt handler for pipe event.
79  */
80 static void gma_pipe_event_handler(struct drm_device *dev, int pipe)
81 {
82 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
83 
84 	uint32_t pipe_stat_val = 0;
85 	uint32_t pipe_stat_reg = gma_pipestat(pipe);
86 	uint32_t pipe_enable = dev_priv->pipestat[pipe];
87 	uint32_t pipe_status = dev_priv->pipestat[pipe] >> 16;
88 	uint32_t pipe_clear;
89 	uint32_t i = 0;
90 
91 	spin_lock(&dev_priv->irqmask_lock);
92 
93 	pipe_stat_val = PSB_RVDC32(pipe_stat_reg);
94 	pipe_stat_val &= pipe_enable | pipe_status;
95 	pipe_stat_val &= pipe_stat_val >> 16;
96 
97 	spin_unlock(&dev_priv->irqmask_lock);
98 
99 	/* Clear the 2nd level interrupt status bits
100 	 * Sometimes the bits are very sticky so we repeat until they unstick */
101 	for (i = 0; i < 0xffff; i++) {
102 		PSB_WVDC32(PSB_RVDC32(pipe_stat_reg), pipe_stat_reg);
103 		pipe_clear = PSB_RVDC32(pipe_stat_reg) & pipe_status;
104 
105 		if (pipe_clear == 0)
106 			break;
107 	}
108 
109 	if (pipe_clear)
110 		dev_err(dev->dev,
111 		"%s, can't clear status bits for pipe %d, its value = 0x%x.\n",
112 		__func__, pipe, PSB_RVDC32(pipe_stat_reg));
113 
114 	if (pipe_stat_val & PIPE_VBLANK_STATUS) {
115 		struct drm_crtc *crtc = drm_crtc_from_index(dev, pipe);
116 		struct gma_crtc *gma_crtc = to_gma_crtc(crtc);
117 		unsigned long flags;
118 
119 		drm_handle_vblank(dev, pipe);
120 
121 		spin_lock_irqsave(&dev->event_lock, flags);
122 		if (gma_crtc->page_flip_event) {
123 			drm_crtc_send_vblank_event(crtc,
124 						   gma_crtc->page_flip_event);
125 			gma_crtc->page_flip_event = NULL;
126 			drm_crtc_vblank_put(crtc);
127 		}
128 		spin_unlock_irqrestore(&dev->event_lock, flags);
129 	}
130 }
131 
132 /*
133  * Display controller interrupt handler.
134  */
135 static void gma_vdc_interrupt(struct drm_device *dev, uint32_t vdc_stat)
136 {
137 	if (vdc_stat & _PSB_IRQ_ASLE)
138 		psb_intel_opregion_asle_intr(dev);
139 
140 	if (vdc_stat & _PSB_VSYNC_PIPEA_FLAG)
141 		gma_pipe_event_handler(dev, 0);
142 
143 	if (vdc_stat & _PSB_VSYNC_PIPEB_FLAG)
144 		gma_pipe_event_handler(dev, 1);
145 }
146 
147 /*
148  * SGX interrupt handler
149  */
150 static void gma_sgx_interrupt(struct drm_device *dev, u32 stat_1, u32 stat_2)
151 {
152 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
153 	u32 val, addr;
154 
155 	if (stat_1 & _PSB_CE_TWOD_COMPLETE)
156 		val = PSB_RSGX32(PSB_CR_2D_BLIT_STATUS);
157 
158 	if (stat_2 & _PSB_CE2_BIF_REQUESTER_FAULT) {
159 		val = PSB_RSGX32(PSB_CR_BIF_INT_STAT);
160 		addr = PSB_RSGX32(PSB_CR_BIF_FAULT);
161 		if (val) {
162 			if (val & _PSB_CBI_STAT_PF_N_RW)
163 				DRM_ERROR("SGX MMU page fault:");
164 			else
165 				DRM_ERROR("SGX MMU read / write protection fault:");
166 
167 			if (val & _PSB_CBI_STAT_FAULT_CACHE)
168 				DRM_ERROR("\tCache requestor");
169 			if (val & _PSB_CBI_STAT_FAULT_TA)
170 				DRM_ERROR("\tTA requestor");
171 			if (val & _PSB_CBI_STAT_FAULT_VDM)
172 				DRM_ERROR("\tVDM requestor");
173 			if (val & _PSB_CBI_STAT_FAULT_2D)
174 				DRM_ERROR("\t2D requestor");
175 			if (val & _PSB_CBI_STAT_FAULT_PBE)
176 				DRM_ERROR("\tPBE requestor");
177 			if (val & _PSB_CBI_STAT_FAULT_TSP)
178 				DRM_ERROR("\tTSP requestor");
179 			if (val & _PSB_CBI_STAT_FAULT_ISP)
180 				DRM_ERROR("\tISP requestor");
181 			if (val & _PSB_CBI_STAT_FAULT_USSEPDS)
182 				DRM_ERROR("\tUSSEPDS requestor");
183 			if (val & _PSB_CBI_STAT_FAULT_HOST)
184 				DRM_ERROR("\tHost requestor");
185 
186 			DRM_ERROR("\tMMU failing address is 0x%08x.\n",
187 				  (unsigned int)addr);
188 		}
189 	}
190 
191 	/* Clear bits */
192 	PSB_WSGX32(stat_1, PSB_CR_EVENT_HOST_CLEAR);
193 	PSB_WSGX32(stat_2, PSB_CR_EVENT_HOST_CLEAR2);
194 	PSB_RSGX32(PSB_CR_EVENT_HOST_CLEAR2);
195 }
196 
197 static irqreturn_t gma_irq_handler(int irq, void *arg)
198 {
199 	struct drm_device *dev = arg;
200 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
201 	uint32_t vdc_stat, dsp_int = 0, sgx_int = 0, hotplug_int = 0;
202 	u32 sgx_stat_1, sgx_stat_2;
203 	int handled = 0;
204 
205 	spin_lock(&dev_priv->irqmask_lock);
206 
207 	vdc_stat = PSB_RVDC32(PSB_INT_IDENTITY_R);
208 
209 	if (vdc_stat & (_PSB_PIPE_EVENT_FLAG|_PSB_IRQ_ASLE))
210 		dsp_int = 1;
211 
212 	if (vdc_stat & _PSB_IRQ_SGX_FLAG)
213 		sgx_int = 1;
214 	if (vdc_stat & _PSB_IRQ_DISP_HOTSYNC)
215 		hotplug_int = 1;
216 
217 	vdc_stat &= dev_priv->vdc_irq_mask;
218 	spin_unlock(&dev_priv->irqmask_lock);
219 
220 	if (dsp_int) {
221 		gma_vdc_interrupt(dev, vdc_stat);
222 		handled = 1;
223 	}
224 
225 	if (sgx_int) {
226 		sgx_stat_1 = PSB_RSGX32(PSB_CR_EVENT_STATUS);
227 		sgx_stat_2 = PSB_RSGX32(PSB_CR_EVENT_STATUS2);
228 		gma_sgx_interrupt(dev, sgx_stat_1, sgx_stat_2);
229 		handled = 1;
230 	}
231 
232 	/* Note: this bit has other meanings on some devices, so we will
233 	   need to address that later if it ever matters */
234 	if (hotplug_int && dev_priv->ops->hotplug) {
235 		handled = dev_priv->ops->hotplug(dev);
236 		REG_WRITE(PORT_HOTPLUG_STAT, REG_READ(PORT_HOTPLUG_STAT));
237 	}
238 
239 	PSB_WVDC32(vdc_stat, PSB_INT_IDENTITY_R);
240 	(void) PSB_RVDC32(PSB_INT_IDENTITY_R);
241 	rmb();
242 
243 	if (!handled)
244 		return IRQ_NONE;
245 
246 	return IRQ_HANDLED;
247 }
248 
249 void gma_irq_preinstall(struct drm_device *dev)
250 {
251 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
252 	unsigned long irqflags;
253 
254 	spin_lock_irqsave(&dev_priv->irqmask_lock, irqflags);
255 
256 	PSB_WVDC32(0xFFFFFFFF, PSB_HWSTAM);
257 	PSB_WVDC32(0x00000000, PSB_INT_MASK_R);
258 	PSB_WVDC32(0x00000000, PSB_INT_ENABLE_R);
259 	PSB_WSGX32(0x00000000, PSB_CR_EVENT_HOST_ENABLE);
260 	PSB_RSGX32(PSB_CR_EVENT_HOST_ENABLE);
261 
262 	if (dev->vblank[0].enabled)
263 		dev_priv->vdc_irq_mask |= _PSB_VSYNC_PIPEA_FLAG;
264 	if (dev->vblank[1].enabled)
265 		dev_priv->vdc_irq_mask |= _PSB_VSYNC_PIPEB_FLAG;
266 
267 	/* Revisit this area - want per device masks ? */
268 	if (dev_priv->ops->hotplug)
269 		dev_priv->vdc_irq_mask |= _PSB_IRQ_DISP_HOTSYNC;
270 	dev_priv->vdc_irq_mask |= _PSB_IRQ_ASLE | _PSB_IRQ_SGX_FLAG;
271 
272 	/* This register is safe even if display island is off */
273 	PSB_WVDC32(~dev_priv->vdc_irq_mask, PSB_INT_MASK_R);
274 	spin_unlock_irqrestore(&dev_priv->irqmask_lock, irqflags);
275 }
276 
277 void gma_irq_postinstall(struct drm_device *dev)
278 {
279 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
280 	unsigned long irqflags;
281 	unsigned int i;
282 
283 	spin_lock_irqsave(&dev_priv->irqmask_lock, irqflags);
284 
285 	/* Enable 2D and MMU fault interrupts */
286 	PSB_WSGX32(_PSB_CE2_BIF_REQUESTER_FAULT, PSB_CR_EVENT_HOST_ENABLE2);
287 	PSB_WSGX32(_PSB_CE_TWOD_COMPLETE, PSB_CR_EVENT_HOST_ENABLE);
288 	PSB_RSGX32(PSB_CR_EVENT_HOST_ENABLE); /* Post */
289 
290 	/* This register is safe even if display island is off */
291 	PSB_WVDC32(dev_priv->vdc_irq_mask, PSB_INT_ENABLE_R);
292 	PSB_WVDC32(0xFFFFFFFF, PSB_HWSTAM);
293 
294 	for (i = 0; i < dev->num_crtcs; ++i) {
295 		if (dev->vblank[i].enabled)
296 			gma_enable_pipestat(dev_priv, i, PIPE_VBLANK_INTERRUPT_ENABLE);
297 		else
298 			gma_disable_pipestat(dev_priv, i, PIPE_VBLANK_INTERRUPT_ENABLE);
299 	}
300 
301 	if (dev_priv->ops->hotplug_enable)
302 		dev_priv->ops->hotplug_enable(dev, true);
303 
304 	spin_unlock_irqrestore(&dev_priv->irqmask_lock, irqflags);
305 }
306 
307 int gma_irq_install(struct drm_device *dev)
308 {
309 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
310 	struct pci_dev *pdev = to_pci_dev(dev->dev);
311 	int ret;
312 
313 	if (dev_priv->use_msi && pci_enable_msi(pdev)) {
314 		dev_warn(dev->dev, "Enabling MSI failed!\n");
315 		dev_priv->use_msi = false;
316 	}
317 
318 	if (pdev->irq == IRQ_NOTCONNECTED)
319 		return -ENOTCONN;
320 
321 	gma_irq_preinstall(dev);
322 
323 	/* PCI devices require shared interrupts. */
324 	ret = request_irq(pdev->irq, gma_irq_handler, IRQF_SHARED, dev->driver->name, dev);
325 	if (ret)
326 		return ret;
327 
328 	gma_irq_postinstall(dev);
329 
330 	return 0;
331 }
332 
333 void gma_irq_uninstall(struct drm_device *dev)
334 {
335 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
336 	struct pci_dev *pdev = to_pci_dev(dev->dev);
337 	unsigned long irqflags;
338 	unsigned int i;
339 
340 	spin_lock_irqsave(&dev_priv->irqmask_lock, irqflags);
341 
342 	if (dev_priv->ops->hotplug_enable)
343 		dev_priv->ops->hotplug_enable(dev, false);
344 
345 	PSB_WVDC32(0xFFFFFFFF, PSB_HWSTAM);
346 
347 	for (i = 0; i < dev->num_crtcs; ++i) {
348 		if (dev->vblank[i].enabled)
349 			gma_disable_pipestat(dev_priv, i, PIPE_VBLANK_INTERRUPT_ENABLE);
350 	}
351 
352 	dev_priv->vdc_irq_mask &= _PSB_IRQ_SGX_FLAG |
353 				  _PSB_IRQ_MSVDX_FLAG |
354 				  _LNC_IRQ_TOPAZ_FLAG;
355 
356 	/* These two registers are safe even if display island is off */
357 	PSB_WVDC32(~dev_priv->vdc_irq_mask, PSB_INT_MASK_R);
358 	PSB_WVDC32(dev_priv->vdc_irq_mask, PSB_INT_ENABLE_R);
359 
360 	wmb();
361 
362 	/* This register is safe even if display island is off */
363 	PSB_WVDC32(PSB_RVDC32(PSB_INT_IDENTITY_R), PSB_INT_IDENTITY_R);
364 	spin_unlock_irqrestore(&dev_priv->irqmask_lock, irqflags);
365 
366 	free_irq(pdev->irq, dev);
367 	if (dev_priv->use_msi)
368 		pci_disable_msi(pdev);
369 }
370 
371 int gma_crtc_enable_vblank(struct drm_crtc *crtc)
372 {
373 	struct drm_device *dev = crtc->dev;
374 	unsigned int pipe = crtc->index;
375 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
376 	unsigned long irqflags;
377 	uint32_t reg_val = 0;
378 	uint32_t pipeconf_reg = gma_pipeconf(pipe);
379 
380 	if (gma_power_begin(dev, false)) {
381 		reg_val = REG_READ(pipeconf_reg);
382 		gma_power_end(dev);
383 	}
384 
385 	if (!(reg_val & PIPEACONF_ENABLE))
386 		return -EINVAL;
387 
388 	spin_lock_irqsave(&dev_priv->irqmask_lock, irqflags);
389 
390 	if (pipe == 0)
391 		dev_priv->vdc_irq_mask |= _PSB_VSYNC_PIPEA_FLAG;
392 	else if (pipe == 1)
393 		dev_priv->vdc_irq_mask |= _PSB_VSYNC_PIPEB_FLAG;
394 
395 	PSB_WVDC32(~dev_priv->vdc_irq_mask, PSB_INT_MASK_R);
396 	PSB_WVDC32(dev_priv->vdc_irq_mask, PSB_INT_ENABLE_R);
397 	gma_enable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_ENABLE);
398 
399 	spin_unlock_irqrestore(&dev_priv->irqmask_lock, irqflags);
400 
401 	return 0;
402 }
403 
404 void gma_crtc_disable_vblank(struct drm_crtc *crtc)
405 {
406 	struct drm_device *dev = crtc->dev;
407 	unsigned int pipe = crtc->index;
408 	struct drm_psb_private *dev_priv = to_drm_psb_private(dev);
409 	unsigned long irqflags;
410 
411 	spin_lock_irqsave(&dev_priv->irqmask_lock, irqflags);
412 
413 	if (pipe == 0)
414 		dev_priv->vdc_irq_mask &= ~_PSB_VSYNC_PIPEA_FLAG;
415 	else if (pipe == 1)
416 		dev_priv->vdc_irq_mask &= ~_PSB_VSYNC_PIPEB_FLAG;
417 
418 	PSB_WVDC32(~dev_priv->vdc_irq_mask, PSB_INT_MASK_R);
419 	PSB_WVDC32(dev_priv->vdc_irq_mask, PSB_INT_ENABLE_R);
420 	gma_disable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_ENABLE);
421 
422 	spin_unlock_irqrestore(&dev_priv->irqmask_lock, irqflags);
423 }
424 
425 /* Called from drm generic code, passed a 'crtc', which
426  * we use as a pipe index
427  */
428 u32 gma_crtc_get_vblank_counter(struct drm_crtc *crtc)
429 {
430 	struct drm_device *dev = crtc->dev;
431 	unsigned int pipe = crtc->index;
432 	uint32_t high_frame = PIPEAFRAMEHIGH;
433 	uint32_t low_frame = PIPEAFRAMEPIXEL;
434 	uint32_t pipeconf_reg = PIPEACONF;
435 	uint32_t reg_val = 0;
436 	uint32_t high1 = 0, high2 = 0, low = 0, count = 0;
437 
438 	switch (pipe) {
439 	case 0:
440 		break;
441 	case 1:
442 		high_frame = PIPEBFRAMEHIGH;
443 		low_frame = PIPEBFRAMEPIXEL;
444 		pipeconf_reg = PIPEBCONF;
445 		break;
446 	case 2:
447 		high_frame = PIPECFRAMEHIGH;
448 		low_frame = PIPECFRAMEPIXEL;
449 		pipeconf_reg = PIPECCONF;
450 		break;
451 	default:
452 		dev_err(dev->dev, "%s, invalid pipe.\n", __func__);
453 		return 0;
454 	}
455 
456 	if (!gma_power_begin(dev, false))
457 		return 0;
458 
459 	reg_val = REG_READ(pipeconf_reg);
460 
461 	if (!(reg_val & PIPEACONF_ENABLE)) {
462 		dev_err(dev->dev, "trying to get vblank count for disabled pipe %u\n",
463 			pipe);
464 		goto err_gma_power_end;
465 	}
466 
467 	/*
468 	 * High & low register fields aren't synchronized, so make sure
469 	 * we get a low value that's stable across two reads of the high
470 	 * register.
471 	 */
472 	do {
473 		high1 = ((REG_READ(high_frame) & PIPE_FRAME_HIGH_MASK) >>
474 			 PIPE_FRAME_HIGH_SHIFT);
475 		low =  ((REG_READ(low_frame) & PIPE_FRAME_LOW_MASK) >>
476 			PIPE_FRAME_LOW_SHIFT);
477 		high2 = ((REG_READ(high_frame) & PIPE_FRAME_HIGH_MASK) >>
478 			 PIPE_FRAME_HIGH_SHIFT);
479 	} while (high1 != high2);
480 
481 	count = (high1 << 8) | low;
482 
483 err_gma_power_end:
484 	gma_power_end(dev);
485 
486 	return count;
487 }
488 
489