1 /**************************************************************************
2  * Copyright (c) 2007-2011, Intel Corporation.
3  * All Rights Reserved.
4  *
5  * This program is free software; you can redistribute it and/or modify it
6  * under the terms and conditions of the GNU General Public License,
7  * version 2, as published by the Free Software Foundation.
8  *
9  * This program is distributed in the hope it will be useful, but WITHOUT
10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
12  * more details.
13  *
14  * You should have received a copy of the GNU General Public License along with
15  * this program; if not, write to the Free Software Foundation, Inc.,
16  * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
17  *
18  **************************************************************************/
19 
20 #include <linux/module.h>
21 #include <linux/kernel.h>
22 #include <linux/errno.h>
23 #include <linux/string.h>
24 #include <linux/mm.h>
25 #include <linux/tty.h>
26 #include <linux/slab.h>
27 #include <linux/delay.h>
28 #include <linux/fb.h>
29 #include <linux/init.h>
30 #include <linux/console.h>
31 
32 #include <drm/drmP.h>
33 #include <drm/drm.h>
34 #include <drm/drm_crtc.h>
35 #include <drm/drm_fb_helper.h>
36 
37 #include "psb_drv.h"
38 #include "psb_intel_reg.h"
39 #include "psb_intel_drv.h"
40 #include "framebuffer.h"
41 #include "gtt.h"
42 
43 static void psb_user_framebuffer_destroy(struct drm_framebuffer *fb);
44 static int psb_user_framebuffer_create_handle(struct drm_framebuffer *fb,
45 					      struct drm_file *file_priv,
46 					      unsigned int *handle);
47 
48 static const struct drm_framebuffer_funcs psb_fb_funcs = {
49 	.destroy = psb_user_framebuffer_destroy,
50 	.create_handle = psb_user_framebuffer_create_handle,
51 };
52 
53 #define CMAP_TOHW(_val, _width) ((((_val) << (_width)) + 0x7FFF - (_val)) >> 16)
54 
55 static int psbfb_setcolreg(unsigned regno, unsigned red, unsigned green,
56 			   unsigned blue, unsigned transp,
57 			   struct fb_info *info)
58 {
59 	struct psb_fbdev *fbdev = info->par;
60 	struct drm_framebuffer *fb = fbdev->psb_fb_helper.fb;
61 	uint32_t v;
62 
63 	if (!fb)
64 		return -ENOMEM;
65 
66 	if (regno > 255)
67 		return 1;
68 
69 	red = CMAP_TOHW(red, info->var.red.length);
70 	blue = CMAP_TOHW(blue, info->var.blue.length);
71 	green = CMAP_TOHW(green, info->var.green.length);
72 	transp = CMAP_TOHW(transp, info->var.transp.length);
73 
74 	v = (red << info->var.red.offset) |
75 	    (green << info->var.green.offset) |
76 	    (blue << info->var.blue.offset) |
77 	    (transp << info->var.transp.offset);
78 
79 	if (regno < 16) {
80 		switch (fb->bits_per_pixel) {
81 		case 16:
82 			((uint32_t *) info->pseudo_palette)[regno] = v;
83 			break;
84 		case 24:
85 		case 32:
86 			((uint32_t *) info->pseudo_palette)[regno] = v;
87 			break;
88 		}
89 	}
90 
91 	return 0;
92 }
93 
94 static int psbfb_pan(struct fb_var_screeninfo *var, struct fb_info *info)
95 {
96 	struct psb_fbdev *fbdev = info->par;
97 	struct psb_framebuffer *psbfb = &fbdev->pfb;
98 	struct drm_device *dev = psbfb->base.dev;
99 
100 	/*
101 	 *	We have to poke our nose in here. The core fb code assumes
102 	 *	panning is part of the hardware that can be invoked before
103 	 *	the actual fb is mapped. In our case that isn't quite true.
104 	 */
105 	if (psbfb->gtt->npage) {
106 		/* GTT roll shifts in 4K pages, we need to shift the right
107 		   number of pages */
108 		int pages = info->fix.line_length >> 12;
109 		psb_gtt_roll(dev, psbfb->gtt, var->yoffset * pages);
110 	}
111         return 0;
112 }
113 
114 static int psbfb_vm_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
115 {
116 	struct psb_framebuffer *psbfb = vma->vm_private_data;
117 	struct drm_device *dev = psbfb->base.dev;
118 	struct drm_psb_private *dev_priv = dev->dev_private;
119 	int page_num;
120 	int i;
121 	unsigned long address;
122 	int ret;
123 	unsigned long pfn;
124 	/* FIXME: assumes fb at stolen base which may not be true */
125 	unsigned long phys_addr = (unsigned long)dev_priv->stolen_base;
126 
127 	page_num = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
128 	address = (unsigned long)vmf->virtual_address - (vmf->pgoff << PAGE_SHIFT);
129 
130 	vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
131 
132 	for (i = 0; i < page_num; i++) {
133 		pfn = (phys_addr >> PAGE_SHIFT);
134 
135 		ret = vm_insert_mixed(vma, address, pfn);
136 		if (unlikely((ret == -EBUSY) || (ret != 0 && i > 0)))
137 			break;
138 		else if (unlikely(ret != 0)) {
139 			ret = (ret == -ENOMEM) ? VM_FAULT_OOM : VM_FAULT_SIGBUS;
140 			return ret;
141 		}
142 		address += PAGE_SIZE;
143 		phys_addr += PAGE_SIZE;
144 	}
145 	return VM_FAULT_NOPAGE;
146 }
147 
148 static void psbfb_vm_open(struct vm_area_struct *vma)
149 {
150 }
151 
152 static void psbfb_vm_close(struct vm_area_struct *vma)
153 {
154 }
155 
156 static struct vm_operations_struct psbfb_vm_ops = {
157 	.fault	= psbfb_vm_fault,
158 	.open	= psbfb_vm_open,
159 	.close	= psbfb_vm_close
160 };
161 
162 static int psbfb_mmap(struct fb_info *info, struct vm_area_struct *vma)
163 {
164 	struct psb_fbdev *fbdev = info->par;
165 	struct psb_framebuffer *psbfb = &fbdev->pfb;
166 
167 	if (vma->vm_pgoff != 0)
168 		return -EINVAL;
169 	if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT))
170 		return -EINVAL;
171 
172 	if (!psbfb->addr_space)
173 		psbfb->addr_space = vma->vm_file->f_mapping;
174 	/*
175 	 * If this is a GEM object then info->screen_base is the virtual
176 	 * kernel remapping of the object. FIXME: Review if this is
177 	 * suitable for our mmap work
178 	 */
179 	vma->vm_ops = &psbfb_vm_ops;
180 	vma->vm_private_data = (void *)psbfb;
181 	vma->vm_flags |= VM_RESERVED | VM_IO |
182 					VM_MIXEDMAP | VM_DONTEXPAND;
183 	return 0;
184 }
185 
186 static int psbfb_ioctl(struct fb_info *info, unsigned int cmd,
187 						unsigned long arg)
188 {
189 	return -ENOTTY;
190 }
191 
192 static struct fb_ops psbfb_ops = {
193 	.owner = THIS_MODULE,
194 	.fb_check_var = drm_fb_helper_check_var,
195 	.fb_set_par = drm_fb_helper_set_par,
196 	.fb_blank = drm_fb_helper_blank,
197 	.fb_setcolreg = psbfb_setcolreg,
198 	.fb_fillrect = cfb_fillrect,
199 	.fb_copyarea = psbfb_copyarea,
200 	.fb_imageblit = cfb_imageblit,
201 	.fb_mmap = psbfb_mmap,
202 	.fb_sync = psbfb_sync,
203 	.fb_ioctl = psbfb_ioctl,
204 };
205 
206 static struct fb_ops psbfb_roll_ops = {
207 	.owner = THIS_MODULE,
208 	.fb_check_var = drm_fb_helper_check_var,
209 	.fb_set_par = drm_fb_helper_set_par,
210 	.fb_blank = drm_fb_helper_blank,
211 	.fb_setcolreg = psbfb_setcolreg,
212 	.fb_fillrect = cfb_fillrect,
213 	.fb_copyarea = cfb_copyarea,
214 	.fb_imageblit = cfb_imageblit,
215 	.fb_pan_display = psbfb_pan,
216 	.fb_mmap = psbfb_mmap,
217 	.fb_ioctl = psbfb_ioctl,
218 };
219 
220 static struct fb_ops psbfb_unaccel_ops = {
221 	.owner = THIS_MODULE,
222 	.fb_check_var = drm_fb_helper_check_var,
223 	.fb_set_par = drm_fb_helper_set_par,
224 	.fb_blank = drm_fb_helper_blank,
225 	.fb_setcolreg = psbfb_setcolreg,
226 	.fb_fillrect = cfb_fillrect,
227 	.fb_copyarea = cfb_copyarea,
228 	.fb_imageblit = cfb_imageblit,
229 	.fb_mmap = psbfb_mmap,
230 	.fb_ioctl = psbfb_ioctl,
231 };
232 
233 /**
234  *	psb_framebuffer_init	-	initialize a framebuffer
235  *	@dev: our DRM device
236  *	@fb: framebuffer to set up
237  *	@mode_cmd: mode description
238  *	@gt: backing object
239  *
240  *	Configure and fill in the boilerplate for our frame buffer. Return
241  *	0 on success or an error code if we fail.
242  */
243 static int psb_framebuffer_init(struct drm_device *dev,
244 					struct psb_framebuffer *fb,
245 					struct drm_mode_fb_cmd2 *mode_cmd,
246 					struct gtt_range *gt)
247 {
248 	u32 bpp, depth;
249 	int ret;
250 
251 	drm_fb_get_bpp_depth(mode_cmd->pixel_format, &depth, &bpp);
252 
253 	if (mode_cmd->pitches[0] & 63)
254 		return -EINVAL;
255 	switch (bpp) {
256 	case 8:
257 	case 16:
258 	case 24:
259 	case 32:
260 		break;
261 	default:
262 		return -EINVAL;
263 	}
264 	ret = drm_framebuffer_init(dev, &fb->base, &psb_fb_funcs);
265 	if (ret) {
266 		dev_err(dev->dev, "framebuffer init failed: %d\n", ret);
267 		return ret;
268 	}
269 	drm_helper_mode_fill_fb_struct(&fb->base, mode_cmd);
270 	fb->gtt = gt;
271 	return 0;
272 }
273 
274 /**
275  *	psb_framebuffer_create	-	create a framebuffer backed by gt
276  *	@dev: our DRM device
277  *	@mode_cmd: the description of the requested mode
278  *	@gt: the backing object
279  *
280  *	Create a framebuffer object backed by the gt, and fill in the
281  *	boilerplate required
282  *
283  *	TODO: review object references
284  */
285 
286 static struct drm_framebuffer *psb_framebuffer_create
287 			(struct drm_device *dev,
288 			 struct drm_mode_fb_cmd2 *mode_cmd,
289 			 struct gtt_range *gt)
290 {
291 	struct psb_framebuffer *fb;
292 	int ret;
293 
294 	fb = kzalloc(sizeof(*fb), GFP_KERNEL);
295 	if (!fb)
296 		return ERR_PTR(-ENOMEM);
297 
298 	ret = psb_framebuffer_init(dev, fb, mode_cmd, gt);
299 	if (ret) {
300 		kfree(fb);
301 		return ERR_PTR(ret);
302 	}
303 	return &fb->base;
304 }
305 
306 /**
307  *	psbfb_alloc		-	allocate frame buffer memory
308  *	@dev: the DRM device
309  *	@aligned_size: space needed
310  *	@force: fall back to GEM buffers if need be
311  *
312  *	Allocate the frame buffer. In the usual case we get a GTT range that
313  *	is stolen memory backed and life is simple. If there isn't sufficient
314  *	we fail as we don't have the virtual mapping space to really vmap it
315  *	and the kernel console code can't handle non linear framebuffers.
316  *
317  *	Re-address this as and if the framebuffer layer grows this ability.
318  */
319 static struct gtt_range *psbfb_alloc(struct drm_device *dev, int aligned_size)
320 {
321 	struct gtt_range *backing;
322 	/* Begin by trying to use stolen memory backing */
323 	backing = psb_gtt_alloc_range(dev, aligned_size, "fb", 1);
324 	if (backing) {
325 		if (drm_gem_private_object_init(dev,
326 					&backing->gem, aligned_size) == 0)
327 			return backing;
328 		psb_gtt_free_range(dev, backing);
329 	}
330 	return NULL;
331 }
332 
333 /**
334  *	psbfb_create		-	create a framebuffer
335  *	@fbdev: the framebuffer device
336  *	@sizes: specification of the layout
337  *
338  *	Create a framebuffer to the specifications provided
339  */
340 static int psbfb_create(struct psb_fbdev *fbdev,
341 				struct drm_fb_helper_surface_size *sizes)
342 {
343 	struct drm_device *dev = fbdev->psb_fb_helper.dev;
344 	struct drm_psb_private *dev_priv = dev->dev_private;
345 	struct fb_info *info;
346 	struct drm_framebuffer *fb;
347 	struct psb_framebuffer *psbfb = &fbdev->pfb;
348 	struct drm_mode_fb_cmd2 mode_cmd;
349 	struct device *device = &dev->pdev->dev;
350 	int size;
351 	int ret;
352 	struct gtt_range *backing;
353 	u32 bpp, depth;
354 	int gtt_roll = 0;
355 	int pitch_lines = 0;
356 
357 	mode_cmd.width = sizes->surface_width;
358 	mode_cmd.height = sizes->surface_height;
359 	bpp = sizes->surface_bpp;
360 	depth = sizes->surface_depth;
361 
362 	/* No 24bit packed */
363 	if (bpp == 24)
364 		bpp = 32;
365 
366 	do {
367 		/*
368 		 * Acceleration via the GTT requires pitch to be
369 		 * power of two aligned. Preferably page but less
370 		 * is ok with some fonts
371 		 */
372         	mode_cmd.pitches[0] =  ALIGN(mode_cmd.width * ((bpp + 7) / 8), 4096 >> pitch_lines);
373 
374         	size = mode_cmd.pitches[0] * mode_cmd.height;
375         	size = ALIGN(size, PAGE_SIZE);
376 
377 		/* Allocate the fb in the GTT with stolen page backing */
378 		backing = psbfb_alloc(dev, size);
379 
380 		if (pitch_lines)
381 			pitch_lines *= 2;
382 		else
383 			pitch_lines = 1;
384 		gtt_roll++;
385 	} while (backing == NULL && pitch_lines <= 16);
386 
387 	/* The final pitch we accepted if we succeeded */
388 	pitch_lines /= 2;
389 
390 	if (backing == NULL) {
391 		/*
392 		 *	We couldn't get the space we wanted, fall back to the
393 		 *	display engine requirement instead.  The HW requires
394 		 *	the pitch to be 64 byte aligned
395 		 */
396 
397 		gtt_roll = 0;	/* Don't use GTT accelerated scrolling */
398 		pitch_lines = 64;
399 
400 		mode_cmd.pitches[0] =  ALIGN(mode_cmd.width * ((bpp + 7) / 8), 64);
401 
402 		size = mode_cmd.pitches[0] * mode_cmd.height;
403 		size = ALIGN(size, PAGE_SIZE);
404 
405 		/* Allocate the framebuffer in the GTT with stolen page backing */
406 		backing = psbfb_alloc(dev, size);
407 		if (backing == NULL)
408 			return -ENOMEM;
409 	}
410 
411 	mutex_lock(&dev->struct_mutex);
412 
413 	info = framebuffer_alloc(0, device);
414 	if (!info) {
415 		ret = -ENOMEM;
416 		goto out_err1;
417 	}
418 	info->par = fbdev;
419 
420 	mode_cmd.pixel_format = drm_mode_legacy_fb_format(bpp, depth);
421 
422 	ret = psb_framebuffer_init(dev, psbfb, &mode_cmd, backing);
423 	if (ret)
424 		goto out_unref;
425 
426 	fb = &psbfb->base;
427 	psbfb->fbdev = info;
428 
429 	fbdev->psb_fb_helper.fb = fb;
430 	fbdev->psb_fb_helper.fbdev = info;
431 
432 	drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
433 	strcpy(info->fix.id, "psbfb");
434 
435 	info->flags = FBINFO_DEFAULT;
436 	if (dev_priv->ops->accel_2d && pitch_lines > 8)	/* 2D engine */
437 		info->fbops = &psbfb_ops;
438 	else if (gtt_roll) {	/* GTT rolling seems best */
439 		info->fbops = &psbfb_roll_ops;
440 		info->flags |= FBINFO_HWACCEL_YPAN;
441 	} else	/* Software */
442 		info->fbops = &psbfb_unaccel_ops;
443 
444 	ret = fb_alloc_cmap(&info->cmap, 256, 0);
445 	if (ret) {
446 		ret = -ENOMEM;
447 		goto out_unref;
448 	}
449 
450 	info->fix.smem_start = dev->mode_config.fb_base;
451 	info->fix.smem_len = size;
452 	info->fix.ywrapstep = gtt_roll;
453 	info->fix.ypanstep = 0;
454 
455 	/* Accessed stolen memory directly */
456 	info->screen_base = (char *)dev_priv->vram_addr +
457 							backing->offset;
458 	info->screen_size = size;
459 
460 	if (dev_priv->gtt.stolen_size) {
461 		info->apertures = alloc_apertures(1);
462 		if (!info->apertures) {
463 			ret = -ENOMEM;
464 			goto out_unref;
465 		}
466 		info->apertures->ranges[0].base = dev->mode_config.fb_base;
467 		info->apertures->ranges[0].size = dev_priv->gtt.stolen_size;
468 	}
469 
470 	drm_fb_helper_fill_var(info, &fbdev->psb_fb_helper,
471 				sizes->fb_width, sizes->fb_height);
472 
473 	info->fix.mmio_start = pci_resource_start(dev->pdev, 0);
474 	info->fix.mmio_len = pci_resource_len(dev->pdev, 0);
475 
476 	/* Use default scratch pixmap (info->pixmap.flags = FB_PIXMAP_SYSTEM) */
477 
478 	dev_info(dev->dev, "allocated %dx%d fb\n",
479 					psbfb->base.width, psbfb->base.height);
480 
481 	mutex_unlock(&dev->struct_mutex);
482 	return 0;
483 out_unref:
484 	if (backing->stolen)
485 		psb_gtt_free_range(dev, backing);
486 	else
487 		drm_gem_object_unreference(&backing->gem);
488 out_err1:
489 	mutex_unlock(&dev->struct_mutex);
490 	psb_gtt_free_range(dev, backing);
491 	return ret;
492 }
493 
494 /**
495  *	psb_user_framebuffer_create	-	create framebuffer
496  *	@dev: our DRM device
497  *	@filp: client file
498  *	@cmd: mode request
499  *
500  *	Create a new framebuffer backed by a userspace GEM object
501  */
502 static struct drm_framebuffer *psb_user_framebuffer_create
503 			(struct drm_device *dev, struct drm_file *filp,
504 			 struct drm_mode_fb_cmd2 *cmd)
505 {
506 	struct gtt_range *r;
507 	struct drm_gem_object *obj;
508 
509 	/*
510 	 *	Find the GEM object and thus the gtt range object that is
511 	 *	to back this space
512 	 */
513 	obj = drm_gem_object_lookup(dev, filp, cmd->handles[0]);
514 	if (obj == NULL)
515 		return ERR_PTR(-ENOENT);
516 
517 	/* Let the core code do all the work */
518 	r = container_of(obj, struct gtt_range, gem);
519 	return psb_framebuffer_create(dev, cmd, r);
520 }
521 
522 static void psbfb_gamma_set(struct drm_crtc *crtc, u16 red, u16 green,
523 							u16 blue, int regno)
524 {
525 	struct psb_intel_crtc *intel_crtc = to_psb_intel_crtc(crtc);
526 
527 	intel_crtc->lut_r[regno] = red >> 8;
528 	intel_crtc->lut_g[regno] = green >> 8;
529 	intel_crtc->lut_b[regno] = blue >> 8;
530 }
531 
532 static void psbfb_gamma_get(struct drm_crtc *crtc, u16 *red,
533 					u16 *green, u16 *blue, int regno)
534 {
535 	struct psb_intel_crtc *intel_crtc = to_psb_intel_crtc(crtc);
536 
537 	*red = intel_crtc->lut_r[regno] << 8;
538 	*green = intel_crtc->lut_g[regno] << 8;
539 	*blue = intel_crtc->lut_b[regno] << 8;
540 }
541 
542 static int psbfb_probe(struct drm_fb_helper *helper,
543 				struct drm_fb_helper_surface_size *sizes)
544 {
545 	struct psb_fbdev *psb_fbdev = (struct psb_fbdev *)helper;
546 	int new_fb = 0;
547 	int ret;
548 
549 	if (!helper->fb) {
550 		ret = psbfb_create(psb_fbdev, sizes);
551 		if (ret)
552 			return ret;
553 		new_fb = 1;
554 	}
555 	return new_fb;
556 }
557 
558 struct drm_fb_helper_funcs psb_fb_helper_funcs = {
559 	.gamma_set = psbfb_gamma_set,
560 	.gamma_get = psbfb_gamma_get,
561 	.fb_probe = psbfb_probe,
562 };
563 
564 static int psb_fbdev_destroy(struct drm_device *dev, struct psb_fbdev *fbdev)
565 {
566 	struct fb_info *info;
567 	struct psb_framebuffer *psbfb = &fbdev->pfb;
568 
569 	if (fbdev->psb_fb_helper.fbdev) {
570 		info = fbdev->psb_fb_helper.fbdev;
571 		unregister_framebuffer(info);
572 		if (info->cmap.len)
573 			fb_dealloc_cmap(&info->cmap);
574 		framebuffer_release(info);
575 	}
576 	drm_fb_helper_fini(&fbdev->psb_fb_helper);
577 	drm_framebuffer_cleanup(&psbfb->base);
578 
579 	if (psbfb->gtt)
580 		drm_gem_object_unreference(&psbfb->gtt->gem);
581 	return 0;
582 }
583 
584 int psb_fbdev_init(struct drm_device *dev)
585 {
586 	struct psb_fbdev *fbdev;
587 	struct drm_psb_private *dev_priv = dev->dev_private;
588 
589 	fbdev = kzalloc(sizeof(struct psb_fbdev), GFP_KERNEL);
590 	if (!fbdev) {
591 		dev_err(dev->dev, "no memory\n");
592 		return -ENOMEM;
593 	}
594 
595 	dev_priv->fbdev = fbdev;
596 	fbdev->psb_fb_helper.funcs = &psb_fb_helper_funcs;
597 
598 	drm_fb_helper_init(dev, &fbdev->psb_fb_helper, dev_priv->ops->crtcs,
599 							INTELFB_CONN_LIMIT);
600 
601 	drm_fb_helper_single_add_all_connectors(&fbdev->psb_fb_helper);
602 	drm_fb_helper_initial_config(&fbdev->psb_fb_helper, 32);
603 	return 0;
604 }
605 
606 static void psb_fbdev_fini(struct drm_device *dev)
607 {
608 	struct drm_psb_private *dev_priv = dev->dev_private;
609 
610 	if (!dev_priv->fbdev)
611 		return;
612 
613 	psb_fbdev_destroy(dev, dev_priv->fbdev);
614 	kfree(dev_priv->fbdev);
615 	dev_priv->fbdev = NULL;
616 }
617 
618 static void psbfb_output_poll_changed(struct drm_device *dev)
619 {
620 	struct drm_psb_private *dev_priv = dev->dev_private;
621 	struct psb_fbdev *fbdev = (struct psb_fbdev *)dev_priv->fbdev;
622 	drm_fb_helper_hotplug_event(&fbdev->psb_fb_helper);
623 }
624 
625 /**
626  *	psb_user_framebuffer_create_handle - add hamdle to a framebuffer
627  *	@fb: framebuffer
628  *	@file_priv: our DRM file
629  *	@handle: returned handle
630  *
631  *	Our framebuffer object is a GTT range which also contains a GEM
632  *	object. We need to turn it into a handle for userspace. GEM will do
633  *	the work for us
634  */
635 static int psb_user_framebuffer_create_handle(struct drm_framebuffer *fb,
636 					      struct drm_file *file_priv,
637 					      unsigned int *handle)
638 {
639 	struct psb_framebuffer *psbfb = to_psb_fb(fb);
640 	struct gtt_range *r = psbfb->gtt;
641 	return drm_gem_handle_create(file_priv, &r->gem, handle);
642 }
643 
644 /**
645  *	psb_user_framebuffer_destroy	-	destruct user created fb
646  *	@fb: framebuffer
647  *
648  *	User framebuffers are backed by GEM objects so all we have to do is
649  *	clean up a bit and drop the reference, GEM will handle the fallout
650  */
651 static void psb_user_framebuffer_destroy(struct drm_framebuffer *fb)
652 {
653 	struct psb_framebuffer *psbfb = to_psb_fb(fb);
654 	struct gtt_range *r = psbfb->gtt;
655 	struct drm_device *dev = fb->dev;
656 	struct drm_psb_private *dev_priv = dev->dev_private;
657 	struct psb_fbdev *fbdev = dev_priv->fbdev;
658 	struct drm_crtc *crtc;
659 	int reset = 0;
660 
661 	/* Should never get stolen memory for a user fb */
662 	WARN_ON(r->stolen);
663 
664 	/* Check if we are erroneously live */
665 	list_for_each_entry(crtc, &dev->mode_config.crtc_list, head)
666 		if (crtc->fb == fb)
667 			reset = 1;
668 
669 	if (reset)
670 		/*
671 		 * Now force a sane response before we permit the DRM CRTC
672 		 * layer to do stupid things like blank the display. Instead
673 		 * we reset this framebuffer as if the user had forced a reset.
674 		 * We must do this before the cleanup so that the DRM layer
675 		 * doesn't get a chance to stick its oar in where it isn't
676 		 * wanted.
677 		 */
678 		drm_fb_helper_restore_fbdev_mode(&fbdev->psb_fb_helper);
679 
680 	/* Let DRM do its clean up */
681 	drm_framebuffer_cleanup(fb);
682 	/*  We are no longer using the resource in GEM */
683 	drm_gem_object_unreference_unlocked(&r->gem);
684 	kfree(fb);
685 }
686 
687 static const struct drm_mode_config_funcs psb_mode_funcs = {
688 	.fb_create = psb_user_framebuffer_create,
689 	.output_poll_changed = psbfb_output_poll_changed,
690 };
691 
692 static int psb_create_backlight_property(struct drm_device *dev)
693 {
694 	struct drm_psb_private *dev_priv = dev->dev_private;
695 	struct drm_property *backlight;
696 
697 	if (dev_priv->backlight_property)
698 		return 0;
699 
700 	backlight = drm_property_create_range(dev, 0, "backlight", 0, 100);
701 
702 	dev_priv->backlight_property = backlight;
703 
704 	return 0;
705 }
706 
707 static void psb_setup_outputs(struct drm_device *dev)
708 {
709 	struct drm_psb_private *dev_priv = dev->dev_private;
710 	struct drm_connector *connector;
711 
712 	drm_mode_create_scaling_mode_property(dev);
713 	psb_create_backlight_property(dev);
714 
715 	dev_priv->ops->output_init(dev);
716 
717 	list_for_each_entry(connector, &dev->mode_config.connector_list,
718 			    head) {
719 		struct psb_intel_encoder *psb_intel_encoder =
720 			psb_intel_attached_encoder(connector);
721 		struct drm_encoder *encoder = &psb_intel_encoder->base;
722 		int crtc_mask = 0, clone_mask = 0;
723 
724 		/* valid crtcs */
725 		switch (psb_intel_encoder->type) {
726 		case INTEL_OUTPUT_ANALOG:
727 			crtc_mask = (1 << 0);
728 			clone_mask = (1 << INTEL_OUTPUT_ANALOG);
729 			break;
730 		case INTEL_OUTPUT_SDVO:
731 			crtc_mask = ((1 << 0) | (1 << 1));
732 			clone_mask = (1 << INTEL_OUTPUT_SDVO);
733 			break;
734 		case INTEL_OUTPUT_LVDS:
735 			if (IS_MRST(dev))
736 				crtc_mask = (1 << 0);
737 			else
738 				crtc_mask = (1 << 1);
739 			clone_mask = (1 << INTEL_OUTPUT_LVDS);
740 			break;
741 		case INTEL_OUTPUT_MIPI:
742 			crtc_mask = (1 << 0);
743 			clone_mask = (1 << INTEL_OUTPUT_MIPI);
744 			break;
745 		case INTEL_OUTPUT_MIPI2:
746 			crtc_mask = (1 << 2);
747 			clone_mask = (1 << INTEL_OUTPUT_MIPI2);
748 			break;
749 		case INTEL_OUTPUT_HDMI:
750 			if (IS_MFLD(dev))
751 				crtc_mask = (1 << 1);
752 			else
753 				crtc_mask = (1 << 0);
754 			clone_mask = (1 << INTEL_OUTPUT_HDMI);
755 			break;
756 		}
757 		encoder->possible_crtcs = crtc_mask;
758 		encoder->possible_clones =
759 		    psb_intel_connector_clones(dev, clone_mask);
760 	}
761 }
762 
763 void psb_modeset_init(struct drm_device *dev)
764 {
765 	struct drm_psb_private *dev_priv = dev->dev_private;
766 	struct psb_intel_mode_device *mode_dev = &dev_priv->mode_dev;
767 	int i;
768 
769 	drm_mode_config_init(dev);
770 
771 	dev->mode_config.min_width = 0;
772 	dev->mode_config.min_height = 0;
773 
774 	dev->mode_config.funcs = (void *) &psb_mode_funcs;
775 
776 	/* set memory base */
777 	/* Oaktrail and Poulsbo should use BAR 2*/
778 	pci_read_config_dword(dev->pdev, PSB_BSM, (u32 *)
779 					&(dev->mode_config.fb_base));
780 
781 	/* num pipes is 2 for PSB but 1 for Mrst */
782 	for (i = 0; i < dev_priv->num_pipe; i++)
783 		psb_intel_crtc_init(dev, i, mode_dev);
784 
785 	dev->mode_config.max_width = 2048;
786 	dev->mode_config.max_height = 2048;
787 
788 	psb_setup_outputs(dev);
789 }
790 
791 void psb_modeset_cleanup(struct drm_device *dev)
792 {
793 	mutex_lock(&dev->struct_mutex);
794 
795 	drm_kms_helper_poll_fini(dev);
796 	psb_fbdev_fini(dev);
797 	drm_mode_config_cleanup(dev);
798 
799 	mutex_unlock(&dev->struct_mutex);
800 }
801