1 /**************************************************************************
2  *
3  * Copyright © 2009-2015 VMware, Inc., Palo Alto, CA., USA
4  * All Rights Reserved.
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a
7  * copy of this software and associated documentation files (the
8  * "Software"), to deal in the Software without restriction, including
9  * without limitation the rights to use, copy, modify, merge, publish,
10  * distribute, sub license, and/or sell copies of the Software, and to
11  * permit persons to whom the Software is furnished to do so, subject to
12  * the following conditions:
13  *
14  * The above copyright notice and this permission notice (including the
15  * next paragraph) shall be included in all copies or substantial portions
16  * of the Software.
17  *
18  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
21  * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
22  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
23  * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
24  * USE OR OTHER DEALINGS IN THE SOFTWARE.
25  *
26  **************************************************************************/
27 
28 #include "vmwgfx_drv.h"
29 #include <drm/vmwgfx_drm.h>
30 #include "vmwgfx_kms.h"
31 #include "device_include/svga3d_caps.h"
32 
33 struct svga_3d_compat_cap {
34 	SVGA3dCapsRecordHeader header;
35 	SVGA3dCapPair pairs[SVGA3D_DEVCAP_MAX];
36 };
37 
38 int vmw_getparam_ioctl(struct drm_device *dev, void *data,
39 		       struct drm_file *file_priv)
40 {
41 	struct vmw_private *dev_priv = vmw_priv(dev);
42 	struct drm_vmw_getparam_arg *param =
43 	    (struct drm_vmw_getparam_arg *)data;
44 	struct vmw_fpriv *vmw_fp = vmw_fpriv(file_priv);
45 
46 	switch (param->param) {
47 	case DRM_VMW_PARAM_NUM_STREAMS:
48 		param->value = vmw_overlay_num_overlays(dev_priv);
49 		break;
50 	case DRM_VMW_PARAM_NUM_FREE_STREAMS:
51 		param->value = vmw_overlay_num_free_overlays(dev_priv);
52 		break;
53 	case DRM_VMW_PARAM_3D:
54 		param->value = vmw_fifo_have_3d(dev_priv) ? 1 : 0;
55 		break;
56 	case DRM_VMW_PARAM_HW_CAPS:
57 		param->value = dev_priv->capabilities;
58 		break;
59 	case DRM_VMW_PARAM_FIFO_CAPS:
60 		param->value = dev_priv->fifo.capabilities;
61 		break;
62 	case DRM_VMW_PARAM_MAX_FB_SIZE:
63 		param->value = dev_priv->prim_bb_mem;
64 		break;
65 	case DRM_VMW_PARAM_FIFO_HW_VERSION:
66 	{
67 		u32 *fifo_mem = dev_priv->mmio_virt;
68 		const struct vmw_fifo_state *fifo = &dev_priv->fifo;
69 
70 		if ((dev_priv->capabilities & SVGA_CAP_GBOBJECTS)) {
71 			param->value = SVGA3D_HWVERSION_WS8_B1;
72 			break;
73 		}
74 
75 		param->value =
76 			vmw_mmio_read(fifo_mem +
77 				      ((fifo->capabilities &
78 					SVGA_FIFO_CAP_3D_HWVERSION_REVISED) ?
79 				       SVGA_FIFO_3D_HWVERSION_REVISED :
80 				       SVGA_FIFO_3D_HWVERSION));
81 		break;
82 	}
83 	case DRM_VMW_PARAM_MAX_SURF_MEMORY:
84 		if ((dev_priv->capabilities & SVGA_CAP_GBOBJECTS) &&
85 		    !vmw_fp->gb_aware)
86 			param->value = dev_priv->max_mob_pages * PAGE_SIZE / 2;
87 		else
88 			param->value = dev_priv->memory_size;
89 		break;
90 	case DRM_VMW_PARAM_3D_CAPS_SIZE:
91 		if ((dev_priv->capabilities & SVGA_CAP_GBOBJECTS) &&
92 		    vmw_fp->gb_aware)
93 			param->value = SVGA3D_DEVCAP_MAX * sizeof(uint32_t);
94 		else if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS)
95 			param->value = sizeof(struct svga_3d_compat_cap) +
96 				sizeof(uint32_t);
97 		else
98 			param->value = (SVGA_FIFO_3D_CAPS_LAST -
99 					SVGA_FIFO_3D_CAPS + 1) *
100 				sizeof(uint32_t);
101 		break;
102 	case DRM_VMW_PARAM_MAX_MOB_MEMORY:
103 		vmw_fp->gb_aware = true;
104 		param->value = dev_priv->max_mob_pages * PAGE_SIZE;
105 		break;
106 	case DRM_VMW_PARAM_MAX_MOB_SIZE:
107 		param->value = dev_priv->max_mob_size;
108 		break;
109 	case DRM_VMW_PARAM_SCREEN_TARGET:
110 		param->value =
111 			(dev_priv->active_display_unit == vmw_du_screen_target);
112 		break;
113 	case DRM_VMW_PARAM_DX:
114 		param->value = dev_priv->has_dx;
115 		break;
116 	default:
117 		return -EINVAL;
118 	}
119 
120 	return 0;
121 }
122 
123 static u32 vmw_mask_multisample(unsigned int cap, u32 fmt_value)
124 {
125 	/* If the header is updated, update the format test as well! */
126 	BUILD_BUG_ON(SVGA3D_DEVCAP_DXFMT_BC5_UNORM + 1 != SVGA3D_DEVCAP_MAX);
127 
128 	if (cap >= SVGA3D_DEVCAP_DXFMT_X8R8G8B8 &&
129 	    cap <= SVGA3D_DEVCAP_DXFMT_BC5_UNORM)
130 		fmt_value &= ~(SVGADX_DXFMT_MULTISAMPLE_2 |
131 			       SVGADX_DXFMT_MULTISAMPLE_4 |
132 			       SVGADX_DXFMT_MULTISAMPLE_8);
133 	else if (cap == SVGA3D_DEVCAP_MULTISAMPLE_MASKABLESAMPLES)
134 		return 0;
135 
136 	return fmt_value;
137 }
138 
139 static int vmw_fill_compat_cap(struct vmw_private *dev_priv, void *bounce,
140 			       size_t size)
141 {
142 	struct svga_3d_compat_cap *compat_cap =
143 		(struct svga_3d_compat_cap *) bounce;
144 	unsigned int i;
145 	size_t pair_offset = offsetof(struct svga_3d_compat_cap, pairs);
146 	unsigned int max_size;
147 
148 	if (size < pair_offset)
149 		return -EINVAL;
150 
151 	max_size = (size - pair_offset) / sizeof(SVGA3dCapPair);
152 
153 	if (max_size > SVGA3D_DEVCAP_MAX)
154 		max_size = SVGA3D_DEVCAP_MAX;
155 
156 	compat_cap->header.length =
157 		(pair_offset + max_size * sizeof(SVGA3dCapPair)) / sizeof(u32);
158 	compat_cap->header.type = SVGA3DCAPS_RECORD_DEVCAPS;
159 
160 	spin_lock(&dev_priv->cap_lock);
161 	for (i = 0; i < max_size; ++i) {
162 		vmw_write(dev_priv, SVGA_REG_DEV_CAP, i);
163 		compat_cap->pairs[i][0] = i;
164 		compat_cap->pairs[i][1] = vmw_mask_multisample
165 			(i, vmw_read(dev_priv, SVGA_REG_DEV_CAP));
166 	}
167 	spin_unlock(&dev_priv->cap_lock);
168 
169 	return 0;
170 }
171 
172 
173 int vmw_get_cap_3d_ioctl(struct drm_device *dev, void *data,
174 			 struct drm_file *file_priv)
175 {
176 	struct drm_vmw_get_3d_cap_arg *arg =
177 		(struct drm_vmw_get_3d_cap_arg *) data;
178 	struct vmw_private *dev_priv = vmw_priv(dev);
179 	uint32_t size;
180 	u32 *fifo_mem;
181 	void __user *buffer = (void __user *)((unsigned long)(arg->buffer));
182 	void *bounce;
183 	int ret;
184 	bool gb_objects = !!(dev_priv->capabilities & SVGA_CAP_GBOBJECTS);
185 	struct vmw_fpriv *vmw_fp = vmw_fpriv(file_priv);
186 
187 	if (unlikely(arg->pad64 != 0 || arg->max_size == 0)) {
188 		DRM_ERROR("Illegal GET_3D_CAP argument.\n");
189 		return -EINVAL;
190 	}
191 
192 	if (gb_objects && vmw_fp->gb_aware)
193 		size = SVGA3D_DEVCAP_MAX * sizeof(uint32_t);
194 	else if (gb_objects)
195 		size = sizeof(struct svga_3d_compat_cap) + sizeof(uint32_t);
196 	else
197 		size = (SVGA_FIFO_3D_CAPS_LAST - SVGA_FIFO_3D_CAPS + 1) *
198 			sizeof(uint32_t);
199 
200 	if (arg->max_size < size)
201 		size = arg->max_size;
202 
203 	bounce = vzalloc(size);
204 	if (unlikely(bounce == NULL)) {
205 		DRM_ERROR("Failed to allocate bounce buffer for 3D caps.\n");
206 		return -ENOMEM;
207 	}
208 
209 	if (gb_objects && vmw_fp->gb_aware) {
210 		int i, num;
211 		uint32_t *bounce32 = (uint32_t *) bounce;
212 
213 		num = size / sizeof(uint32_t);
214 		if (num > SVGA3D_DEVCAP_MAX)
215 			num = SVGA3D_DEVCAP_MAX;
216 
217 		spin_lock(&dev_priv->cap_lock);
218 		for (i = 0; i < num; ++i) {
219 			vmw_write(dev_priv, SVGA_REG_DEV_CAP, i);
220 			*bounce32++ = vmw_mask_multisample
221 				(i, vmw_read(dev_priv, SVGA_REG_DEV_CAP));
222 		}
223 		spin_unlock(&dev_priv->cap_lock);
224 	} else if (gb_objects) {
225 		ret = vmw_fill_compat_cap(dev_priv, bounce, size);
226 		if (unlikely(ret != 0))
227 			goto out_err;
228 	} else {
229 		fifo_mem = dev_priv->mmio_virt;
230 		memcpy(bounce, &fifo_mem[SVGA_FIFO_3D_CAPS], size);
231 	}
232 
233 	ret = copy_to_user(buffer, bounce, size);
234 	if (ret)
235 		ret = -EFAULT;
236 out_err:
237 	vfree(bounce);
238 
239 	if (unlikely(ret != 0))
240 		DRM_ERROR("Failed to report 3D caps info.\n");
241 
242 	return ret;
243 }
244 
245 int vmw_present_ioctl(struct drm_device *dev, void *data,
246 		      struct drm_file *file_priv)
247 {
248 	struct ttm_object_file *tfile = vmw_fpriv(file_priv)->tfile;
249 	struct vmw_private *dev_priv = vmw_priv(dev);
250 	struct drm_vmw_present_arg *arg =
251 		(struct drm_vmw_present_arg *)data;
252 	struct vmw_surface *surface;
253 	struct drm_vmw_rect __user *clips_ptr;
254 	struct drm_vmw_rect *clips = NULL;
255 	struct drm_framebuffer *fb;
256 	struct vmw_framebuffer *vfb;
257 	struct vmw_resource *res;
258 	uint32_t num_clips;
259 	int ret;
260 
261 	num_clips = arg->num_clips;
262 	clips_ptr = (struct drm_vmw_rect __user *)(unsigned long)arg->clips_ptr;
263 
264 	if (unlikely(num_clips == 0))
265 		return 0;
266 
267 	if (clips_ptr == NULL) {
268 		DRM_ERROR("Variable clips_ptr must be specified.\n");
269 		ret = -EINVAL;
270 		goto out_clips;
271 	}
272 
273 	clips = kcalloc(num_clips, sizeof(*clips), GFP_KERNEL);
274 	if (clips == NULL) {
275 		DRM_ERROR("Failed to allocate clip rect list.\n");
276 		ret = -ENOMEM;
277 		goto out_clips;
278 	}
279 
280 	ret = copy_from_user(clips, clips_ptr, num_clips * sizeof(*clips));
281 	if (ret) {
282 		DRM_ERROR("Failed to copy clip rects from userspace.\n");
283 		ret = -EFAULT;
284 		goto out_no_copy;
285 	}
286 
287 	drm_modeset_lock_all(dev);
288 
289 	fb = drm_framebuffer_lookup(dev, file_priv, arg->fb_id);
290 	if (!fb) {
291 		DRM_ERROR("Invalid framebuffer id.\n");
292 		ret = -ENOENT;
293 		goto out_no_fb;
294 	}
295 	vfb = vmw_framebuffer_to_vfb(fb);
296 
297 	ret = ttm_read_lock(&dev_priv->reservation_sem, true);
298 	if (unlikely(ret != 0))
299 		goto out_no_ttm_lock;
300 
301 	ret = vmw_user_resource_lookup_handle(dev_priv, tfile, arg->sid,
302 					      user_surface_converter,
303 					      &res);
304 	if (ret)
305 		goto out_no_surface;
306 
307 	surface = vmw_res_to_srf(res);
308 	ret = vmw_kms_present(dev_priv, file_priv,
309 			      vfb, surface, arg->sid,
310 			      arg->dest_x, arg->dest_y,
311 			      clips, num_clips);
312 
313 	/* vmw_user_surface_lookup takes one ref so does new_fb */
314 	vmw_surface_unreference(&surface);
315 
316 out_no_surface:
317 	ttm_read_unlock(&dev_priv->reservation_sem);
318 out_no_ttm_lock:
319 	drm_framebuffer_unreference(fb);
320 out_no_fb:
321 	drm_modeset_unlock_all(dev);
322 out_no_copy:
323 	kfree(clips);
324 out_clips:
325 	return ret;
326 }
327 
328 int vmw_present_readback_ioctl(struct drm_device *dev, void *data,
329 			       struct drm_file *file_priv)
330 {
331 	struct vmw_private *dev_priv = vmw_priv(dev);
332 	struct drm_vmw_present_readback_arg *arg =
333 		(struct drm_vmw_present_readback_arg *)data;
334 	struct drm_vmw_fence_rep __user *user_fence_rep =
335 		(struct drm_vmw_fence_rep __user *)
336 		(unsigned long)arg->fence_rep;
337 	struct drm_vmw_rect __user *clips_ptr;
338 	struct drm_vmw_rect *clips = NULL;
339 	struct drm_framebuffer *fb;
340 	struct vmw_framebuffer *vfb;
341 	uint32_t num_clips;
342 	int ret;
343 
344 	num_clips = arg->num_clips;
345 	clips_ptr = (struct drm_vmw_rect __user *)(unsigned long)arg->clips_ptr;
346 
347 	if (unlikely(num_clips == 0))
348 		return 0;
349 
350 	if (clips_ptr == NULL) {
351 		DRM_ERROR("Argument clips_ptr must be specified.\n");
352 		ret = -EINVAL;
353 		goto out_clips;
354 	}
355 
356 	clips = kcalloc(num_clips, sizeof(*clips), GFP_KERNEL);
357 	if (clips == NULL) {
358 		DRM_ERROR("Failed to allocate clip rect list.\n");
359 		ret = -ENOMEM;
360 		goto out_clips;
361 	}
362 
363 	ret = copy_from_user(clips, clips_ptr, num_clips * sizeof(*clips));
364 	if (ret) {
365 		DRM_ERROR("Failed to copy clip rects from userspace.\n");
366 		ret = -EFAULT;
367 		goto out_no_copy;
368 	}
369 
370 	drm_modeset_lock_all(dev);
371 
372 	fb = drm_framebuffer_lookup(dev, file_priv, arg->fb_id);
373 	if (!fb) {
374 		DRM_ERROR("Invalid framebuffer id.\n");
375 		ret = -ENOENT;
376 		goto out_no_fb;
377 	}
378 
379 	vfb = vmw_framebuffer_to_vfb(fb);
380 	if (!vfb->dmabuf) {
381 		DRM_ERROR("Framebuffer not dmabuf backed.\n");
382 		ret = -EINVAL;
383 		goto out_no_ttm_lock;
384 	}
385 
386 	ret = ttm_read_lock(&dev_priv->reservation_sem, true);
387 	if (unlikely(ret != 0))
388 		goto out_no_ttm_lock;
389 
390 	ret = vmw_kms_readback(dev_priv, file_priv,
391 			       vfb, user_fence_rep,
392 			       clips, num_clips);
393 
394 	ttm_read_unlock(&dev_priv->reservation_sem);
395 out_no_ttm_lock:
396 	drm_framebuffer_unreference(fb);
397 out_no_fb:
398 	drm_modeset_unlock_all(dev);
399 out_no_copy:
400 	kfree(clips);
401 out_clips:
402 	return ret;
403 }
404 
405 
406 /**
407  * vmw_fops_poll - wrapper around the drm_poll function
408  *
409  * @filp: See the linux fops poll documentation.
410  * @wait: See the linux fops poll documentation.
411  *
412  * Wrapper around the drm_poll function that makes sure the device is
413  * processing the fifo if drm_poll decides to wait.
414  */
415 __poll_t vmw_fops_poll(struct file *filp, struct poll_table_struct *wait)
416 {
417 	struct drm_file *file_priv = filp->private_data;
418 	struct vmw_private *dev_priv =
419 		vmw_priv(file_priv->minor->dev);
420 
421 	vmw_fifo_ping_host(dev_priv, SVGA_SYNC_GENERIC);
422 	return drm_poll(filp, wait);
423 }
424 
425 
426 /**
427  * vmw_fops_read - wrapper around the drm_read function
428  *
429  * @filp: See the linux fops read documentation.
430  * @buffer: See the linux fops read documentation.
431  * @count: See the linux fops read documentation.
432  * offset: See the linux fops read documentation.
433  *
434  * Wrapper around the drm_read function that makes sure the device is
435  * processing the fifo if drm_read decides to wait.
436  */
437 ssize_t vmw_fops_read(struct file *filp, char __user *buffer,
438 		      size_t count, loff_t *offset)
439 {
440 	struct drm_file *file_priv = filp->private_data;
441 	struct vmw_private *dev_priv =
442 		vmw_priv(file_priv->minor->dev);
443 
444 	vmw_fifo_ping_host(dev_priv, SVGA_SYNC_GENERIC);
445 	return drm_read(filp, buffer, count, offset);
446 }
447