xref: /openbmc/linux/drivers/gpu/drm/tegra/gem.c (revision ecfb9f40)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * NVIDIA Tegra DRM GEM helper functions
4  *
5  * Copyright (C) 2012 Sascha Hauer, Pengutronix
6  * Copyright (C) 2013-2015 NVIDIA CORPORATION, All rights reserved.
7  *
8  * Based on the GEM/CMA helpers
9  *
10  * Copyright (c) 2011 Samsung Electronics Co., Ltd.
11  */
12 
13 #include <linux/dma-buf.h>
14 #include <linux/iommu.h>
15 #include <linux/module.h>
16 
17 #include <drm/drm_drv.h>
18 #include <drm/drm_prime.h>
19 #include <drm/tegra_drm.h>
20 
21 #include "drm.h"
22 #include "gem.h"
23 
24 MODULE_IMPORT_NS(DMA_BUF);
25 
26 static unsigned int sg_dma_count_chunks(struct scatterlist *sgl, unsigned int nents)
27 {
28 	dma_addr_t next = ~(dma_addr_t)0;
29 	unsigned int count = 0, i;
30 	struct scatterlist *s;
31 
32 	for_each_sg(sgl, s, nents, i) {
33 		/* sg_dma_address(s) is only valid for entries that have sg_dma_len(s) != 0. */
34 		if (!sg_dma_len(s))
35 			continue;
36 
37 		if (sg_dma_address(s) != next) {
38 			next = sg_dma_address(s) + sg_dma_len(s);
39 			count++;
40 		}
41 	}
42 
43 	return count;
44 }
45 
46 static inline unsigned int sgt_dma_count_chunks(struct sg_table *sgt)
47 {
48 	return sg_dma_count_chunks(sgt->sgl, sgt->nents);
49 }
50 
51 static void tegra_bo_put(struct host1x_bo *bo)
52 {
53 	struct tegra_bo *obj = host1x_to_tegra_bo(bo);
54 
55 	drm_gem_object_put(&obj->gem);
56 }
57 
58 static struct host1x_bo_mapping *tegra_bo_pin(struct device *dev, struct host1x_bo *bo,
59 					      enum dma_data_direction direction)
60 {
61 	struct tegra_bo *obj = host1x_to_tegra_bo(bo);
62 	struct drm_gem_object *gem = &obj->gem;
63 	struct host1x_bo_mapping *map;
64 	int err;
65 
66 	map = kzalloc(sizeof(*map), GFP_KERNEL);
67 	if (!map)
68 		return ERR_PTR(-ENOMEM);
69 
70 	kref_init(&map->ref);
71 	map->bo = host1x_bo_get(bo);
72 	map->direction = direction;
73 	map->dev = dev;
74 
75 	/*
76 	 * Imported buffers need special treatment to satisfy the semantics of DMA-BUF.
77 	 */
78 	if (gem->import_attach) {
79 		struct dma_buf *buf = gem->import_attach->dmabuf;
80 
81 		map->attach = dma_buf_attach(buf, dev);
82 		if (IS_ERR(map->attach)) {
83 			err = PTR_ERR(map->attach);
84 			goto free;
85 		}
86 
87 		map->sgt = dma_buf_map_attachment_unlocked(map->attach, direction);
88 		if (IS_ERR(map->sgt)) {
89 			dma_buf_detach(buf, map->attach);
90 			err = PTR_ERR(map->sgt);
91 			map->sgt = NULL;
92 			goto free;
93 		}
94 
95 		err = sgt_dma_count_chunks(map->sgt);
96 		map->size = gem->size;
97 
98 		goto out;
99 	}
100 
101 	/*
102 	 * If we don't have a mapping for this buffer yet, return an SG table
103 	 * so that host1x can do the mapping for us via the DMA API.
104 	 */
105 	map->sgt = kzalloc(sizeof(*map->sgt), GFP_KERNEL);
106 	if (!map->sgt) {
107 		err = -ENOMEM;
108 		goto free;
109 	}
110 
111 	if (obj->pages) {
112 		/*
113 		 * If the buffer object was allocated from the explicit IOMMU
114 		 * API code paths, construct an SG table from the pages.
115 		 */
116 		err = sg_alloc_table_from_pages(map->sgt, obj->pages, obj->num_pages, 0, gem->size,
117 						GFP_KERNEL);
118 		if (err < 0)
119 			goto free;
120 	} else {
121 		/*
122 		 * If the buffer object had no pages allocated and if it was
123 		 * not imported, it had to be allocated with the DMA API, so
124 		 * the DMA API helper can be used.
125 		 */
126 		err = dma_get_sgtable(dev, map->sgt, obj->vaddr, obj->iova, gem->size);
127 		if (err < 0)
128 			goto free;
129 	}
130 
131 	err = dma_map_sgtable(dev, map->sgt, direction, 0);
132 	if (err)
133 		goto free_sgt;
134 
135 out:
136 	/*
137 	 * If we've manually mapped the buffer object through the IOMMU, make sure to return the
138 	 * existing IOVA address of our mapping.
139 	 */
140 	if (!obj->mm) {
141 		map->phys = sg_dma_address(map->sgt->sgl);
142 		map->chunks = err;
143 	} else {
144 		map->phys = obj->iova;
145 		map->chunks = 1;
146 	}
147 
148 	map->size = gem->size;
149 
150 	return map;
151 
152 free_sgt:
153 	sg_free_table(map->sgt);
154 free:
155 	kfree(map->sgt);
156 	kfree(map);
157 	return ERR_PTR(err);
158 }
159 
160 static void tegra_bo_unpin(struct host1x_bo_mapping *map)
161 {
162 	if (map->attach) {
163 		dma_buf_unmap_attachment_unlocked(map->attach, map->sgt,
164 						  map->direction);
165 		dma_buf_detach(map->attach->dmabuf, map->attach);
166 	} else {
167 		dma_unmap_sgtable(map->dev, map->sgt, map->direction, 0);
168 		sg_free_table(map->sgt);
169 		kfree(map->sgt);
170 	}
171 
172 	host1x_bo_put(map->bo);
173 	kfree(map);
174 }
175 
176 static void *tegra_bo_mmap(struct host1x_bo *bo)
177 {
178 	struct tegra_bo *obj = host1x_to_tegra_bo(bo);
179 	struct iosys_map map;
180 	int ret;
181 
182 	if (obj->vaddr) {
183 		return obj->vaddr;
184 	} else if (obj->gem.import_attach) {
185 		ret = dma_buf_vmap_unlocked(obj->gem.import_attach->dmabuf, &map);
186 		return ret ? NULL : map.vaddr;
187 	} else {
188 		return vmap(obj->pages, obj->num_pages, VM_MAP,
189 			    pgprot_writecombine(PAGE_KERNEL));
190 	}
191 }
192 
193 static void tegra_bo_munmap(struct host1x_bo *bo, void *addr)
194 {
195 	struct tegra_bo *obj = host1x_to_tegra_bo(bo);
196 	struct iosys_map map = IOSYS_MAP_INIT_VADDR(addr);
197 
198 	if (obj->vaddr)
199 		return;
200 	else if (obj->gem.import_attach)
201 		dma_buf_vunmap_unlocked(obj->gem.import_attach->dmabuf, &map);
202 	else
203 		vunmap(addr);
204 }
205 
206 static struct host1x_bo *tegra_bo_get(struct host1x_bo *bo)
207 {
208 	struct tegra_bo *obj = host1x_to_tegra_bo(bo);
209 
210 	drm_gem_object_get(&obj->gem);
211 
212 	return bo;
213 }
214 
215 static const struct host1x_bo_ops tegra_bo_ops = {
216 	.get = tegra_bo_get,
217 	.put = tegra_bo_put,
218 	.pin = tegra_bo_pin,
219 	.unpin = tegra_bo_unpin,
220 	.mmap = tegra_bo_mmap,
221 	.munmap = tegra_bo_munmap,
222 };
223 
224 static int tegra_bo_iommu_map(struct tegra_drm *tegra, struct tegra_bo *bo)
225 {
226 	int prot = IOMMU_READ | IOMMU_WRITE;
227 	int err;
228 
229 	if (bo->mm)
230 		return -EBUSY;
231 
232 	bo->mm = kzalloc(sizeof(*bo->mm), GFP_KERNEL);
233 	if (!bo->mm)
234 		return -ENOMEM;
235 
236 	mutex_lock(&tegra->mm_lock);
237 
238 	err = drm_mm_insert_node_generic(&tegra->mm,
239 					 bo->mm, bo->gem.size, PAGE_SIZE, 0, 0);
240 	if (err < 0) {
241 		dev_err(tegra->drm->dev, "out of I/O virtual memory: %d\n",
242 			err);
243 		goto unlock;
244 	}
245 
246 	bo->iova = bo->mm->start;
247 
248 	bo->size = iommu_map_sgtable(tegra->domain, bo->iova, bo->sgt, prot);
249 	if (!bo->size) {
250 		dev_err(tegra->drm->dev, "failed to map buffer\n");
251 		err = -ENOMEM;
252 		goto remove;
253 	}
254 
255 	mutex_unlock(&tegra->mm_lock);
256 
257 	return 0;
258 
259 remove:
260 	drm_mm_remove_node(bo->mm);
261 unlock:
262 	mutex_unlock(&tegra->mm_lock);
263 	kfree(bo->mm);
264 	return err;
265 }
266 
267 static int tegra_bo_iommu_unmap(struct tegra_drm *tegra, struct tegra_bo *bo)
268 {
269 	if (!bo->mm)
270 		return 0;
271 
272 	mutex_lock(&tegra->mm_lock);
273 	iommu_unmap(tegra->domain, bo->iova, bo->size);
274 	drm_mm_remove_node(bo->mm);
275 	mutex_unlock(&tegra->mm_lock);
276 
277 	kfree(bo->mm);
278 
279 	return 0;
280 }
281 
282 static const struct drm_gem_object_funcs tegra_gem_object_funcs = {
283 	.free = tegra_bo_free_object,
284 	.export = tegra_gem_prime_export,
285 	.vm_ops = &tegra_bo_vm_ops,
286 };
287 
288 static struct tegra_bo *tegra_bo_alloc_object(struct drm_device *drm,
289 					      size_t size)
290 {
291 	struct tegra_bo *bo;
292 	int err;
293 
294 	bo = kzalloc(sizeof(*bo), GFP_KERNEL);
295 	if (!bo)
296 		return ERR_PTR(-ENOMEM);
297 
298 	bo->gem.funcs = &tegra_gem_object_funcs;
299 
300 	host1x_bo_init(&bo->base, &tegra_bo_ops);
301 	size = round_up(size, PAGE_SIZE);
302 
303 	err = drm_gem_object_init(drm, &bo->gem, size);
304 	if (err < 0)
305 		goto free;
306 
307 	err = drm_gem_create_mmap_offset(&bo->gem);
308 	if (err < 0)
309 		goto release;
310 
311 	return bo;
312 
313 release:
314 	drm_gem_object_release(&bo->gem);
315 free:
316 	kfree(bo);
317 	return ERR_PTR(err);
318 }
319 
320 static void tegra_bo_free(struct drm_device *drm, struct tegra_bo *bo)
321 {
322 	if (bo->pages) {
323 		dma_unmap_sgtable(drm->dev, bo->sgt, DMA_FROM_DEVICE, 0);
324 		drm_gem_put_pages(&bo->gem, bo->pages, true, true);
325 		sg_free_table(bo->sgt);
326 		kfree(bo->sgt);
327 	} else if (bo->vaddr) {
328 		dma_free_wc(drm->dev, bo->gem.size, bo->vaddr, bo->iova);
329 	}
330 }
331 
332 static int tegra_bo_get_pages(struct drm_device *drm, struct tegra_bo *bo)
333 {
334 	int err;
335 
336 	bo->pages = drm_gem_get_pages(&bo->gem);
337 	if (IS_ERR(bo->pages))
338 		return PTR_ERR(bo->pages);
339 
340 	bo->num_pages = bo->gem.size >> PAGE_SHIFT;
341 
342 	bo->sgt = drm_prime_pages_to_sg(bo->gem.dev, bo->pages, bo->num_pages);
343 	if (IS_ERR(bo->sgt)) {
344 		err = PTR_ERR(bo->sgt);
345 		goto put_pages;
346 	}
347 
348 	err = dma_map_sgtable(drm->dev, bo->sgt, DMA_FROM_DEVICE, 0);
349 	if (err)
350 		goto free_sgt;
351 
352 	return 0;
353 
354 free_sgt:
355 	sg_free_table(bo->sgt);
356 	kfree(bo->sgt);
357 put_pages:
358 	drm_gem_put_pages(&bo->gem, bo->pages, false, false);
359 	return err;
360 }
361 
362 static int tegra_bo_alloc(struct drm_device *drm, struct tegra_bo *bo)
363 {
364 	struct tegra_drm *tegra = drm->dev_private;
365 	int err;
366 
367 	if (tegra->domain) {
368 		err = tegra_bo_get_pages(drm, bo);
369 		if (err < 0)
370 			return err;
371 
372 		err = tegra_bo_iommu_map(tegra, bo);
373 		if (err < 0) {
374 			tegra_bo_free(drm, bo);
375 			return err;
376 		}
377 	} else {
378 		size_t size = bo->gem.size;
379 
380 		bo->vaddr = dma_alloc_wc(drm->dev, size, &bo->iova,
381 					 GFP_KERNEL | __GFP_NOWARN);
382 		if (!bo->vaddr) {
383 			dev_err(drm->dev,
384 				"failed to allocate buffer of size %zu\n",
385 				size);
386 			return -ENOMEM;
387 		}
388 	}
389 
390 	return 0;
391 }
392 
393 struct tegra_bo *tegra_bo_create(struct drm_device *drm, size_t size,
394 				 unsigned long flags)
395 {
396 	struct tegra_bo *bo;
397 	int err;
398 
399 	bo = tegra_bo_alloc_object(drm, size);
400 	if (IS_ERR(bo))
401 		return bo;
402 
403 	err = tegra_bo_alloc(drm, bo);
404 	if (err < 0)
405 		goto release;
406 
407 	if (flags & DRM_TEGRA_GEM_CREATE_TILED)
408 		bo->tiling.mode = TEGRA_BO_TILING_MODE_TILED;
409 
410 	if (flags & DRM_TEGRA_GEM_CREATE_BOTTOM_UP)
411 		bo->flags |= TEGRA_BO_BOTTOM_UP;
412 
413 	return bo;
414 
415 release:
416 	drm_gem_object_release(&bo->gem);
417 	kfree(bo);
418 	return ERR_PTR(err);
419 }
420 
421 struct tegra_bo *tegra_bo_create_with_handle(struct drm_file *file,
422 					     struct drm_device *drm,
423 					     size_t size,
424 					     unsigned long flags,
425 					     u32 *handle)
426 {
427 	struct tegra_bo *bo;
428 	int err;
429 
430 	bo = tegra_bo_create(drm, size, flags);
431 	if (IS_ERR(bo))
432 		return bo;
433 
434 	err = drm_gem_handle_create(file, &bo->gem, handle);
435 	if (err) {
436 		tegra_bo_free_object(&bo->gem);
437 		return ERR_PTR(err);
438 	}
439 
440 	drm_gem_object_put(&bo->gem);
441 
442 	return bo;
443 }
444 
445 static struct tegra_bo *tegra_bo_import(struct drm_device *drm,
446 					struct dma_buf *buf)
447 {
448 	struct tegra_drm *tegra = drm->dev_private;
449 	struct dma_buf_attachment *attach;
450 	struct tegra_bo *bo;
451 	int err;
452 
453 	bo = tegra_bo_alloc_object(drm, buf->size);
454 	if (IS_ERR(bo))
455 		return bo;
456 
457 	attach = dma_buf_attach(buf, drm->dev);
458 	if (IS_ERR(attach)) {
459 		err = PTR_ERR(attach);
460 		goto free;
461 	}
462 
463 	get_dma_buf(buf);
464 
465 	bo->sgt = dma_buf_map_attachment_unlocked(attach, DMA_TO_DEVICE);
466 	if (IS_ERR(bo->sgt)) {
467 		err = PTR_ERR(bo->sgt);
468 		goto detach;
469 	}
470 
471 	if (tegra->domain) {
472 		err = tegra_bo_iommu_map(tegra, bo);
473 		if (err < 0)
474 			goto detach;
475 	}
476 
477 	bo->gem.import_attach = attach;
478 
479 	return bo;
480 
481 detach:
482 	if (!IS_ERR_OR_NULL(bo->sgt))
483 		dma_buf_unmap_attachment_unlocked(attach, bo->sgt, DMA_TO_DEVICE);
484 
485 	dma_buf_detach(buf, attach);
486 	dma_buf_put(buf);
487 free:
488 	drm_gem_object_release(&bo->gem);
489 	kfree(bo);
490 	return ERR_PTR(err);
491 }
492 
493 void tegra_bo_free_object(struct drm_gem_object *gem)
494 {
495 	struct tegra_drm *tegra = gem->dev->dev_private;
496 	struct host1x_bo_mapping *mapping, *tmp;
497 	struct tegra_bo *bo = to_tegra_bo(gem);
498 
499 	/* remove all mappings of this buffer object from any caches */
500 	list_for_each_entry_safe(mapping, tmp, &bo->base.mappings, list) {
501 		if (mapping->cache)
502 			host1x_bo_unpin(mapping);
503 		else
504 			dev_err(gem->dev->dev, "mapping %p stale for device %s\n", mapping,
505 				dev_name(mapping->dev));
506 	}
507 
508 	if (tegra->domain)
509 		tegra_bo_iommu_unmap(tegra, bo);
510 
511 	if (gem->import_attach) {
512 		dma_buf_unmap_attachment_unlocked(gem->import_attach, bo->sgt,
513 						  DMA_TO_DEVICE);
514 		drm_prime_gem_destroy(gem, NULL);
515 	} else {
516 		tegra_bo_free(gem->dev, bo);
517 	}
518 
519 	drm_gem_object_release(gem);
520 	kfree(bo);
521 }
522 
523 int tegra_bo_dumb_create(struct drm_file *file, struct drm_device *drm,
524 			 struct drm_mode_create_dumb *args)
525 {
526 	unsigned int min_pitch = DIV_ROUND_UP(args->width * args->bpp, 8);
527 	struct tegra_drm *tegra = drm->dev_private;
528 	struct tegra_bo *bo;
529 
530 	args->pitch = round_up(min_pitch, tegra->pitch_align);
531 	args->size = args->pitch * args->height;
532 
533 	bo = tegra_bo_create_with_handle(file, drm, args->size, 0,
534 					 &args->handle);
535 	if (IS_ERR(bo))
536 		return PTR_ERR(bo);
537 
538 	return 0;
539 }
540 
541 static vm_fault_t tegra_bo_fault(struct vm_fault *vmf)
542 {
543 	struct vm_area_struct *vma = vmf->vma;
544 	struct drm_gem_object *gem = vma->vm_private_data;
545 	struct tegra_bo *bo = to_tegra_bo(gem);
546 	struct page *page;
547 	pgoff_t offset;
548 
549 	if (!bo->pages)
550 		return VM_FAULT_SIGBUS;
551 
552 	offset = (vmf->address - vma->vm_start) >> PAGE_SHIFT;
553 	page = bo->pages[offset];
554 
555 	return vmf_insert_page(vma, vmf->address, page);
556 }
557 
558 const struct vm_operations_struct tegra_bo_vm_ops = {
559 	.fault = tegra_bo_fault,
560 	.open = drm_gem_vm_open,
561 	.close = drm_gem_vm_close,
562 };
563 
564 int __tegra_gem_mmap(struct drm_gem_object *gem, struct vm_area_struct *vma)
565 {
566 	struct tegra_bo *bo = to_tegra_bo(gem);
567 
568 	if (!bo->pages) {
569 		unsigned long vm_pgoff = vma->vm_pgoff;
570 		int err;
571 
572 		/*
573 		 * Clear the VM_PFNMAP flag that was set by drm_gem_mmap(),
574 		 * and set the vm_pgoff (used as a fake buffer offset by DRM)
575 		 * to 0 as we want to map the whole buffer.
576 		 */
577 		vma->vm_flags &= ~VM_PFNMAP;
578 		vma->vm_pgoff = 0;
579 
580 		err = dma_mmap_wc(gem->dev->dev, vma, bo->vaddr, bo->iova,
581 				  gem->size);
582 		if (err < 0) {
583 			drm_gem_vm_close(vma);
584 			return err;
585 		}
586 
587 		vma->vm_pgoff = vm_pgoff;
588 	} else {
589 		pgprot_t prot = vm_get_page_prot(vma->vm_flags);
590 
591 		vma->vm_flags |= VM_MIXEDMAP;
592 		vma->vm_flags &= ~VM_PFNMAP;
593 
594 		vma->vm_page_prot = pgprot_writecombine(prot);
595 	}
596 
597 	return 0;
598 }
599 
600 int tegra_drm_mmap(struct file *file, struct vm_area_struct *vma)
601 {
602 	struct drm_gem_object *gem;
603 	int err;
604 
605 	err = drm_gem_mmap(file, vma);
606 	if (err < 0)
607 		return err;
608 
609 	gem = vma->vm_private_data;
610 
611 	return __tegra_gem_mmap(gem, vma);
612 }
613 
614 static struct sg_table *
615 tegra_gem_prime_map_dma_buf(struct dma_buf_attachment *attach,
616 			    enum dma_data_direction dir)
617 {
618 	struct drm_gem_object *gem = attach->dmabuf->priv;
619 	struct tegra_bo *bo = to_tegra_bo(gem);
620 	struct sg_table *sgt;
621 
622 	sgt = kmalloc(sizeof(*sgt), GFP_KERNEL);
623 	if (!sgt)
624 		return NULL;
625 
626 	if (bo->pages) {
627 		if (sg_alloc_table_from_pages(sgt, bo->pages, bo->num_pages,
628 					      0, gem->size, GFP_KERNEL) < 0)
629 			goto free;
630 	} else {
631 		if (dma_get_sgtable(attach->dev, sgt, bo->vaddr, bo->iova,
632 				    gem->size) < 0)
633 			goto free;
634 	}
635 
636 	if (dma_map_sgtable(attach->dev, sgt, dir, 0))
637 		goto free;
638 
639 	return sgt;
640 
641 free:
642 	sg_free_table(sgt);
643 	kfree(sgt);
644 	return NULL;
645 }
646 
647 static void tegra_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach,
648 					  struct sg_table *sgt,
649 					  enum dma_data_direction dir)
650 {
651 	struct drm_gem_object *gem = attach->dmabuf->priv;
652 	struct tegra_bo *bo = to_tegra_bo(gem);
653 
654 	if (bo->pages)
655 		dma_unmap_sgtable(attach->dev, sgt, dir, 0);
656 
657 	sg_free_table(sgt);
658 	kfree(sgt);
659 }
660 
661 static void tegra_gem_prime_release(struct dma_buf *buf)
662 {
663 	drm_gem_dmabuf_release(buf);
664 }
665 
666 static int tegra_gem_prime_begin_cpu_access(struct dma_buf *buf,
667 					    enum dma_data_direction direction)
668 {
669 	struct drm_gem_object *gem = buf->priv;
670 	struct tegra_bo *bo = to_tegra_bo(gem);
671 	struct drm_device *drm = gem->dev;
672 
673 	if (bo->pages)
674 		dma_sync_sgtable_for_cpu(drm->dev, bo->sgt, DMA_FROM_DEVICE);
675 
676 	return 0;
677 }
678 
679 static int tegra_gem_prime_end_cpu_access(struct dma_buf *buf,
680 					  enum dma_data_direction direction)
681 {
682 	struct drm_gem_object *gem = buf->priv;
683 	struct tegra_bo *bo = to_tegra_bo(gem);
684 	struct drm_device *drm = gem->dev;
685 
686 	if (bo->pages)
687 		dma_sync_sgtable_for_device(drm->dev, bo->sgt, DMA_TO_DEVICE);
688 
689 	return 0;
690 }
691 
692 static int tegra_gem_prime_mmap(struct dma_buf *buf, struct vm_area_struct *vma)
693 {
694 	struct drm_gem_object *gem = buf->priv;
695 	int err;
696 
697 	dma_resv_assert_held(buf->resv);
698 
699 	err = drm_gem_mmap_obj(gem, gem->size, vma);
700 	if (err < 0)
701 		return err;
702 
703 	return __tegra_gem_mmap(gem, vma);
704 }
705 
706 static int tegra_gem_prime_vmap(struct dma_buf *buf, struct iosys_map *map)
707 {
708 	struct drm_gem_object *gem = buf->priv;
709 	struct tegra_bo *bo = to_tegra_bo(gem);
710 	void *vaddr;
711 
712 	vaddr = tegra_bo_mmap(&bo->base);
713 	if (IS_ERR(vaddr))
714 		return PTR_ERR(vaddr);
715 
716 	iosys_map_set_vaddr(map, vaddr);
717 
718 	return 0;
719 }
720 
721 static void tegra_gem_prime_vunmap(struct dma_buf *buf, struct iosys_map *map)
722 {
723 	struct drm_gem_object *gem = buf->priv;
724 	struct tegra_bo *bo = to_tegra_bo(gem);
725 
726 	tegra_bo_munmap(&bo->base, map->vaddr);
727 }
728 
729 static const struct dma_buf_ops tegra_gem_prime_dmabuf_ops = {
730 	.map_dma_buf = tegra_gem_prime_map_dma_buf,
731 	.unmap_dma_buf = tegra_gem_prime_unmap_dma_buf,
732 	.release = tegra_gem_prime_release,
733 	.begin_cpu_access = tegra_gem_prime_begin_cpu_access,
734 	.end_cpu_access = tegra_gem_prime_end_cpu_access,
735 	.mmap = tegra_gem_prime_mmap,
736 	.vmap = tegra_gem_prime_vmap,
737 	.vunmap = tegra_gem_prime_vunmap,
738 };
739 
740 struct dma_buf *tegra_gem_prime_export(struct drm_gem_object *gem,
741 				       int flags)
742 {
743 	DEFINE_DMA_BUF_EXPORT_INFO(exp_info);
744 
745 	exp_info.exp_name = KBUILD_MODNAME;
746 	exp_info.owner = gem->dev->driver->fops->owner;
747 	exp_info.ops = &tegra_gem_prime_dmabuf_ops;
748 	exp_info.size = gem->size;
749 	exp_info.flags = flags;
750 	exp_info.priv = gem;
751 
752 	return drm_gem_dmabuf_export(gem->dev, &exp_info);
753 }
754 
755 struct drm_gem_object *tegra_gem_prime_import(struct drm_device *drm,
756 					      struct dma_buf *buf)
757 {
758 	struct tegra_bo *bo;
759 
760 	if (buf->ops == &tegra_gem_prime_dmabuf_ops) {
761 		struct drm_gem_object *gem = buf->priv;
762 
763 		if (gem->dev == drm) {
764 			drm_gem_object_get(gem);
765 			return gem;
766 		}
767 	}
768 
769 	bo = tegra_bo_import(drm, buf);
770 	if (IS_ERR(bo))
771 		return ERR_CAST(bo);
772 
773 	return &bo->gem;
774 }
775 
776 struct host1x_bo *tegra_gem_lookup(struct drm_file *file, u32 handle)
777 {
778 	struct drm_gem_object *gem;
779 	struct tegra_bo *bo;
780 
781 	gem = drm_gem_object_lookup(file, handle);
782 	if (!gem)
783 		return NULL;
784 
785 	bo = to_tegra_bo(gem);
786 	return &bo->base;
787 }
788