1 /* 2 * NVIDIA Tegra DRM GEM helper functions 3 * 4 * Copyright (C) 2012 Sascha Hauer, Pengutronix 5 * Copyright (C) 2013 NVIDIA CORPORATION, All rights reserved. 6 * 7 * Based on the GEM/CMA helpers 8 * 9 * Copyright (c) 2011 Samsung Electronics Co., Ltd. 10 * 11 * This program is free software; you can redistribute it and/or modify 12 * it under the terms of the GNU General Public License version 2 as 13 * published by the Free Software Foundation. 14 */ 15 16 #include <linux/dma-buf.h> 17 #include <drm/tegra_drm.h> 18 19 #include "gem.h" 20 21 static inline struct tegra_bo *host1x_to_tegra_bo(struct host1x_bo *bo) 22 { 23 return container_of(bo, struct tegra_bo, base); 24 } 25 26 static void tegra_bo_put(struct host1x_bo *bo) 27 { 28 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 29 struct drm_device *drm = obj->gem.dev; 30 31 mutex_lock(&drm->struct_mutex); 32 drm_gem_object_unreference(&obj->gem); 33 mutex_unlock(&drm->struct_mutex); 34 } 35 36 static dma_addr_t tegra_bo_pin(struct host1x_bo *bo, struct sg_table **sgt) 37 { 38 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 39 40 return obj->paddr; 41 } 42 43 static void tegra_bo_unpin(struct host1x_bo *bo, struct sg_table *sgt) 44 { 45 } 46 47 static void *tegra_bo_mmap(struct host1x_bo *bo) 48 { 49 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 50 51 return obj->vaddr; 52 } 53 54 static void tegra_bo_munmap(struct host1x_bo *bo, void *addr) 55 { 56 } 57 58 static void *tegra_bo_kmap(struct host1x_bo *bo, unsigned int page) 59 { 60 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 61 62 return obj->vaddr + page * PAGE_SIZE; 63 } 64 65 static void tegra_bo_kunmap(struct host1x_bo *bo, unsigned int page, 66 void *addr) 67 { 68 } 69 70 static struct host1x_bo *tegra_bo_get(struct host1x_bo *bo) 71 { 72 struct tegra_bo *obj = host1x_to_tegra_bo(bo); 73 struct drm_device *drm = obj->gem.dev; 74 75 mutex_lock(&drm->struct_mutex); 76 drm_gem_object_reference(&obj->gem); 77 mutex_unlock(&drm->struct_mutex); 78 79 return bo; 80 } 81 82 static const struct host1x_bo_ops tegra_bo_ops = { 83 .get = tegra_bo_get, 84 .put = tegra_bo_put, 85 .pin = tegra_bo_pin, 86 .unpin = tegra_bo_unpin, 87 .mmap = tegra_bo_mmap, 88 .munmap = tegra_bo_munmap, 89 .kmap = tegra_bo_kmap, 90 .kunmap = tegra_bo_kunmap, 91 }; 92 93 static void tegra_bo_destroy(struct drm_device *drm, struct tegra_bo *bo) 94 { 95 dma_free_writecombine(drm->dev, bo->gem.size, bo->vaddr, bo->paddr); 96 } 97 98 struct tegra_bo *tegra_bo_create(struct drm_device *drm, unsigned int size, 99 unsigned long flags) 100 { 101 struct tegra_bo *bo; 102 int err; 103 104 bo = kzalloc(sizeof(*bo), GFP_KERNEL); 105 if (!bo) 106 return ERR_PTR(-ENOMEM); 107 108 host1x_bo_init(&bo->base, &tegra_bo_ops); 109 size = round_up(size, PAGE_SIZE); 110 111 bo->vaddr = dma_alloc_writecombine(drm->dev, size, &bo->paddr, 112 GFP_KERNEL | __GFP_NOWARN); 113 if (!bo->vaddr) { 114 dev_err(drm->dev, "failed to allocate buffer with size %u\n", 115 size); 116 err = -ENOMEM; 117 goto err_dma; 118 } 119 120 err = drm_gem_object_init(drm, &bo->gem, size); 121 if (err) 122 goto err_init; 123 124 err = drm_gem_create_mmap_offset(&bo->gem); 125 if (err) 126 goto err_mmap; 127 128 if (flags & DRM_TEGRA_GEM_CREATE_TILED) 129 bo->flags |= TEGRA_BO_TILED; 130 131 if (flags & DRM_TEGRA_GEM_CREATE_BOTTOM_UP) 132 bo->flags |= TEGRA_BO_BOTTOM_UP; 133 134 return bo; 135 136 err_mmap: 137 drm_gem_object_release(&bo->gem); 138 err_init: 139 tegra_bo_destroy(drm, bo); 140 err_dma: 141 kfree(bo); 142 143 return ERR_PTR(err); 144 } 145 146 struct tegra_bo *tegra_bo_create_with_handle(struct drm_file *file, 147 struct drm_device *drm, 148 unsigned int size, 149 unsigned long flags, 150 unsigned int *handle) 151 { 152 struct tegra_bo *bo; 153 int ret; 154 155 bo = tegra_bo_create(drm, size, flags); 156 if (IS_ERR(bo)) 157 return bo; 158 159 ret = drm_gem_handle_create(file, &bo->gem, handle); 160 if (ret) 161 goto err; 162 163 drm_gem_object_unreference_unlocked(&bo->gem); 164 165 return bo; 166 167 err: 168 tegra_bo_free_object(&bo->gem); 169 return ERR_PTR(ret); 170 } 171 172 struct tegra_bo *tegra_bo_import(struct drm_device *drm, struct dma_buf *buf) 173 { 174 struct dma_buf_attachment *attach; 175 struct tegra_bo *bo; 176 ssize_t size; 177 int err; 178 179 bo = kzalloc(sizeof(*bo), GFP_KERNEL); 180 if (!bo) 181 return ERR_PTR(-ENOMEM); 182 183 host1x_bo_init(&bo->base, &tegra_bo_ops); 184 size = round_up(buf->size, PAGE_SIZE); 185 186 err = drm_gem_object_init(drm, &bo->gem, size); 187 if (err < 0) 188 goto free; 189 190 err = drm_gem_create_mmap_offset(&bo->gem); 191 if (err < 0) 192 goto release; 193 194 attach = dma_buf_attach(buf, drm->dev); 195 if (IS_ERR(attach)) { 196 err = PTR_ERR(attach); 197 goto free_mmap; 198 } 199 200 get_dma_buf(buf); 201 202 bo->sgt = dma_buf_map_attachment(attach, DMA_TO_DEVICE); 203 if (!bo->sgt) { 204 err = -ENOMEM; 205 goto detach; 206 } 207 208 if (IS_ERR(bo->sgt)) { 209 err = PTR_ERR(bo->sgt); 210 goto detach; 211 } 212 213 if (bo->sgt->nents > 1) { 214 err = -EINVAL; 215 goto detach; 216 } 217 218 bo->paddr = sg_dma_address(bo->sgt->sgl); 219 bo->gem.import_attach = attach; 220 221 return bo; 222 223 detach: 224 if (!IS_ERR_OR_NULL(bo->sgt)) 225 dma_buf_unmap_attachment(attach, bo->sgt, DMA_TO_DEVICE); 226 227 dma_buf_detach(buf, attach); 228 dma_buf_put(buf); 229 free_mmap: 230 drm_gem_free_mmap_offset(&bo->gem); 231 release: 232 drm_gem_object_release(&bo->gem); 233 free: 234 kfree(bo); 235 236 return ERR_PTR(err); 237 } 238 239 void tegra_bo_free_object(struct drm_gem_object *gem) 240 { 241 struct tegra_bo *bo = to_tegra_bo(gem); 242 243 if (gem->import_attach) { 244 dma_buf_unmap_attachment(gem->import_attach, bo->sgt, 245 DMA_TO_DEVICE); 246 drm_prime_gem_destroy(gem, NULL); 247 } else { 248 tegra_bo_destroy(gem->dev, bo); 249 } 250 251 drm_gem_free_mmap_offset(gem); 252 drm_gem_object_release(gem); 253 254 kfree(bo); 255 } 256 257 int tegra_bo_dumb_create(struct drm_file *file, struct drm_device *drm, 258 struct drm_mode_create_dumb *args) 259 { 260 int min_pitch = DIV_ROUND_UP(args->width * args->bpp, 8); 261 struct tegra_bo *bo; 262 263 if (args->pitch < min_pitch) 264 args->pitch = min_pitch; 265 266 if (args->size < args->pitch * args->height) 267 args->size = args->pitch * args->height; 268 269 bo = tegra_bo_create_with_handle(file, drm, args->size, 0, 270 &args->handle); 271 if (IS_ERR(bo)) 272 return PTR_ERR(bo); 273 274 return 0; 275 } 276 277 int tegra_bo_dumb_map_offset(struct drm_file *file, struct drm_device *drm, 278 uint32_t handle, uint64_t *offset) 279 { 280 struct drm_gem_object *gem; 281 struct tegra_bo *bo; 282 283 mutex_lock(&drm->struct_mutex); 284 285 gem = drm_gem_object_lookup(drm, file, handle); 286 if (!gem) { 287 dev_err(drm->dev, "failed to lookup GEM object\n"); 288 mutex_unlock(&drm->struct_mutex); 289 return -EINVAL; 290 } 291 292 bo = to_tegra_bo(gem); 293 294 *offset = drm_vma_node_offset_addr(&bo->gem.vma_node); 295 296 drm_gem_object_unreference(gem); 297 298 mutex_unlock(&drm->struct_mutex); 299 300 return 0; 301 } 302 303 const struct vm_operations_struct tegra_bo_vm_ops = { 304 .open = drm_gem_vm_open, 305 .close = drm_gem_vm_close, 306 }; 307 308 int tegra_drm_mmap(struct file *file, struct vm_area_struct *vma) 309 { 310 struct drm_gem_object *gem; 311 struct tegra_bo *bo; 312 int ret; 313 314 ret = drm_gem_mmap(file, vma); 315 if (ret) 316 return ret; 317 318 gem = vma->vm_private_data; 319 bo = to_tegra_bo(gem); 320 321 ret = remap_pfn_range(vma, vma->vm_start, bo->paddr >> PAGE_SHIFT, 322 vma->vm_end - vma->vm_start, vma->vm_page_prot); 323 if (ret) 324 drm_gem_vm_close(vma); 325 326 return ret; 327 } 328 329 static struct sg_table * 330 tegra_gem_prime_map_dma_buf(struct dma_buf_attachment *attach, 331 enum dma_data_direction dir) 332 { 333 struct drm_gem_object *gem = attach->dmabuf->priv; 334 struct tegra_bo *bo = to_tegra_bo(gem); 335 struct sg_table *sgt; 336 337 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); 338 if (!sgt) 339 return NULL; 340 341 if (sg_alloc_table(sgt, 1, GFP_KERNEL)) { 342 kfree(sgt); 343 return NULL; 344 } 345 346 sg_dma_address(sgt->sgl) = bo->paddr; 347 sg_dma_len(sgt->sgl) = gem->size; 348 349 return sgt; 350 } 351 352 static void tegra_gem_prime_unmap_dma_buf(struct dma_buf_attachment *attach, 353 struct sg_table *sgt, 354 enum dma_data_direction dir) 355 { 356 sg_free_table(sgt); 357 kfree(sgt); 358 } 359 360 static void tegra_gem_prime_release(struct dma_buf *buf) 361 { 362 drm_gem_dmabuf_release(buf); 363 } 364 365 static void *tegra_gem_prime_kmap_atomic(struct dma_buf *buf, 366 unsigned long page) 367 { 368 return NULL; 369 } 370 371 static void tegra_gem_prime_kunmap_atomic(struct dma_buf *buf, 372 unsigned long page, 373 void *addr) 374 { 375 } 376 377 static void *tegra_gem_prime_kmap(struct dma_buf *buf, unsigned long page) 378 { 379 return NULL; 380 } 381 382 static void tegra_gem_prime_kunmap(struct dma_buf *buf, unsigned long page, 383 void *addr) 384 { 385 } 386 387 static int tegra_gem_prime_mmap(struct dma_buf *buf, struct vm_area_struct *vma) 388 { 389 return -EINVAL; 390 } 391 392 static void *tegra_gem_prime_vmap(struct dma_buf *buf) 393 { 394 struct drm_gem_object *gem = buf->priv; 395 struct tegra_bo *bo = to_tegra_bo(gem); 396 397 return bo->vaddr; 398 } 399 400 static void tegra_gem_prime_vunmap(struct dma_buf *buf, void *vaddr) 401 { 402 } 403 404 static const struct dma_buf_ops tegra_gem_prime_dmabuf_ops = { 405 .map_dma_buf = tegra_gem_prime_map_dma_buf, 406 .unmap_dma_buf = tegra_gem_prime_unmap_dma_buf, 407 .release = tegra_gem_prime_release, 408 .kmap_atomic = tegra_gem_prime_kmap_atomic, 409 .kunmap_atomic = tegra_gem_prime_kunmap_atomic, 410 .kmap = tegra_gem_prime_kmap, 411 .kunmap = tegra_gem_prime_kunmap, 412 .mmap = tegra_gem_prime_mmap, 413 .vmap = tegra_gem_prime_vmap, 414 .vunmap = tegra_gem_prime_vunmap, 415 }; 416 417 struct dma_buf *tegra_gem_prime_export(struct drm_device *drm, 418 struct drm_gem_object *gem, 419 int flags) 420 { 421 return dma_buf_export(gem, &tegra_gem_prime_dmabuf_ops, gem->size, 422 flags); 423 } 424 425 struct drm_gem_object *tegra_gem_prime_import(struct drm_device *drm, 426 struct dma_buf *buf) 427 { 428 struct tegra_bo *bo; 429 430 if (buf->ops == &tegra_gem_prime_dmabuf_ops) { 431 struct drm_gem_object *gem = buf->priv; 432 433 if (gem->dev == drm) { 434 drm_gem_object_reference(gem); 435 return gem; 436 } 437 } 438 439 bo = tegra_bo_import(drm, buf); 440 if (IS_ERR(bo)) 441 return ERR_CAST(bo); 442 443 return &bo->gem; 444 } 445