1 /* SPDX-License-Identifier: GPL-2.0 OR MIT */ 2 /* 3 * Copyright 2021-2023 VMware, Inc. 4 * 5 * Permission is hereby granted, free of charge, to any person 6 * obtaining a copy of this software and associated documentation 7 * files (the "Software"), to deal in the Software without 8 * restriction, including without limitation the rights to use, copy, 9 * modify, merge, publish, distribute, sublicense, and/or sell copies 10 * of the Software, and to permit persons to whom the Software is 11 * furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be 14 * included in all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 17 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 18 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 19 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 20 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 21 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 22 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 23 * SOFTWARE. 24 * 25 */ 26 27 #include "vmwgfx_bo.h" 28 #include "vmwgfx_drv.h" 29 30 #include "drm/drm_prime.h" 31 #include "drm/drm_gem_ttm_helper.h" 32 33 static void vmw_gem_object_free(struct drm_gem_object *gobj) 34 { 35 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(gobj); 36 if (bo) 37 ttm_bo_put(bo); 38 } 39 40 static int vmw_gem_object_open(struct drm_gem_object *obj, 41 struct drm_file *file_priv) 42 { 43 return 0; 44 } 45 46 static void vmw_gem_object_close(struct drm_gem_object *obj, 47 struct drm_file *file_priv) 48 { 49 } 50 51 static int vmw_gem_pin_private(struct drm_gem_object *obj, bool do_pin) 52 { 53 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(obj); 54 struct vmw_bo *vbo = to_vmw_bo(obj); 55 int ret; 56 57 ret = ttm_bo_reserve(bo, false, false, NULL); 58 if (unlikely(ret != 0)) 59 goto err; 60 61 vmw_bo_pin_reserved(vbo, do_pin); 62 63 ttm_bo_unreserve(bo); 64 65 err: 66 return ret; 67 } 68 69 70 static int vmw_gem_object_pin(struct drm_gem_object *obj) 71 { 72 return vmw_gem_pin_private(obj, true); 73 } 74 75 static void vmw_gem_object_unpin(struct drm_gem_object *obj) 76 { 77 vmw_gem_pin_private(obj, false); 78 } 79 80 static struct sg_table *vmw_gem_object_get_sg_table(struct drm_gem_object *obj) 81 { 82 struct ttm_buffer_object *bo = drm_gem_ttm_of_gem(obj); 83 struct vmw_ttm_tt *vmw_tt = 84 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm); 85 86 if (vmw_tt->vsgt.sgt) 87 return vmw_tt->vsgt.sgt; 88 89 return drm_prime_pages_to_sg(obj->dev, vmw_tt->dma_ttm.pages, vmw_tt->dma_ttm.num_pages); 90 } 91 92 static const struct vm_operations_struct vmw_vm_ops = { 93 .pfn_mkwrite = vmw_bo_vm_mkwrite, 94 .page_mkwrite = vmw_bo_vm_mkwrite, 95 .fault = vmw_bo_vm_fault, 96 .open = ttm_bo_vm_open, 97 .close = ttm_bo_vm_close, 98 }; 99 100 static const struct drm_gem_object_funcs vmw_gem_object_funcs = { 101 .free = vmw_gem_object_free, 102 .open = vmw_gem_object_open, 103 .close = vmw_gem_object_close, 104 .print_info = drm_gem_ttm_print_info, 105 .pin = vmw_gem_object_pin, 106 .unpin = vmw_gem_object_unpin, 107 .get_sg_table = vmw_gem_object_get_sg_table, 108 .vmap = drm_gem_ttm_vmap, 109 .vunmap = drm_gem_ttm_vunmap, 110 .mmap = drm_gem_ttm_mmap, 111 .vm_ops = &vmw_vm_ops, 112 }; 113 114 int vmw_gem_object_create_with_handle(struct vmw_private *dev_priv, 115 struct drm_file *filp, 116 uint32_t size, 117 uint32_t *handle, 118 struct vmw_bo **p_vbo) 119 { 120 int ret; 121 struct vmw_bo_params params = { 122 .domain = (dev_priv->has_mob) ? VMW_BO_DOMAIN_SYS : VMW_BO_DOMAIN_VRAM, 123 .busy_domain = VMW_BO_DOMAIN_SYS, 124 .bo_type = ttm_bo_type_device, 125 .size = size, 126 .pin = false 127 }; 128 129 ret = vmw_bo_create(dev_priv, ¶ms, p_vbo); 130 if (ret != 0) 131 goto out_no_bo; 132 133 (*p_vbo)->tbo.base.funcs = &vmw_gem_object_funcs; 134 135 ret = drm_gem_handle_create(filp, &(*p_vbo)->tbo.base, handle); 136 out_no_bo: 137 return ret; 138 } 139 140 141 int vmw_gem_object_create_ioctl(struct drm_device *dev, void *data, 142 struct drm_file *filp) 143 { 144 struct vmw_private *dev_priv = vmw_priv(dev); 145 union drm_vmw_alloc_dmabuf_arg *arg = 146 (union drm_vmw_alloc_dmabuf_arg *)data; 147 struct drm_vmw_alloc_dmabuf_req *req = &arg->req; 148 struct drm_vmw_dmabuf_rep *rep = &arg->rep; 149 struct vmw_bo *vbo; 150 uint32_t handle; 151 int ret; 152 153 ret = vmw_gem_object_create_with_handle(dev_priv, filp, 154 req->size, &handle, &vbo); 155 if (ret) 156 goto out_no_bo; 157 158 rep->handle = handle; 159 rep->map_handle = drm_vma_node_offset_addr(&vbo->tbo.base.vma_node); 160 rep->cur_gmr_id = handle; 161 rep->cur_gmr_offset = 0; 162 /* drop reference from allocate - handle holds it now */ 163 drm_gem_object_put(&vbo->tbo.base); 164 out_no_bo: 165 return ret; 166 } 167 168 #if defined(CONFIG_DEBUG_FS) 169 170 static void vmw_bo_print_info(int id, struct vmw_bo *bo, struct seq_file *m) 171 { 172 const char *placement; 173 const char *type; 174 175 switch (bo->tbo.resource->mem_type) { 176 case TTM_PL_SYSTEM: 177 placement = " CPU"; 178 break; 179 case VMW_PL_GMR: 180 placement = " GMR"; 181 break; 182 case VMW_PL_MOB: 183 placement = " MOB"; 184 break; 185 case VMW_PL_SYSTEM: 186 placement = "VCPU"; 187 break; 188 case TTM_PL_VRAM: 189 placement = "VRAM"; 190 break; 191 default: 192 placement = "None"; 193 break; 194 } 195 196 switch (bo->tbo.type) { 197 case ttm_bo_type_device: 198 type = "device"; 199 break; 200 case ttm_bo_type_kernel: 201 type = "kernel"; 202 break; 203 case ttm_bo_type_sg: 204 type = "sg "; 205 break; 206 default: 207 type = "none "; 208 break; 209 } 210 211 seq_printf(m, "\t\t0x%08x: %12zu bytes %s, type = %s", 212 id, bo->tbo.base.size, placement, type); 213 seq_printf(m, ", priority = %u, pin_count = %u, GEM refs = %d, TTM refs = %d", 214 bo->tbo.priority, 215 bo->tbo.pin_count, 216 kref_read(&bo->tbo.base.refcount), 217 kref_read(&bo->tbo.kref)); 218 seq_puts(m, "\n"); 219 } 220 221 static int vmw_debugfs_gem_info_show(struct seq_file *m, void *unused) 222 { 223 struct vmw_private *vdev = (struct vmw_private *)m->private; 224 struct drm_device *dev = &vdev->drm; 225 struct drm_file *file; 226 int r; 227 228 r = mutex_lock_interruptible(&dev->filelist_mutex); 229 if (r) 230 return r; 231 232 list_for_each_entry(file, &dev->filelist, lhead) { 233 struct task_struct *task; 234 struct drm_gem_object *gobj; 235 int id; 236 237 /* 238 * Although we have a valid reference on file->pid, that does 239 * not guarantee that the task_struct who called get_pid() is 240 * still alive (e.g. get_pid(current) => fork() => exit()). 241 * Therefore, we need to protect this ->comm access using RCU. 242 */ 243 rcu_read_lock(); 244 task = pid_task(file->pid, PIDTYPE_TGID); 245 seq_printf(m, "pid %8d command %s:\n", pid_nr(file->pid), 246 task ? task->comm : "<unknown>"); 247 rcu_read_unlock(); 248 249 spin_lock(&file->table_lock); 250 idr_for_each_entry(&file->object_idr, gobj, id) { 251 struct vmw_bo *bo = to_vmw_bo(gobj); 252 253 vmw_bo_print_info(id, bo, m); 254 } 255 spin_unlock(&file->table_lock); 256 } 257 258 mutex_unlock(&dev->filelist_mutex); 259 return 0; 260 } 261 262 DEFINE_SHOW_ATTRIBUTE(vmw_debugfs_gem_info); 263 264 #endif 265 266 void vmw_debugfs_gem_init(struct vmw_private *vdev) 267 { 268 #if defined(CONFIG_DEBUG_FS) 269 struct drm_minor *minor = vdev->drm.primary; 270 struct dentry *root = minor->debugfs_root; 271 272 debugfs_create_file("vmwgfx_gem_info", 0444, root, vdev, 273 &vmw_debugfs_gem_info_fops); 274 #endif 275 } 276