1 /* 2 * Copyright 2016 Advanced Micro Devices, Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 * Authors: Christian König 23 */ 24 25 #include <drm/ttm/ttm_range_manager.h> 26 27 #include "amdgpu.h" 28 29 struct amdgpu_gtt_node { 30 struct ttm_buffer_object *tbo; 31 struct ttm_range_mgr_node base; 32 }; 33 34 static inline struct amdgpu_gtt_mgr * 35 to_gtt_mgr(struct ttm_resource_manager *man) 36 { 37 return container_of(man, struct amdgpu_gtt_mgr, manager); 38 } 39 40 static inline struct amdgpu_gtt_node * 41 to_amdgpu_gtt_node(struct ttm_resource *res) 42 { 43 return container_of(res, struct amdgpu_gtt_node, base.base); 44 } 45 46 /** 47 * DOC: mem_info_gtt_total 48 * 49 * The amdgpu driver provides a sysfs API for reporting current total size of 50 * the GTT. 51 * The file mem_info_gtt_total is used for this, and returns the total size of 52 * the GTT block, in bytes 53 */ 54 static ssize_t amdgpu_mem_info_gtt_total_show(struct device *dev, 55 struct device_attribute *attr, 56 char *buf) 57 { 58 struct drm_device *ddev = dev_get_drvdata(dev); 59 struct amdgpu_device *adev = drm_to_adev(ddev); 60 struct ttm_resource_manager *man; 61 62 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); 63 return sysfs_emit(buf, "%llu\n", man->size * PAGE_SIZE); 64 } 65 66 /** 67 * DOC: mem_info_gtt_used 68 * 69 * The amdgpu driver provides a sysfs API for reporting current total amount of 70 * used GTT. 71 * The file mem_info_gtt_used is used for this, and returns the current used 72 * size of the GTT block, in bytes 73 */ 74 static ssize_t amdgpu_mem_info_gtt_used_show(struct device *dev, 75 struct device_attribute *attr, 76 char *buf) 77 { 78 struct drm_device *ddev = dev_get_drvdata(dev); 79 struct amdgpu_device *adev = drm_to_adev(ddev); 80 struct ttm_resource_manager *man; 81 82 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); 83 return sysfs_emit(buf, "%llu\n", amdgpu_gtt_mgr_usage(man)); 84 } 85 86 static DEVICE_ATTR(mem_info_gtt_total, S_IRUGO, 87 amdgpu_mem_info_gtt_total_show, NULL); 88 static DEVICE_ATTR(mem_info_gtt_used, S_IRUGO, 89 amdgpu_mem_info_gtt_used_show, NULL); 90 91 static struct attribute *amdgpu_gtt_mgr_attributes[] = { 92 &dev_attr_mem_info_gtt_total.attr, 93 &dev_attr_mem_info_gtt_used.attr, 94 NULL 95 }; 96 97 const struct attribute_group amdgpu_gtt_mgr_attr_group = { 98 .attrs = amdgpu_gtt_mgr_attributes 99 }; 100 101 /** 102 * amdgpu_gtt_mgr_has_gart_addr - Check if mem has address space 103 * 104 * @res: the mem object to check 105 * 106 * Check if a mem object has already address space allocated. 107 */ 108 bool amdgpu_gtt_mgr_has_gart_addr(struct ttm_resource *res) 109 { 110 struct amdgpu_gtt_node *node = to_amdgpu_gtt_node(res); 111 112 return drm_mm_node_allocated(&node->base.mm_nodes[0]); 113 } 114 115 /** 116 * amdgpu_gtt_mgr_new - allocate a new node 117 * 118 * @man: TTM memory type manager 119 * @tbo: TTM BO we need this range for 120 * @place: placement flags and restrictions 121 * @res: the resulting mem object 122 * 123 * Dummy, allocate the node but no space for it yet. 124 */ 125 static int amdgpu_gtt_mgr_new(struct ttm_resource_manager *man, 126 struct ttm_buffer_object *tbo, 127 const struct ttm_place *place, 128 struct ttm_resource **res) 129 { 130 struct amdgpu_gtt_mgr *mgr = to_gtt_mgr(man); 131 uint32_t num_pages = PFN_UP(tbo->base.size); 132 struct amdgpu_gtt_node *node; 133 int r; 134 135 if (!(place->flags & TTM_PL_FLAG_TEMPORARY) && 136 atomic64_add_return(num_pages, &mgr->used) > man->size) { 137 atomic64_sub(num_pages, &mgr->used); 138 return -ENOSPC; 139 } 140 141 node = kzalloc(struct_size(node, base.mm_nodes, 1), GFP_KERNEL); 142 if (!node) { 143 r = -ENOMEM; 144 goto err_out; 145 } 146 147 node->tbo = tbo; 148 ttm_resource_init(tbo, place, &node->base.base); 149 150 if (place->lpfn) { 151 spin_lock(&mgr->lock); 152 r = drm_mm_insert_node_in_range(&mgr->mm, 153 &node->base.mm_nodes[0], 154 num_pages, tbo->page_alignment, 155 0, place->fpfn, place->lpfn, 156 DRM_MM_INSERT_BEST); 157 spin_unlock(&mgr->lock); 158 if (unlikely(r)) 159 goto err_free; 160 161 node->base.base.start = node->base.mm_nodes[0].start; 162 } else { 163 node->base.mm_nodes[0].start = 0; 164 node->base.mm_nodes[0].size = node->base.base.num_pages; 165 node->base.base.start = AMDGPU_BO_INVALID_OFFSET; 166 } 167 168 *res = &node->base.base; 169 return 0; 170 171 err_free: 172 ttm_resource_fini(man, &node->base.base); 173 kfree(node); 174 175 err_out: 176 if (!(place->flags & TTM_PL_FLAG_TEMPORARY)) 177 atomic64_sub(num_pages, &mgr->used); 178 179 return r; 180 } 181 182 /** 183 * amdgpu_gtt_mgr_del - free ranges 184 * 185 * @man: TTM memory type manager 186 * @res: TTM memory object 187 * 188 * Free the allocated GTT again. 189 */ 190 static void amdgpu_gtt_mgr_del(struct ttm_resource_manager *man, 191 struct ttm_resource *res) 192 { 193 struct amdgpu_gtt_node *node = to_amdgpu_gtt_node(res); 194 struct amdgpu_gtt_mgr *mgr = to_gtt_mgr(man); 195 196 spin_lock(&mgr->lock); 197 if (drm_mm_node_allocated(&node->base.mm_nodes[0])) 198 drm_mm_remove_node(&node->base.mm_nodes[0]); 199 spin_unlock(&mgr->lock); 200 201 if (!(res->placement & TTM_PL_FLAG_TEMPORARY)) 202 atomic64_sub(res->num_pages, &mgr->used); 203 204 ttm_resource_fini(man, res); 205 kfree(node); 206 } 207 208 /** 209 * amdgpu_gtt_mgr_usage - return usage of GTT domain 210 * 211 * @man: TTM memory type manager 212 * 213 * Return how many bytes are used in the GTT domain 214 */ 215 uint64_t amdgpu_gtt_mgr_usage(struct ttm_resource_manager *man) 216 { 217 struct amdgpu_gtt_mgr *mgr = to_gtt_mgr(man); 218 219 return atomic64_read(&mgr->used) * PAGE_SIZE; 220 } 221 222 /** 223 * amdgpu_gtt_mgr_recover - re-init gart 224 * 225 * @man: TTM memory type manager 226 * 227 * Re-init the gart for each known BO in the GTT. 228 */ 229 int amdgpu_gtt_mgr_recover(struct ttm_resource_manager *man) 230 { 231 struct amdgpu_gtt_mgr *mgr = to_gtt_mgr(man); 232 struct amdgpu_device *adev; 233 struct amdgpu_gtt_node *node; 234 struct drm_mm_node *mm_node; 235 int r = 0; 236 237 adev = container_of(mgr, typeof(*adev), mman.gtt_mgr); 238 spin_lock(&mgr->lock); 239 drm_mm_for_each_node(mm_node, &mgr->mm) { 240 node = container_of(mm_node, typeof(*node), base.mm_nodes[0]); 241 r = amdgpu_ttm_recover_gart(node->tbo); 242 if (r) 243 break; 244 } 245 spin_unlock(&mgr->lock); 246 247 amdgpu_gart_invalidate_tlb(adev); 248 249 return r; 250 } 251 252 /** 253 * amdgpu_gtt_mgr_debug - dump VRAM table 254 * 255 * @man: TTM memory type manager 256 * @printer: DRM printer to use 257 * 258 * Dump the table content using printk. 259 */ 260 static void amdgpu_gtt_mgr_debug(struct ttm_resource_manager *man, 261 struct drm_printer *printer) 262 { 263 struct amdgpu_gtt_mgr *mgr = to_gtt_mgr(man); 264 265 spin_lock(&mgr->lock); 266 drm_mm_print(&mgr->mm, printer); 267 spin_unlock(&mgr->lock); 268 269 drm_printf(printer, "man size:%llu pages, gtt used:%llu pages\n", 270 man->size, atomic64_read(&mgr->used)); 271 } 272 273 static const struct ttm_resource_manager_func amdgpu_gtt_mgr_func = { 274 .alloc = amdgpu_gtt_mgr_new, 275 .free = amdgpu_gtt_mgr_del, 276 .debug = amdgpu_gtt_mgr_debug 277 }; 278 279 /** 280 * amdgpu_gtt_mgr_init - init GTT manager and DRM MM 281 * 282 * @adev: amdgpu_device pointer 283 * @gtt_size: maximum size of GTT 284 * 285 * Allocate and initialize the GTT manager. 286 */ 287 int amdgpu_gtt_mgr_init(struct amdgpu_device *adev, uint64_t gtt_size) 288 { 289 struct amdgpu_gtt_mgr *mgr = &adev->mman.gtt_mgr; 290 struct ttm_resource_manager *man = &mgr->manager; 291 uint64_t start, size; 292 293 man->use_tt = true; 294 man->func = &amdgpu_gtt_mgr_func; 295 296 ttm_resource_manager_init(man, gtt_size >> PAGE_SHIFT); 297 298 start = AMDGPU_GTT_MAX_TRANSFER_SIZE * AMDGPU_GTT_NUM_TRANSFER_WINDOWS; 299 size = (adev->gmc.gart_size >> PAGE_SHIFT) - start; 300 drm_mm_init(&mgr->mm, start, size); 301 spin_lock_init(&mgr->lock); 302 atomic64_set(&mgr->used, 0); 303 304 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, &mgr->manager); 305 ttm_resource_manager_set_used(man, true); 306 return 0; 307 } 308 309 /** 310 * amdgpu_gtt_mgr_fini - free and destroy GTT manager 311 * 312 * @adev: amdgpu_device pointer 313 * 314 * Destroy and free the GTT manager, returns -EBUSY if ranges are still 315 * allocated inside it. 316 */ 317 void amdgpu_gtt_mgr_fini(struct amdgpu_device *adev) 318 { 319 struct amdgpu_gtt_mgr *mgr = &adev->mman.gtt_mgr; 320 struct ttm_resource_manager *man = &mgr->manager; 321 int ret; 322 323 ttm_resource_manager_set_used(man, false); 324 325 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); 326 if (ret) 327 return; 328 329 spin_lock(&mgr->lock); 330 drm_mm_takedown(&mgr->mm); 331 spin_unlock(&mgr->lock); 332 333 ttm_resource_manager_cleanup(man); 334 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, NULL); 335 } 336