1 /* 2 * Copyright 2008 Advanced Micro Devices, Inc. 3 * Copyright 2008 Red Hat Inc. 4 * Copyright 2009 Jerome Glisse. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the "Software"), 8 * to deal in the Software without restriction, including without limitation 9 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 10 * and/or sell copies of the Software, and to permit persons to whom the 11 * Software is furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 22 * OTHER DEALINGS IN THE SOFTWARE. 23 * 24 */ 25 26 #include <linux/kthread.h> 27 #include <linux/pci.h> 28 #include <linux/uaccess.h> 29 #include <linux/pm_runtime.h> 30 31 #include "amdgpu.h" 32 #include "amdgpu_pm.h" 33 #include "amdgpu_dm_debugfs.h" 34 #include "amdgpu_ras.h" 35 #include "amdgpu_rap.h" 36 #include "amdgpu_securedisplay.h" 37 #include "amdgpu_fw_attestation.h" 38 #include "amdgpu_umr.h" 39 40 #if defined(CONFIG_DEBUG_FS) 41 42 /** 43 * amdgpu_debugfs_process_reg_op - Handle MMIO register reads/writes 44 * 45 * @read: True if reading 46 * @f: open file handle 47 * @buf: User buffer to write/read to 48 * @size: Number of bytes to write/read 49 * @pos: Offset to seek to 50 * 51 * This debugfs entry has special meaning on the offset being sought. 52 * Various bits have different meanings: 53 * 54 * Bit 62: Indicates a GRBM bank switch is needed 55 * Bit 61: Indicates a SRBM bank switch is needed (implies bit 62 is 56 * zero) 57 * Bits 24..33: The SE or ME selector if needed 58 * Bits 34..43: The SH (or SA) or PIPE selector if needed 59 * Bits 44..53: The INSTANCE (or CU/WGP) or QUEUE selector if needed 60 * 61 * Bit 23: Indicates that the PM power gating lock should be held 62 * This is necessary to read registers that might be 63 * unreliable during a power gating transistion. 64 * 65 * The lower bits are the BYTE offset of the register to read. This 66 * allows reading multiple registers in a single call and having 67 * the returned size reflect that. 68 */ 69 static int amdgpu_debugfs_process_reg_op(bool read, struct file *f, 70 char __user *buf, size_t size, loff_t *pos) 71 { 72 struct amdgpu_device *adev = file_inode(f)->i_private; 73 ssize_t result = 0; 74 int r; 75 bool pm_pg_lock, use_bank, use_ring; 76 unsigned instance_bank, sh_bank, se_bank, me, pipe, queue, vmid; 77 78 pm_pg_lock = use_bank = use_ring = false; 79 instance_bank = sh_bank = se_bank = me = pipe = queue = vmid = 0; 80 81 if (size & 0x3 || *pos & 0x3 || 82 ((*pos & (1ULL << 62)) && (*pos & (1ULL << 61)))) 83 return -EINVAL; 84 85 /* are we reading registers for which a PG lock is necessary? */ 86 pm_pg_lock = (*pos >> 23) & 1; 87 88 if (*pos & (1ULL << 62)) { 89 se_bank = (*pos & GENMASK_ULL(33, 24)) >> 24; 90 sh_bank = (*pos & GENMASK_ULL(43, 34)) >> 34; 91 instance_bank = (*pos & GENMASK_ULL(53, 44)) >> 44; 92 93 if (se_bank == 0x3FF) 94 se_bank = 0xFFFFFFFF; 95 if (sh_bank == 0x3FF) 96 sh_bank = 0xFFFFFFFF; 97 if (instance_bank == 0x3FF) 98 instance_bank = 0xFFFFFFFF; 99 use_bank = true; 100 } else if (*pos & (1ULL << 61)) { 101 102 me = (*pos & GENMASK_ULL(33, 24)) >> 24; 103 pipe = (*pos & GENMASK_ULL(43, 34)) >> 34; 104 queue = (*pos & GENMASK_ULL(53, 44)) >> 44; 105 vmid = (*pos & GENMASK_ULL(58, 54)) >> 54; 106 107 use_ring = true; 108 } else { 109 use_bank = use_ring = false; 110 } 111 112 *pos &= (1UL << 22) - 1; 113 114 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 115 if (r < 0) { 116 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 117 return r; 118 } 119 120 r = amdgpu_virt_enable_access_debugfs(adev); 121 if (r < 0) { 122 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 123 return r; 124 } 125 126 if (use_bank) { 127 if ((sh_bank != 0xFFFFFFFF && sh_bank >= adev->gfx.config.max_sh_per_se) || 128 (se_bank != 0xFFFFFFFF && se_bank >= adev->gfx.config.max_shader_engines)) { 129 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 130 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 131 amdgpu_virt_disable_access_debugfs(adev); 132 return -EINVAL; 133 } 134 mutex_lock(&adev->grbm_idx_mutex); 135 amdgpu_gfx_select_se_sh(adev, se_bank, 136 sh_bank, instance_bank); 137 } else if (use_ring) { 138 mutex_lock(&adev->srbm_mutex); 139 amdgpu_gfx_select_me_pipe_q(adev, me, pipe, queue, vmid); 140 } 141 142 if (pm_pg_lock) 143 mutex_lock(&adev->pm.mutex); 144 145 while (size) { 146 uint32_t value; 147 148 if (read) { 149 value = RREG32(*pos >> 2); 150 r = put_user(value, (uint32_t *)buf); 151 } else { 152 r = get_user(value, (uint32_t *)buf); 153 if (!r) 154 amdgpu_mm_wreg_mmio_rlc(adev, *pos >> 2, value); 155 } 156 if (r) { 157 result = r; 158 goto end; 159 } 160 161 result += 4; 162 buf += 4; 163 *pos += 4; 164 size -= 4; 165 } 166 167 end: 168 if (use_bank) { 169 amdgpu_gfx_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff); 170 mutex_unlock(&adev->grbm_idx_mutex); 171 } else if (use_ring) { 172 amdgpu_gfx_select_me_pipe_q(adev, 0, 0, 0, 0); 173 mutex_unlock(&adev->srbm_mutex); 174 } 175 176 if (pm_pg_lock) 177 mutex_unlock(&adev->pm.mutex); 178 179 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 180 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 181 182 amdgpu_virt_disable_access_debugfs(adev); 183 return result; 184 } 185 186 /* 187 * amdgpu_debugfs_regs_read - Callback for reading MMIO registers 188 */ 189 static ssize_t amdgpu_debugfs_regs_read(struct file *f, char __user *buf, 190 size_t size, loff_t *pos) 191 { 192 return amdgpu_debugfs_process_reg_op(true, f, buf, size, pos); 193 } 194 195 /* 196 * amdgpu_debugfs_regs_write - Callback for writing MMIO registers 197 */ 198 static ssize_t amdgpu_debugfs_regs_write(struct file *f, const char __user *buf, 199 size_t size, loff_t *pos) 200 { 201 return amdgpu_debugfs_process_reg_op(false, f, (char __user *)buf, size, pos); 202 } 203 204 static int amdgpu_debugfs_regs2_open(struct inode *inode, struct file *file) 205 { 206 struct amdgpu_debugfs_regs2_data *rd; 207 208 rd = kzalloc(sizeof *rd, GFP_KERNEL); 209 if (!rd) 210 return -ENOMEM; 211 rd->adev = file_inode(file)->i_private; 212 file->private_data = rd; 213 mutex_init(&rd->lock); 214 215 return 0; 216 } 217 218 static int amdgpu_debugfs_regs2_release(struct inode *inode, struct file *file) 219 { 220 struct amdgpu_debugfs_regs2_data *rd = file->private_data; 221 mutex_destroy(&rd->lock); 222 kfree(file->private_data); 223 return 0; 224 } 225 226 static ssize_t amdgpu_debugfs_regs2_op(struct file *f, char __user *buf, u32 offset, size_t size, int write_en) 227 { 228 struct amdgpu_debugfs_regs2_data *rd = f->private_data; 229 struct amdgpu_device *adev = rd->adev; 230 ssize_t result = 0; 231 int r; 232 uint32_t value; 233 234 if (size & 0x3 || offset & 0x3) 235 return -EINVAL; 236 237 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 238 if (r < 0) { 239 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 240 return r; 241 } 242 243 r = amdgpu_virt_enable_access_debugfs(adev); 244 if (r < 0) { 245 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 246 return r; 247 } 248 249 mutex_lock(&rd->lock); 250 251 if (rd->id.use_grbm) { 252 if ((rd->id.grbm.sh != 0xFFFFFFFF && rd->id.grbm.sh >= adev->gfx.config.max_sh_per_se) || 253 (rd->id.grbm.se != 0xFFFFFFFF && rd->id.grbm.se >= adev->gfx.config.max_shader_engines)) { 254 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 255 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 256 amdgpu_virt_disable_access_debugfs(adev); 257 mutex_unlock(&rd->lock); 258 return -EINVAL; 259 } 260 mutex_lock(&adev->grbm_idx_mutex); 261 amdgpu_gfx_select_se_sh(adev, rd->id.grbm.se, 262 rd->id.grbm.sh, 263 rd->id.grbm.instance); 264 } 265 266 if (rd->id.use_srbm) { 267 mutex_lock(&adev->srbm_mutex); 268 amdgpu_gfx_select_me_pipe_q(adev, rd->id.srbm.me, rd->id.srbm.pipe, 269 rd->id.srbm.queue, rd->id.srbm.vmid); 270 } 271 272 if (rd->id.pg_lock) 273 mutex_lock(&adev->pm.mutex); 274 275 while (size) { 276 if (!write_en) { 277 value = RREG32(offset >> 2); 278 r = put_user(value, (uint32_t *)buf); 279 } else { 280 r = get_user(value, (uint32_t *)buf); 281 if (!r) 282 amdgpu_mm_wreg_mmio_rlc(adev, offset >> 2, value); 283 } 284 if (r) { 285 result = r; 286 goto end; 287 } 288 offset += 4; 289 size -= 4; 290 result += 4; 291 buf += 4; 292 } 293 end: 294 if (rd->id.use_grbm) { 295 amdgpu_gfx_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff); 296 mutex_unlock(&adev->grbm_idx_mutex); 297 } 298 299 if (rd->id.use_srbm) { 300 amdgpu_gfx_select_me_pipe_q(adev, 0, 0, 0, 0); 301 mutex_unlock(&adev->srbm_mutex); 302 } 303 304 if (rd->id.pg_lock) 305 mutex_unlock(&adev->pm.mutex); 306 307 mutex_unlock(&rd->lock); 308 309 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 310 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 311 312 amdgpu_virt_disable_access_debugfs(adev); 313 return result; 314 } 315 316 static long amdgpu_debugfs_regs2_ioctl(struct file *f, unsigned int cmd, unsigned long data) 317 { 318 struct amdgpu_debugfs_regs2_data *rd = f->private_data; 319 int r; 320 321 switch (cmd) { 322 case AMDGPU_DEBUGFS_REGS2_IOC_SET_STATE: 323 mutex_lock(&rd->lock); 324 r = copy_from_user(&rd->id, (struct amdgpu_debugfs_regs2_iocdata *)data, sizeof rd->id); 325 mutex_unlock(&rd->lock); 326 return r ? -EINVAL : 0; 327 default: 328 return -EINVAL; 329 } 330 return 0; 331 } 332 333 static ssize_t amdgpu_debugfs_regs2_read(struct file *f, char __user *buf, size_t size, loff_t *pos) 334 { 335 return amdgpu_debugfs_regs2_op(f, buf, *pos, size, 0); 336 } 337 338 static ssize_t amdgpu_debugfs_regs2_write(struct file *f, const char __user *buf, size_t size, loff_t *pos) 339 { 340 return amdgpu_debugfs_regs2_op(f, (char __user *)buf, *pos, size, 1); 341 } 342 343 344 /** 345 * amdgpu_debugfs_regs_pcie_read - Read from a PCIE register 346 * 347 * @f: open file handle 348 * @buf: User buffer to store read data in 349 * @size: Number of bytes to read 350 * @pos: Offset to seek to 351 * 352 * The lower bits are the BYTE offset of the register to read. This 353 * allows reading multiple registers in a single call and having 354 * the returned size reflect that. 355 */ 356 static ssize_t amdgpu_debugfs_regs_pcie_read(struct file *f, char __user *buf, 357 size_t size, loff_t *pos) 358 { 359 struct amdgpu_device *adev = file_inode(f)->i_private; 360 ssize_t result = 0; 361 int r; 362 363 if (size & 0x3 || *pos & 0x3) 364 return -EINVAL; 365 366 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 367 if (r < 0) { 368 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 369 return r; 370 } 371 372 r = amdgpu_virt_enable_access_debugfs(adev); 373 if (r < 0) { 374 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 375 return r; 376 } 377 378 while (size) { 379 uint32_t value; 380 381 value = RREG32_PCIE(*pos); 382 r = put_user(value, (uint32_t *)buf); 383 if (r) { 384 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 385 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 386 amdgpu_virt_disable_access_debugfs(adev); 387 return r; 388 } 389 390 result += 4; 391 buf += 4; 392 *pos += 4; 393 size -= 4; 394 } 395 396 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 397 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 398 399 amdgpu_virt_disable_access_debugfs(adev); 400 return result; 401 } 402 403 /** 404 * amdgpu_debugfs_regs_pcie_write - Write to a PCIE register 405 * 406 * @f: open file handle 407 * @buf: User buffer to write data from 408 * @size: Number of bytes to write 409 * @pos: Offset to seek to 410 * 411 * The lower bits are the BYTE offset of the register to write. This 412 * allows writing multiple registers in a single call and having 413 * the returned size reflect that. 414 */ 415 static ssize_t amdgpu_debugfs_regs_pcie_write(struct file *f, const char __user *buf, 416 size_t size, loff_t *pos) 417 { 418 struct amdgpu_device *adev = file_inode(f)->i_private; 419 ssize_t result = 0; 420 int r; 421 422 if (size & 0x3 || *pos & 0x3) 423 return -EINVAL; 424 425 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 426 if (r < 0) { 427 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 428 return r; 429 } 430 431 r = amdgpu_virt_enable_access_debugfs(adev); 432 if (r < 0) { 433 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 434 return r; 435 } 436 437 while (size) { 438 uint32_t value; 439 440 r = get_user(value, (uint32_t *)buf); 441 if (r) { 442 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 443 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 444 amdgpu_virt_disable_access_debugfs(adev); 445 return r; 446 } 447 448 WREG32_PCIE(*pos, value); 449 450 result += 4; 451 buf += 4; 452 *pos += 4; 453 size -= 4; 454 } 455 456 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 457 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 458 459 amdgpu_virt_disable_access_debugfs(adev); 460 return result; 461 } 462 463 /** 464 * amdgpu_debugfs_regs_didt_read - Read from a DIDT register 465 * 466 * @f: open file handle 467 * @buf: User buffer to store read data in 468 * @size: Number of bytes to read 469 * @pos: Offset to seek to 470 * 471 * The lower bits are the BYTE offset of the register to read. This 472 * allows reading multiple registers in a single call and having 473 * the returned size reflect that. 474 */ 475 static ssize_t amdgpu_debugfs_regs_didt_read(struct file *f, char __user *buf, 476 size_t size, loff_t *pos) 477 { 478 struct amdgpu_device *adev = file_inode(f)->i_private; 479 ssize_t result = 0; 480 int r; 481 482 if (size & 0x3 || *pos & 0x3) 483 return -EINVAL; 484 485 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 486 if (r < 0) { 487 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 488 return r; 489 } 490 491 r = amdgpu_virt_enable_access_debugfs(adev); 492 if (r < 0) { 493 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 494 return r; 495 } 496 497 while (size) { 498 uint32_t value; 499 500 value = RREG32_DIDT(*pos >> 2); 501 r = put_user(value, (uint32_t *)buf); 502 if (r) { 503 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 504 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 505 amdgpu_virt_disable_access_debugfs(adev); 506 return r; 507 } 508 509 result += 4; 510 buf += 4; 511 *pos += 4; 512 size -= 4; 513 } 514 515 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 516 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 517 518 amdgpu_virt_disable_access_debugfs(adev); 519 return result; 520 } 521 522 /** 523 * amdgpu_debugfs_regs_didt_write - Write to a DIDT register 524 * 525 * @f: open file handle 526 * @buf: User buffer to write data from 527 * @size: Number of bytes to write 528 * @pos: Offset to seek to 529 * 530 * The lower bits are the BYTE offset of the register to write. This 531 * allows writing multiple registers in a single call and having 532 * the returned size reflect that. 533 */ 534 static ssize_t amdgpu_debugfs_regs_didt_write(struct file *f, const char __user *buf, 535 size_t size, loff_t *pos) 536 { 537 struct amdgpu_device *adev = file_inode(f)->i_private; 538 ssize_t result = 0; 539 int r; 540 541 if (size & 0x3 || *pos & 0x3) 542 return -EINVAL; 543 544 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 545 if (r < 0) { 546 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 547 return r; 548 } 549 550 r = amdgpu_virt_enable_access_debugfs(adev); 551 if (r < 0) { 552 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 553 return r; 554 } 555 556 while (size) { 557 uint32_t value; 558 559 r = get_user(value, (uint32_t *)buf); 560 if (r) { 561 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 562 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 563 amdgpu_virt_disable_access_debugfs(adev); 564 return r; 565 } 566 567 WREG32_DIDT(*pos >> 2, value); 568 569 result += 4; 570 buf += 4; 571 *pos += 4; 572 size -= 4; 573 } 574 575 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 576 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 577 578 amdgpu_virt_disable_access_debugfs(adev); 579 return result; 580 } 581 582 /** 583 * amdgpu_debugfs_regs_smc_read - Read from a SMC register 584 * 585 * @f: open file handle 586 * @buf: User buffer to store read data in 587 * @size: Number of bytes to read 588 * @pos: Offset to seek to 589 * 590 * The lower bits are the BYTE offset of the register to read. This 591 * allows reading multiple registers in a single call and having 592 * the returned size reflect that. 593 */ 594 static ssize_t amdgpu_debugfs_regs_smc_read(struct file *f, char __user *buf, 595 size_t size, loff_t *pos) 596 { 597 struct amdgpu_device *adev = file_inode(f)->i_private; 598 ssize_t result = 0; 599 int r; 600 601 if (size & 0x3 || *pos & 0x3) 602 return -EINVAL; 603 604 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 605 if (r < 0) { 606 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 607 return r; 608 } 609 610 r = amdgpu_virt_enable_access_debugfs(adev); 611 if (r < 0) { 612 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 613 return r; 614 } 615 616 while (size) { 617 uint32_t value; 618 619 value = RREG32_SMC(*pos); 620 r = put_user(value, (uint32_t *)buf); 621 if (r) { 622 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 623 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 624 amdgpu_virt_disable_access_debugfs(adev); 625 return r; 626 } 627 628 result += 4; 629 buf += 4; 630 *pos += 4; 631 size -= 4; 632 } 633 634 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 635 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 636 637 amdgpu_virt_disable_access_debugfs(adev); 638 return result; 639 } 640 641 /** 642 * amdgpu_debugfs_regs_smc_write - Write to a SMC register 643 * 644 * @f: open file handle 645 * @buf: User buffer to write data from 646 * @size: Number of bytes to write 647 * @pos: Offset to seek to 648 * 649 * The lower bits are the BYTE offset of the register to write. This 650 * allows writing multiple registers in a single call and having 651 * the returned size reflect that. 652 */ 653 static ssize_t amdgpu_debugfs_regs_smc_write(struct file *f, const char __user *buf, 654 size_t size, loff_t *pos) 655 { 656 struct amdgpu_device *adev = file_inode(f)->i_private; 657 ssize_t result = 0; 658 int r; 659 660 if (size & 0x3 || *pos & 0x3) 661 return -EINVAL; 662 663 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 664 if (r < 0) { 665 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 666 return r; 667 } 668 669 r = amdgpu_virt_enable_access_debugfs(adev); 670 if (r < 0) { 671 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 672 return r; 673 } 674 675 while (size) { 676 uint32_t value; 677 678 r = get_user(value, (uint32_t *)buf); 679 if (r) { 680 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 681 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 682 amdgpu_virt_disable_access_debugfs(adev); 683 return r; 684 } 685 686 WREG32_SMC(*pos, value); 687 688 result += 4; 689 buf += 4; 690 *pos += 4; 691 size -= 4; 692 } 693 694 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 695 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 696 697 amdgpu_virt_disable_access_debugfs(adev); 698 return result; 699 } 700 701 /** 702 * amdgpu_debugfs_gca_config_read - Read from gfx config data 703 * 704 * @f: open file handle 705 * @buf: User buffer to store read data in 706 * @size: Number of bytes to read 707 * @pos: Offset to seek to 708 * 709 * This file is used to access configuration data in a somewhat 710 * stable fashion. The format is a series of DWORDs with the first 711 * indicating which revision it is. New content is appended to the 712 * end so that older software can still read the data. 713 */ 714 715 static ssize_t amdgpu_debugfs_gca_config_read(struct file *f, char __user *buf, 716 size_t size, loff_t *pos) 717 { 718 struct amdgpu_device *adev = file_inode(f)->i_private; 719 ssize_t result = 0; 720 int r; 721 uint32_t *config, no_regs = 0; 722 723 if (size & 0x3 || *pos & 0x3) 724 return -EINVAL; 725 726 config = kmalloc_array(256, sizeof(*config), GFP_KERNEL); 727 if (!config) 728 return -ENOMEM; 729 730 /* version, increment each time something is added */ 731 config[no_regs++] = 4; 732 config[no_regs++] = adev->gfx.config.max_shader_engines; 733 config[no_regs++] = adev->gfx.config.max_tile_pipes; 734 config[no_regs++] = adev->gfx.config.max_cu_per_sh; 735 config[no_regs++] = adev->gfx.config.max_sh_per_se; 736 config[no_regs++] = adev->gfx.config.max_backends_per_se; 737 config[no_regs++] = adev->gfx.config.max_texture_channel_caches; 738 config[no_regs++] = adev->gfx.config.max_gprs; 739 config[no_regs++] = adev->gfx.config.max_gs_threads; 740 config[no_regs++] = adev->gfx.config.max_hw_contexts; 741 config[no_regs++] = adev->gfx.config.sc_prim_fifo_size_frontend; 742 config[no_regs++] = adev->gfx.config.sc_prim_fifo_size_backend; 743 config[no_regs++] = adev->gfx.config.sc_hiz_tile_fifo_size; 744 config[no_regs++] = adev->gfx.config.sc_earlyz_tile_fifo_size; 745 config[no_regs++] = adev->gfx.config.num_tile_pipes; 746 config[no_regs++] = adev->gfx.config.backend_enable_mask; 747 config[no_regs++] = adev->gfx.config.mem_max_burst_length_bytes; 748 config[no_regs++] = adev->gfx.config.mem_row_size_in_kb; 749 config[no_regs++] = adev->gfx.config.shader_engine_tile_size; 750 config[no_regs++] = adev->gfx.config.num_gpus; 751 config[no_regs++] = adev->gfx.config.multi_gpu_tile_size; 752 config[no_regs++] = adev->gfx.config.mc_arb_ramcfg; 753 config[no_regs++] = adev->gfx.config.gb_addr_config; 754 config[no_regs++] = adev->gfx.config.num_rbs; 755 756 /* rev==1 */ 757 config[no_regs++] = adev->rev_id; 758 config[no_regs++] = adev->pg_flags; 759 config[no_regs++] = adev->cg_flags; 760 761 /* rev==2 */ 762 config[no_regs++] = adev->family; 763 config[no_regs++] = adev->external_rev_id; 764 765 /* rev==3 */ 766 config[no_regs++] = adev->pdev->device; 767 config[no_regs++] = adev->pdev->revision; 768 config[no_regs++] = adev->pdev->subsystem_device; 769 config[no_regs++] = adev->pdev->subsystem_vendor; 770 771 /* rev==4 APU flag */ 772 config[no_regs++] = adev->flags & AMD_IS_APU ? 1 : 0; 773 774 while (size && (*pos < no_regs * 4)) { 775 uint32_t value; 776 777 value = config[*pos >> 2]; 778 r = put_user(value, (uint32_t *)buf); 779 if (r) { 780 kfree(config); 781 return r; 782 } 783 784 result += 4; 785 buf += 4; 786 *pos += 4; 787 size -= 4; 788 } 789 790 kfree(config); 791 return result; 792 } 793 794 /** 795 * amdgpu_debugfs_sensor_read - Read from the powerplay sensors 796 * 797 * @f: open file handle 798 * @buf: User buffer to store read data in 799 * @size: Number of bytes to read 800 * @pos: Offset to seek to 801 * 802 * The offset is treated as the BYTE address of one of the sensors 803 * enumerated in amd/include/kgd_pp_interface.h under the 804 * 'amd_pp_sensors' enumeration. For instance to read the UVD VCLK 805 * you would use the offset 3 * 4 = 12. 806 */ 807 static ssize_t amdgpu_debugfs_sensor_read(struct file *f, char __user *buf, 808 size_t size, loff_t *pos) 809 { 810 struct amdgpu_device *adev = file_inode(f)->i_private; 811 int idx, x, outsize, r, valuesize; 812 uint32_t values[16]; 813 814 if (size & 3 || *pos & 0x3) 815 return -EINVAL; 816 817 if (!adev->pm.dpm_enabled) 818 return -EINVAL; 819 820 /* convert offset to sensor number */ 821 idx = *pos >> 2; 822 823 valuesize = sizeof(values); 824 825 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 826 if (r < 0) { 827 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 828 return r; 829 } 830 831 r = amdgpu_virt_enable_access_debugfs(adev); 832 if (r < 0) { 833 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 834 return r; 835 } 836 837 r = amdgpu_dpm_read_sensor(adev, idx, &values[0], &valuesize); 838 839 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 840 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 841 842 if (r) { 843 amdgpu_virt_disable_access_debugfs(adev); 844 return r; 845 } 846 847 if (size > valuesize) { 848 amdgpu_virt_disable_access_debugfs(adev); 849 return -EINVAL; 850 } 851 852 outsize = 0; 853 x = 0; 854 if (!r) { 855 while (size) { 856 r = put_user(values[x++], (int32_t *)buf); 857 buf += 4; 858 size -= 4; 859 outsize += 4; 860 } 861 } 862 863 amdgpu_virt_disable_access_debugfs(adev); 864 return !r ? outsize : r; 865 } 866 867 /** amdgpu_debugfs_wave_read - Read WAVE STATUS data 868 * 869 * @f: open file handle 870 * @buf: User buffer to store read data in 871 * @size: Number of bytes to read 872 * @pos: Offset to seek to 873 * 874 * The offset being sought changes which wave that the status data 875 * will be returned for. The bits are used as follows: 876 * 877 * Bits 0..6: Byte offset into data 878 * Bits 7..14: SE selector 879 * Bits 15..22: SH/SA selector 880 * Bits 23..30: CU/{WGP+SIMD} selector 881 * Bits 31..36: WAVE ID selector 882 * Bits 37..44: SIMD ID selector 883 * 884 * The returned data begins with one DWORD of version information 885 * Followed by WAVE STATUS registers relevant to the GFX IP version 886 * being used. See gfx_v8_0_read_wave_data() for an example output. 887 */ 888 static ssize_t amdgpu_debugfs_wave_read(struct file *f, char __user *buf, 889 size_t size, loff_t *pos) 890 { 891 struct amdgpu_device *adev = f->f_inode->i_private; 892 int r, x; 893 ssize_t result = 0; 894 uint32_t offset, se, sh, cu, wave, simd, data[32]; 895 896 if (size & 3 || *pos & 3) 897 return -EINVAL; 898 899 /* decode offset */ 900 offset = (*pos & GENMASK_ULL(6, 0)); 901 se = (*pos & GENMASK_ULL(14, 7)) >> 7; 902 sh = (*pos & GENMASK_ULL(22, 15)) >> 15; 903 cu = (*pos & GENMASK_ULL(30, 23)) >> 23; 904 wave = (*pos & GENMASK_ULL(36, 31)) >> 31; 905 simd = (*pos & GENMASK_ULL(44, 37)) >> 37; 906 907 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 908 if (r < 0) { 909 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 910 return r; 911 } 912 913 r = amdgpu_virt_enable_access_debugfs(adev); 914 if (r < 0) { 915 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 916 return r; 917 } 918 919 /* switch to the specific se/sh/cu */ 920 mutex_lock(&adev->grbm_idx_mutex); 921 amdgpu_gfx_select_se_sh(adev, se, sh, cu); 922 923 x = 0; 924 if (adev->gfx.funcs->read_wave_data) 925 adev->gfx.funcs->read_wave_data(adev, simd, wave, data, &x); 926 927 amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF); 928 mutex_unlock(&adev->grbm_idx_mutex); 929 930 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 931 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 932 933 if (!x) { 934 amdgpu_virt_disable_access_debugfs(adev); 935 return -EINVAL; 936 } 937 938 while (size && (offset < x * 4)) { 939 uint32_t value; 940 941 value = data[offset >> 2]; 942 r = put_user(value, (uint32_t *)buf); 943 if (r) { 944 amdgpu_virt_disable_access_debugfs(adev); 945 return r; 946 } 947 948 result += 4; 949 buf += 4; 950 offset += 4; 951 size -= 4; 952 } 953 954 amdgpu_virt_disable_access_debugfs(adev); 955 return result; 956 } 957 958 /** amdgpu_debugfs_gpr_read - Read wave gprs 959 * 960 * @f: open file handle 961 * @buf: User buffer to store read data in 962 * @size: Number of bytes to read 963 * @pos: Offset to seek to 964 * 965 * The offset being sought changes which wave that the status data 966 * will be returned for. The bits are used as follows: 967 * 968 * Bits 0..11: Byte offset into data 969 * Bits 12..19: SE selector 970 * Bits 20..27: SH/SA selector 971 * Bits 28..35: CU/{WGP+SIMD} selector 972 * Bits 36..43: WAVE ID selector 973 * Bits 37..44: SIMD ID selector 974 * Bits 52..59: Thread selector 975 * Bits 60..61: Bank selector (VGPR=0,SGPR=1) 976 * 977 * The return data comes from the SGPR or VGPR register bank for 978 * the selected operational unit. 979 */ 980 static ssize_t amdgpu_debugfs_gpr_read(struct file *f, char __user *buf, 981 size_t size, loff_t *pos) 982 { 983 struct amdgpu_device *adev = f->f_inode->i_private; 984 int r; 985 ssize_t result = 0; 986 uint32_t offset, se, sh, cu, wave, simd, thread, bank, *data; 987 988 if (size > 4096 || size & 3 || *pos & 3) 989 return -EINVAL; 990 991 /* decode offset */ 992 offset = (*pos & GENMASK_ULL(11, 0)) >> 2; 993 se = (*pos & GENMASK_ULL(19, 12)) >> 12; 994 sh = (*pos & GENMASK_ULL(27, 20)) >> 20; 995 cu = (*pos & GENMASK_ULL(35, 28)) >> 28; 996 wave = (*pos & GENMASK_ULL(43, 36)) >> 36; 997 simd = (*pos & GENMASK_ULL(51, 44)) >> 44; 998 thread = (*pos & GENMASK_ULL(59, 52)) >> 52; 999 bank = (*pos & GENMASK_ULL(61, 60)) >> 60; 1000 1001 data = kcalloc(1024, sizeof(*data), GFP_KERNEL); 1002 if (!data) 1003 return -ENOMEM; 1004 1005 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 1006 if (r < 0) 1007 goto err; 1008 1009 r = amdgpu_virt_enable_access_debugfs(adev); 1010 if (r < 0) 1011 goto err; 1012 1013 /* switch to the specific se/sh/cu */ 1014 mutex_lock(&adev->grbm_idx_mutex); 1015 amdgpu_gfx_select_se_sh(adev, se, sh, cu); 1016 1017 if (bank == 0) { 1018 if (adev->gfx.funcs->read_wave_vgprs) 1019 adev->gfx.funcs->read_wave_vgprs(adev, simd, wave, thread, offset, size>>2, data); 1020 } else { 1021 if (adev->gfx.funcs->read_wave_sgprs) 1022 adev->gfx.funcs->read_wave_sgprs(adev, simd, wave, offset, size>>2, data); 1023 } 1024 1025 amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF); 1026 mutex_unlock(&adev->grbm_idx_mutex); 1027 1028 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1029 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1030 1031 while (size) { 1032 uint32_t value; 1033 1034 value = data[result >> 2]; 1035 r = put_user(value, (uint32_t *)buf); 1036 if (r) { 1037 amdgpu_virt_disable_access_debugfs(adev); 1038 goto err; 1039 } 1040 1041 result += 4; 1042 buf += 4; 1043 size -= 4; 1044 } 1045 1046 kfree(data); 1047 amdgpu_virt_disable_access_debugfs(adev); 1048 return result; 1049 1050 err: 1051 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1052 kfree(data); 1053 return r; 1054 } 1055 1056 /** 1057 * amdgpu_debugfs_gfxoff_write - Enable/disable GFXOFF 1058 * 1059 * @f: open file handle 1060 * @buf: User buffer to write data from 1061 * @size: Number of bytes to write 1062 * @pos: Offset to seek to 1063 * 1064 * Write a 32-bit zero to disable or a 32-bit non-zero to enable 1065 */ 1066 static ssize_t amdgpu_debugfs_gfxoff_write(struct file *f, const char __user *buf, 1067 size_t size, loff_t *pos) 1068 { 1069 struct amdgpu_device *adev = file_inode(f)->i_private; 1070 ssize_t result = 0; 1071 int r; 1072 1073 if (size & 0x3 || *pos & 0x3) 1074 return -EINVAL; 1075 1076 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 1077 if (r < 0) { 1078 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1079 return r; 1080 } 1081 1082 while (size) { 1083 uint32_t value; 1084 1085 r = get_user(value, (uint32_t *)buf); 1086 if (r) { 1087 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1088 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1089 return r; 1090 } 1091 1092 amdgpu_gfx_off_ctrl(adev, value ? true : false); 1093 1094 result += 4; 1095 buf += 4; 1096 *pos += 4; 1097 size -= 4; 1098 } 1099 1100 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1101 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1102 1103 return result; 1104 } 1105 1106 1107 /** 1108 * amdgpu_debugfs_gfxoff_read - read gfxoff status 1109 * 1110 * @f: open file handle 1111 * @buf: User buffer to store read data in 1112 * @size: Number of bytes to read 1113 * @pos: Offset to seek to 1114 */ 1115 static ssize_t amdgpu_debugfs_gfxoff_read(struct file *f, char __user *buf, 1116 size_t size, loff_t *pos) 1117 { 1118 struct amdgpu_device *adev = file_inode(f)->i_private; 1119 ssize_t result = 0; 1120 int r; 1121 1122 if (size & 0x3 || *pos & 0x3) 1123 return -EINVAL; 1124 1125 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 1126 if (r < 0) { 1127 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1128 return r; 1129 } 1130 1131 while (size) { 1132 uint32_t value; 1133 1134 r = amdgpu_get_gfx_off_status(adev, &value); 1135 if (r) { 1136 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1137 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1138 return r; 1139 } 1140 1141 r = put_user(value, (uint32_t *)buf); 1142 if (r) { 1143 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1144 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1145 return r; 1146 } 1147 1148 result += 4; 1149 buf += 4; 1150 *pos += 4; 1151 size -= 4; 1152 } 1153 1154 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1155 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1156 1157 return result; 1158 } 1159 1160 static const struct file_operations amdgpu_debugfs_regs2_fops = { 1161 .owner = THIS_MODULE, 1162 .unlocked_ioctl = amdgpu_debugfs_regs2_ioctl, 1163 .read = amdgpu_debugfs_regs2_read, 1164 .write = amdgpu_debugfs_regs2_write, 1165 .open = amdgpu_debugfs_regs2_open, 1166 .release = amdgpu_debugfs_regs2_release, 1167 .llseek = default_llseek 1168 }; 1169 1170 static const struct file_operations amdgpu_debugfs_regs_fops = { 1171 .owner = THIS_MODULE, 1172 .read = amdgpu_debugfs_regs_read, 1173 .write = amdgpu_debugfs_regs_write, 1174 .llseek = default_llseek 1175 }; 1176 static const struct file_operations amdgpu_debugfs_regs_didt_fops = { 1177 .owner = THIS_MODULE, 1178 .read = amdgpu_debugfs_regs_didt_read, 1179 .write = amdgpu_debugfs_regs_didt_write, 1180 .llseek = default_llseek 1181 }; 1182 static const struct file_operations amdgpu_debugfs_regs_pcie_fops = { 1183 .owner = THIS_MODULE, 1184 .read = amdgpu_debugfs_regs_pcie_read, 1185 .write = amdgpu_debugfs_regs_pcie_write, 1186 .llseek = default_llseek 1187 }; 1188 static const struct file_operations amdgpu_debugfs_regs_smc_fops = { 1189 .owner = THIS_MODULE, 1190 .read = amdgpu_debugfs_regs_smc_read, 1191 .write = amdgpu_debugfs_regs_smc_write, 1192 .llseek = default_llseek 1193 }; 1194 1195 static const struct file_operations amdgpu_debugfs_gca_config_fops = { 1196 .owner = THIS_MODULE, 1197 .read = amdgpu_debugfs_gca_config_read, 1198 .llseek = default_llseek 1199 }; 1200 1201 static const struct file_operations amdgpu_debugfs_sensors_fops = { 1202 .owner = THIS_MODULE, 1203 .read = amdgpu_debugfs_sensor_read, 1204 .llseek = default_llseek 1205 }; 1206 1207 static const struct file_operations amdgpu_debugfs_wave_fops = { 1208 .owner = THIS_MODULE, 1209 .read = amdgpu_debugfs_wave_read, 1210 .llseek = default_llseek 1211 }; 1212 static const struct file_operations amdgpu_debugfs_gpr_fops = { 1213 .owner = THIS_MODULE, 1214 .read = amdgpu_debugfs_gpr_read, 1215 .llseek = default_llseek 1216 }; 1217 1218 static const struct file_operations amdgpu_debugfs_gfxoff_fops = { 1219 .owner = THIS_MODULE, 1220 .read = amdgpu_debugfs_gfxoff_read, 1221 .write = amdgpu_debugfs_gfxoff_write, 1222 .llseek = default_llseek 1223 }; 1224 1225 static const struct file_operations *debugfs_regs[] = { 1226 &amdgpu_debugfs_regs_fops, 1227 &amdgpu_debugfs_regs2_fops, 1228 &amdgpu_debugfs_regs_didt_fops, 1229 &amdgpu_debugfs_regs_pcie_fops, 1230 &amdgpu_debugfs_regs_smc_fops, 1231 &amdgpu_debugfs_gca_config_fops, 1232 &amdgpu_debugfs_sensors_fops, 1233 &amdgpu_debugfs_wave_fops, 1234 &amdgpu_debugfs_gpr_fops, 1235 &amdgpu_debugfs_gfxoff_fops, 1236 }; 1237 1238 static const char *debugfs_regs_names[] = { 1239 "amdgpu_regs", 1240 "amdgpu_regs2", 1241 "amdgpu_regs_didt", 1242 "amdgpu_regs_pcie", 1243 "amdgpu_regs_smc", 1244 "amdgpu_gca_config", 1245 "amdgpu_sensors", 1246 "amdgpu_wave", 1247 "amdgpu_gpr", 1248 "amdgpu_gfxoff", 1249 }; 1250 1251 /** 1252 * amdgpu_debugfs_regs_init - Initialize debugfs entries that provide 1253 * register access. 1254 * 1255 * @adev: The device to attach the debugfs entries to 1256 */ 1257 int amdgpu_debugfs_regs_init(struct amdgpu_device *adev) 1258 { 1259 struct drm_minor *minor = adev_to_drm(adev)->primary; 1260 struct dentry *ent, *root = minor->debugfs_root; 1261 unsigned int i; 1262 1263 for (i = 0; i < ARRAY_SIZE(debugfs_regs); i++) { 1264 ent = debugfs_create_file(debugfs_regs_names[i], 1265 S_IFREG | S_IRUGO, root, 1266 adev, debugfs_regs[i]); 1267 if (!i && !IS_ERR_OR_NULL(ent)) 1268 i_size_write(ent->d_inode, adev->rmmio_size); 1269 } 1270 1271 return 0; 1272 } 1273 1274 static int amdgpu_debugfs_test_ib_show(struct seq_file *m, void *unused) 1275 { 1276 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 1277 struct drm_device *dev = adev_to_drm(adev); 1278 int r = 0, i; 1279 1280 r = pm_runtime_get_sync(dev->dev); 1281 if (r < 0) { 1282 pm_runtime_put_autosuspend(dev->dev); 1283 return r; 1284 } 1285 1286 /* Avoid accidently unparking the sched thread during GPU reset */ 1287 r = down_write_killable(&adev->reset_sem); 1288 if (r) 1289 return r; 1290 1291 /* hold on the scheduler */ 1292 for (i = 0; i < AMDGPU_MAX_RINGS; i++) { 1293 struct amdgpu_ring *ring = adev->rings[i]; 1294 1295 if (!ring || !ring->sched.thread) 1296 continue; 1297 kthread_park(ring->sched.thread); 1298 } 1299 1300 seq_printf(m, "run ib test:\n"); 1301 r = amdgpu_ib_ring_tests(adev); 1302 if (r) 1303 seq_printf(m, "ib ring tests failed (%d).\n", r); 1304 else 1305 seq_printf(m, "ib ring tests passed.\n"); 1306 1307 /* go on the scheduler */ 1308 for (i = 0; i < AMDGPU_MAX_RINGS; i++) { 1309 struct amdgpu_ring *ring = adev->rings[i]; 1310 1311 if (!ring || !ring->sched.thread) 1312 continue; 1313 kthread_unpark(ring->sched.thread); 1314 } 1315 1316 up_write(&adev->reset_sem); 1317 1318 pm_runtime_mark_last_busy(dev->dev); 1319 pm_runtime_put_autosuspend(dev->dev); 1320 1321 return 0; 1322 } 1323 1324 static int amdgpu_debugfs_evict_vram(void *data, u64 *val) 1325 { 1326 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1327 struct drm_device *dev = adev_to_drm(adev); 1328 int r; 1329 1330 r = pm_runtime_get_sync(dev->dev); 1331 if (r < 0) { 1332 pm_runtime_put_autosuspend(dev->dev); 1333 return r; 1334 } 1335 1336 *val = amdgpu_ttm_evict_resources(adev, TTM_PL_VRAM); 1337 1338 pm_runtime_mark_last_busy(dev->dev); 1339 pm_runtime_put_autosuspend(dev->dev); 1340 1341 return 0; 1342 } 1343 1344 1345 static int amdgpu_debugfs_evict_gtt(void *data, u64 *val) 1346 { 1347 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1348 struct drm_device *dev = adev_to_drm(adev); 1349 int r; 1350 1351 r = pm_runtime_get_sync(dev->dev); 1352 if (r < 0) { 1353 pm_runtime_put_autosuspend(dev->dev); 1354 return r; 1355 } 1356 1357 *val = amdgpu_ttm_evict_resources(adev, TTM_PL_TT); 1358 1359 pm_runtime_mark_last_busy(dev->dev); 1360 pm_runtime_put_autosuspend(dev->dev); 1361 1362 return 0; 1363 } 1364 1365 static int amdgpu_debugfs_benchmark(void *data, u64 val) 1366 { 1367 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1368 struct drm_device *dev = adev_to_drm(adev); 1369 int r; 1370 1371 r = pm_runtime_get_sync(dev->dev); 1372 if (r < 0) { 1373 pm_runtime_put_autosuspend(dev->dev); 1374 return r; 1375 } 1376 1377 r = amdgpu_benchmark(adev, val); 1378 1379 pm_runtime_mark_last_busy(dev->dev); 1380 pm_runtime_put_autosuspend(dev->dev); 1381 1382 return r; 1383 } 1384 1385 static int amdgpu_debugfs_vm_info_show(struct seq_file *m, void *unused) 1386 { 1387 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 1388 struct drm_device *dev = adev_to_drm(adev); 1389 struct drm_file *file; 1390 int r; 1391 1392 r = mutex_lock_interruptible(&dev->filelist_mutex); 1393 if (r) 1394 return r; 1395 1396 list_for_each_entry(file, &dev->filelist, lhead) { 1397 struct amdgpu_fpriv *fpriv = file->driver_priv; 1398 struct amdgpu_vm *vm = &fpriv->vm; 1399 1400 seq_printf(m, "pid:%d\tProcess:%s ----------\n", 1401 vm->task_info.pid, vm->task_info.process_name); 1402 r = amdgpu_bo_reserve(vm->root.bo, true); 1403 if (r) 1404 break; 1405 amdgpu_debugfs_vm_bo_info(vm, m); 1406 amdgpu_bo_unreserve(vm->root.bo); 1407 } 1408 1409 mutex_unlock(&dev->filelist_mutex); 1410 1411 return r; 1412 } 1413 1414 DEFINE_SHOW_ATTRIBUTE(amdgpu_debugfs_test_ib); 1415 DEFINE_SHOW_ATTRIBUTE(amdgpu_debugfs_vm_info); 1416 DEFINE_DEBUGFS_ATTRIBUTE(amdgpu_evict_vram_fops, amdgpu_debugfs_evict_vram, 1417 NULL, "%lld\n"); 1418 DEFINE_DEBUGFS_ATTRIBUTE(amdgpu_evict_gtt_fops, amdgpu_debugfs_evict_gtt, 1419 NULL, "%lld\n"); 1420 DEFINE_DEBUGFS_ATTRIBUTE(amdgpu_benchmark_fops, NULL, amdgpu_debugfs_benchmark, 1421 "%lld\n"); 1422 1423 static void amdgpu_ib_preempt_fences_swap(struct amdgpu_ring *ring, 1424 struct dma_fence **fences) 1425 { 1426 struct amdgpu_fence_driver *drv = &ring->fence_drv; 1427 uint32_t sync_seq, last_seq; 1428 1429 last_seq = atomic_read(&ring->fence_drv.last_seq); 1430 sync_seq = ring->fence_drv.sync_seq; 1431 1432 last_seq &= drv->num_fences_mask; 1433 sync_seq &= drv->num_fences_mask; 1434 1435 do { 1436 struct dma_fence *fence, **ptr; 1437 1438 ++last_seq; 1439 last_seq &= drv->num_fences_mask; 1440 ptr = &drv->fences[last_seq]; 1441 1442 fence = rcu_dereference_protected(*ptr, 1); 1443 RCU_INIT_POINTER(*ptr, NULL); 1444 1445 if (!fence) 1446 continue; 1447 1448 fences[last_seq] = fence; 1449 1450 } while (last_seq != sync_seq); 1451 } 1452 1453 static void amdgpu_ib_preempt_signal_fences(struct dma_fence **fences, 1454 int length) 1455 { 1456 int i; 1457 struct dma_fence *fence; 1458 1459 for (i = 0; i < length; i++) { 1460 fence = fences[i]; 1461 if (!fence) 1462 continue; 1463 dma_fence_signal(fence); 1464 dma_fence_put(fence); 1465 } 1466 } 1467 1468 static void amdgpu_ib_preempt_job_recovery(struct drm_gpu_scheduler *sched) 1469 { 1470 struct drm_sched_job *s_job; 1471 struct dma_fence *fence; 1472 1473 spin_lock(&sched->job_list_lock); 1474 list_for_each_entry(s_job, &sched->pending_list, list) { 1475 fence = sched->ops->run_job(s_job); 1476 dma_fence_put(fence); 1477 } 1478 spin_unlock(&sched->job_list_lock); 1479 } 1480 1481 static void amdgpu_ib_preempt_mark_partial_job(struct amdgpu_ring *ring) 1482 { 1483 struct amdgpu_job *job; 1484 struct drm_sched_job *s_job, *tmp; 1485 uint32_t preempt_seq; 1486 struct dma_fence *fence, **ptr; 1487 struct amdgpu_fence_driver *drv = &ring->fence_drv; 1488 struct drm_gpu_scheduler *sched = &ring->sched; 1489 bool preempted = true; 1490 1491 if (ring->funcs->type != AMDGPU_RING_TYPE_GFX) 1492 return; 1493 1494 preempt_seq = le32_to_cpu(*(drv->cpu_addr + 2)); 1495 if (preempt_seq <= atomic_read(&drv->last_seq)) { 1496 preempted = false; 1497 goto no_preempt; 1498 } 1499 1500 preempt_seq &= drv->num_fences_mask; 1501 ptr = &drv->fences[preempt_seq]; 1502 fence = rcu_dereference_protected(*ptr, 1); 1503 1504 no_preempt: 1505 spin_lock(&sched->job_list_lock); 1506 list_for_each_entry_safe(s_job, tmp, &sched->pending_list, list) { 1507 if (dma_fence_is_signaled(&s_job->s_fence->finished)) { 1508 /* remove job from ring_mirror_list */ 1509 list_del_init(&s_job->list); 1510 sched->ops->free_job(s_job); 1511 continue; 1512 } 1513 job = to_amdgpu_job(s_job); 1514 if (preempted && (&job->hw_fence) == fence) 1515 /* mark the job as preempted */ 1516 job->preemption_status |= AMDGPU_IB_PREEMPTED; 1517 } 1518 spin_unlock(&sched->job_list_lock); 1519 } 1520 1521 static int amdgpu_debugfs_ib_preempt(void *data, u64 val) 1522 { 1523 int r, resched, length; 1524 struct amdgpu_ring *ring; 1525 struct dma_fence **fences = NULL; 1526 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1527 1528 if (val >= AMDGPU_MAX_RINGS) 1529 return -EINVAL; 1530 1531 ring = adev->rings[val]; 1532 1533 if (!ring || !ring->funcs->preempt_ib || !ring->sched.thread) 1534 return -EINVAL; 1535 1536 /* the last preemption failed */ 1537 if (ring->trail_seq != le32_to_cpu(*ring->trail_fence_cpu_addr)) 1538 return -EBUSY; 1539 1540 length = ring->fence_drv.num_fences_mask + 1; 1541 fences = kcalloc(length, sizeof(void *), GFP_KERNEL); 1542 if (!fences) 1543 return -ENOMEM; 1544 1545 /* Avoid accidently unparking the sched thread during GPU reset */ 1546 r = down_read_killable(&adev->reset_sem); 1547 if (r) 1548 goto pro_end; 1549 1550 /* stop the scheduler */ 1551 kthread_park(ring->sched.thread); 1552 1553 resched = ttm_bo_lock_delayed_workqueue(&adev->mman.bdev); 1554 1555 /* preempt the IB */ 1556 r = amdgpu_ring_preempt_ib(ring); 1557 if (r) { 1558 DRM_WARN("failed to preempt ring %d\n", ring->idx); 1559 goto failure; 1560 } 1561 1562 amdgpu_fence_process(ring); 1563 1564 if (atomic_read(&ring->fence_drv.last_seq) != 1565 ring->fence_drv.sync_seq) { 1566 DRM_INFO("ring %d was preempted\n", ring->idx); 1567 1568 amdgpu_ib_preempt_mark_partial_job(ring); 1569 1570 /* swap out the old fences */ 1571 amdgpu_ib_preempt_fences_swap(ring, fences); 1572 1573 amdgpu_fence_driver_force_completion(ring); 1574 1575 /* resubmit unfinished jobs */ 1576 amdgpu_ib_preempt_job_recovery(&ring->sched); 1577 1578 /* wait for jobs finished */ 1579 amdgpu_fence_wait_empty(ring); 1580 1581 /* signal the old fences */ 1582 amdgpu_ib_preempt_signal_fences(fences, length); 1583 } 1584 1585 failure: 1586 /* restart the scheduler */ 1587 kthread_unpark(ring->sched.thread); 1588 1589 up_read(&adev->reset_sem); 1590 1591 ttm_bo_unlock_delayed_workqueue(&adev->mman.bdev, resched); 1592 1593 pro_end: 1594 kfree(fences); 1595 1596 return r; 1597 } 1598 1599 static int amdgpu_debugfs_sclk_set(void *data, u64 val) 1600 { 1601 int ret = 0; 1602 uint32_t max_freq, min_freq; 1603 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1604 1605 if (amdgpu_sriov_vf(adev) && !amdgpu_sriov_is_pp_one_vf(adev)) 1606 return -EINVAL; 1607 1608 ret = pm_runtime_get_sync(adev_to_drm(adev)->dev); 1609 if (ret < 0) { 1610 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1611 return ret; 1612 } 1613 1614 ret = amdgpu_dpm_get_dpm_freq_range(adev, PP_SCLK, &min_freq, &max_freq); 1615 if (ret == -EOPNOTSUPP) { 1616 ret = 0; 1617 goto out; 1618 } 1619 if (ret || val > max_freq || val < min_freq) { 1620 ret = -EINVAL; 1621 goto out; 1622 } 1623 1624 ret = amdgpu_dpm_set_soft_freq_range(adev, PP_SCLK, (uint32_t)val, (uint32_t)val); 1625 if (ret) 1626 ret = -EINVAL; 1627 1628 out: 1629 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1630 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1631 1632 return ret; 1633 } 1634 1635 DEFINE_DEBUGFS_ATTRIBUTE(fops_ib_preempt, NULL, 1636 amdgpu_debugfs_ib_preempt, "%llu\n"); 1637 1638 DEFINE_DEBUGFS_ATTRIBUTE(fops_sclk_set, NULL, 1639 amdgpu_debugfs_sclk_set, "%llu\n"); 1640 1641 int amdgpu_debugfs_init(struct amdgpu_device *adev) 1642 { 1643 struct dentry *root = adev_to_drm(adev)->primary->debugfs_root; 1644 struct dentry *ent; 1645 int r, i; 1646 1647 if (!debugfs_initialized()) 1648 return 0; 1649 1650 debugfs_create_x32("amdgpu_smu_debug", 0600, root, 1651 &adev->pm.smu_debug_mask); 1652 1653 ent = debugfs_create_file("amdgpu_preempt_ib", 0600, root, adev, 1654 &fops_ib_preempt); 1655 if (IS_ERR(ent)) { 1656 DRM_ERROR("unable to create amdgpu_preempt_ib debugsfs file\n"); 1657 return PTR_ERR(ent); 1658 } 1659 1660 ent = debugfs_create_file("amdgpu_force_sclk", 0200, root, adev, 1661 &fops_sclk_set); 1662 if (IS_ERR(ent)) { 1663 DRM_ERROR("unable to create amdgpu_set_sclk debugsfs file\n"); 1664 return PTR_ERR(ent); 1665 } 1666 1667 /* Register debugfs entries for amdgpu_ttm */ 1668 amdgpu_ttm_debugfs_init(adev); 1669 amdgpu_debugfs_pm_init(adev); 1670 amdgpu_debugfs_sa_init(adev); 1671 amdgpu_debugfs_fence_init(adev); 1672 amdgpu_debugfs_gem_init(adev); 1673 1674 r = amdgpu_debugfs_regs_init(adev); 1675 if (r) 1676 DRM_ERROR("registering register debugfs failed (%d).\n", r); 1677 1678 amdgpu_debugfs_firmware_init(adev); 1679 1680 #if defined(CONFIG_DRM_AMD_DC) 1681 if (amdgpu_device_has_dc_support(adev)) 1682 dtn_debugfs_init(adev); 1683 #endif 1684 1685 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 1686 struct amdgpu_ring *ring = adev->rings[i]; 1687 1688 if (!ring) 1689 continue; 1690 1691 amdgpu_debugfs_ring_init(adev, ring); 1692 } 1693 1694 amdgpu_ras_debugfs_create_all(adev); 1695 amdgpu_rap_debugfs_init(adev); 1696 amdgpu_securedisplay_debugfs_init(adev); 1697 amdgpu_fw_attestation_debugfs_init(adev); 1698 1699 debugfs_create_file("amdgpu_evict_vram", 0444, root, adev, 1700 &amdgpu_evict_vram_fops); 1701 debugfs_create_file("amdgpu_evict_gtt", 0444, root, adev, 1702 &amdgpu_evict_gtt_fops); 1703 debugfs_create_file("amdgpu_test_ib", 0444, root, adev, 1704 &amdgpu_debugfs_test_ib_fops); 1705 debugfs_create_file("amdgpu_vm_info", 0444, root, adev, 1706 &amdgpu_debugfs_vm_info_fops); 1707 debugfs_create_file("amdgpu_benchmark", 0200, root, adev, 1708 &amdgpu_benchmark_fops); 1709 1710 adev->debugfs_vbios_blob.data = adev->bios; 1711 adev->debugfs_vbios_blob.size = adev->bios_size; 1712 debugfs_create_blob("amdgpu_vbios", 0444, root, 1713 &adev->debugfs_vbios_blob); 1714 1715 adev->debugfs_discovery_blob.data = adev->mman.discovery_bin; 1716 adev->debugfs_discovery_blob.size = adev->mman.discovery_tmr_size; 1717 debugfs_create_blob("amdgpu_discovery", 0444, root, 1718 &adev->debugfs_discovery_blob); 1719 1720 return 0; 1721 } 1722 1723 #else 1724 int amdgpu_debugfs_init(struct amdgpu_device *adev) 1725 { 1726 return 0; 1727 } 1728 int amdgpu_debugfs_regs_init(struct amdgpu_device *adev) 1729 { 1730 return 0; 1731 } 1732 #endif 1733