1 /* 2 * Copyright 2008 Advanced Micro Devices, Inc. 3 * Copyright 2008 Red Hat Inc. 4 * Copyright 2009 Jerome Glisse. 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a 7 * copy of this software and associated documentation files (the "Software"), 8 * to deal in the Software without restriction, including without limitation 9 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 10 * and/or sell copies of the Software, and to permit persons to whom the 11 * Software is furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 22 * OTHER DEALINGS IN THE SOFTWARE. 23 * 24 */ 25 26 #include <linux/kthread.h> 27 #include <linux/pci.h> 28 #include <linux/uaccess.h> 29 #include <linux/pm_runtime.h> 30 31 #include "amdgpu.h" 32 #include "amdgpu_pm.h" 33 #include "amdgpu_dm_debugfs.h" 34 #include "amdgpu_ras.h" 35 #include "amdgpu_rap.h" 36 #include "amdgpu_securedisplay.h" 37 #include "amdgpu_fw_attestation.h" 38 39 #if defined(CONFIG_DEBUG_FS) 40 41 /** 42 * amdgpu_debugfs_process_reg_op - Handle MMIO register reads/writes 43 * 44 * @read: True if reading 45 * @f: open file handle 46 * @buf: User buffer to write/read to 47 * @size: Number of bytes to write/read 48 * @pos: Offset to seek to 49 * 50 * This debugfs entry has special meaning on the offset being sought. 51 * Various bits have different meanings: 52 * 53 * Bit 62: Indicates a GRBM bank switch is needed 54 * Bit 61: Indicates a SRBM bank switch is needed (implies bit 62 is 55 * zero) 56 * Bits 24..33: The SE or ME selector if needed 57 * Bits 34..43: The SH (or SA) or PIPE selector if needed 58 * Bits 44..53: The INSTANCE (or CU/WGP) or QUEUE selector if needed 59 * 60 * Bit 23: Indicates that the PM power gating lock should be held 61 * This is necessary to read registers that might be 62 * unreliable during a power gating transistion. 63 * 64 * The lower bits are the BYTE offset of the register to read. This 65 * allows reading multiple registers in a single call and having 66 * the returned size reflect that. 67 */ 68 static int amdgpu_debugfs_process_reg_op(bool read, struct file *f, 69 char __user *buf, size_t size, loff_t *pos) 70 { 71 struct amdgpu_device *adev = file_inode(f)->i_private; 72 ssize_t result = 0; 73 int r; 74 bool pm_pg_lock, use_bank, use_ring; 75 unsigned instance_bank, sh_bank, se_bank, me, pipe, queue, vmid; 76 77 pm_pg_lock = use_bank = use_ring = false; 78 instance_bank = sh_bank = se_bank = me = pipe = queue = vmid = 0; 79 80 if (size & 0x3 || *pos & 0x3 || 81 ((*pos & (1ULL << 62)) && (*pos & (1ULL << 61)))) 82 return -EINVAL; 83 84 /* are we reading registers for which a PG lock is necessary? */ 85 pm_pg_lock = (*pos >> 23) & 1; 86 87 if (*pos & (1ULL << 62)) { 88 se_bank = (*pos & GENMASK_ULL(33, 24)) >> 24; 89 sh_bank = (*pos & GENMASK_ULL(43, 34)) >> 34; 90 instance_bank = (*pos & GENMASK_ULL(53, 44)) >> 44; 91 92 if (se_bank == 0x3FF) 93 se_bank = 0xFFFFFFFF; 94 if (sh_bank == 0x3FF) 95 sh_bank = 0xFFFFFFFF; 96 if (instance_bank == 0x3FF) 97 instance_bank = 0xFFFFFFFF; 98 use_bank = true; 99 } else if (*pos & (1ULL << 61)) { 100 101 me = (*pos & GENMASK_ULL(33, 24)) >> 24; 102 pipe = (*pos & GENMASK_ULL(43, 34)) >> 34; 103 queue = (*pos & GENMASK_ULL(53, 44)) >> 44; 104 vmid = (*pos & GENMASK_ULL(58, 54)) >> 54; 105 106 use_ring = true; 107 } else { 108 use_bank = use_ring = false; 109 } 110 111 *pos &= (1UL << 22) - 1; 112 113 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 114 if (r < 0) { 115 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 116 return r; 117 } 118 119 r = amdgpu_virt_enable_access_debugfs(adev); 120 if (r < 0) { 121 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 122 return r; 123 } 124 125 if (use_bank) { 126 if ((sh_bank != 0xFFFFFFFF && sh_bank >= adev->gfx.config.max_sh_per_se) || 127 (se_bank != 0xFFFFFFFF && se_bank >= adev->gfx.config.max_shader_engines)) { 128 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 129 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 130 amdgpu_virt_disable_access_debugfs(adev); 131 return -EINVAL; 132 } 133 mutex_lock(&adev->grbm_idx_mutex); 134 amdgpu_gfx_select_se_sh(adev, se_bank, 135 sh_bank, instance_bank); 136 } else if (use_ring) { 137 mutex_lock(&adev->srbm_mutex); 138 amdgpu_gfx_select_me_pipe_q(adev, me, pipe, queue, vmid); 139 } 140 141 if (pm_pg_lock) 142 mutex_lock(&adev->pm.mutex); 143 144 while (size) { 145 uint32_t value; 146 147 if (read) { 148 value = RREG32(*pos >> 2); 149 r = put_user(value, (uint32_t *)buf); 150 } else { 151 r = get_user(value, (uint32_t *)buf); 152 if (!r) 153 amdgpu_mm_wreg_mmio_rlc(adev, *pos >> 2, value); 154 } 155 if (r) { 156 result = r; 157 goto end; 158 } 159 160 result += 4; 161 buf += 4; 162 *pos += 4; 163 size -= 4; 164 } 165 166 end: 167 if (use_bank) { 168 amdgpu_gfx_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff); 169 mutex_unlock(&adev->grbm_idx_mutex); 170 } else if (use_ring) { 171 amdgpu_gfx_select_me_pipe_q(adev, 0, 0, 0, 0); 172 mutex_unlock(&adev->srbm_mutex); 173 } 174 175 if (pm_pg_lock) 176 mutex_unlock(&adev->pm.mutex); 177 178 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 179 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 180 181 amdgpu_virt_disable_access_debugfs(adev); 182 return result; 183 } 184 185 /* 186 * amdgpu_debugfs_regs_read - Callback for reading MMIO registers 187 */ 188 static ssize_t amdgpu_debugfs_regs_read(struct file *f, char __user *buf, 189 size_t size, loff_t *pos) 190 { 191 return amdgpu_debugfs_process_reg_op(true, f, buf, size, pos); 192 } 193 194 /* 195 * amdgpu_debugfs_regs_write - Callback for writing MMIO registers 196 */ 197 static ssize_t amdgpu_debugfs_regs_write(struct file *f, const char __user *buf, 198 size_t size, loff_t *pos) 199 { 200 return amdgpu_debugfs_process_reg_op(false, f, (char __user *)buf, size, pos); 201 } 202 203 204 /** 205 * amdgpu_debugfs_regs_pcie_read - Read from a PCIE register 206 * 207 * @f: open file handle 208 * @buf: User buffer to store read data in 209 * @size: Number of bytes to read 210 * @pos: Offset to seek to 211 * 212 * The lower bits are the BYTE offset of the register to read. This 213 * allows reading multiple registers in a single call and having 214 * the returned size reflect that. 215 */ 216 static ssize_t amdgpu_debugfs_regs_pcie_read(struct file *f, char __user *buf, 217 size_t size, loff_t *pos) 218 { 219 struct amdgpu_device *adev = file_inode(f)->i_private; 220 ssize_t result = 0; 221 int r; 222 223 if (size & 0x3 || *pos & 0x3) 224 return -EINVAL; 225 226 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 227 if (r < 0) { 228 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 229 return r; 230 } 231 232 r = amdgpu_virt_enable_access_debugfs(adev); 233 if (r < 0) { 234 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 235 return r; 236 } 237 238 while (size) { 239 uint32_t value; 240 241 value = RREG32_PCIE(*pos); 242 r = put_user(value, (uint32_t *)buf); 243 if (r) { 244 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 245 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 246 amdgpu_virt_disable_access_debugfs(adev); 247 return r; 248 } 249 250 result += 4; 251 buf += 4; 252 *pos += 4; 253 size -= 4; 254 } 255 256 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 257 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 258 259 amdgpu_virt_disable_access_debugfs(adev); 260 return result; 261 } 262 263 /** 264 * amdgpu_debugfs_regs_pcie_write - Write to a PCIE register 265 * 266 * @f: open file handle 267 * @buf: User buffer to write data from 268 * @size: Number of bytes to write 269 * @pos: Offset to seek to 270 * 271 * The lower bits are the BYTE offset of the register to write. This 272 * allows writing multiple registers in a single call and having 273 * the returned size reflect that. 274 */ 275 static ssize_t amdgpu_debugfs_regs_pcie_write(struct file *f, const char __user *buf, 276 size_t size, loff_t *pos) 277 { 278 struct amdgpu_device *adev = file_inode(f)->i_private; 279 ssize_t result = 0; 280 int r; 281 282 if (size & 0x3 || *pos & 0x3) 283 return -EINVAL; 284 285 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 286 if (r < 0) { 287 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 288 return r; 289 } 290 291 r = amdgpu_virt_enable_access_debugfs(adev); 292 if (r < 0) { 293 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 294 return r; 295 } 296 297 while (size) { 298 uint32_t value; 299 300 r = get_user(value, (uint32_t *)buf); 301 if (r) { 302 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 303 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 304 amdgpu_virt_disable_access_debugfs(adev); 305 return r; 306 } 307 308 WREG32_PCIE(*pos, value); 309 310 result += 4; 311 buf += 4; 312 *pos += 4; 313 size -= 4; 314 } 315 316 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 317 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 318 319 amdgpu_virt_disable_access_debugfs(adev); 320 return result; 321 } 322 323 /** 324 * amdgpu_debugfs_regs_didt_read - Read from a DIDT register 325 * 326 * @f: open file handle 327 * @buf: User buffer to store read data in 328 * @size: Number of bytes to read 329 * @pos: Offset to seek to 330 * 331 * The lower bits are the BYTE offset of the register to read. This 332 * allows reading multiple registers in a single call and having 333 * the returned size reflect that. 334 */ 335 static ssize_t amdgpu_debugfs_regs_didt_read(struct file *f, char __user *buf, 336 size_t size, loff_t *pos) 337 { 338 struct amdgpu_device *adev = file_inode(f)->i_private; 339 ssize_t result = 0; 340 int r; 341 342 if (size & 0x3 || *pos & 0x3) 343 return -EINVAL; 344 345 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 346 if (r < 0) { 347 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 348 return r; 349 } 350 351 r = amdgpu_virt_enable_access_debugfs(adev); 352 if (r < 0) { 353 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 354 return r; 355 } 356 357 while (size) { 358 uint32_t value; 359 360 value = RREG32_DIDT(*pos >> 2); 361 r = put_user(value, (uint32_t *)buf); 362 if (r) { 363 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 364 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 365 amdgpu_virt_disable_access_debugfs(adev); 366 return r; 367 } 368 369 result += 4; 370 buf += 4; 371 *pos += 4; 372 size -= 4; 373 } 374 375 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 376 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 377 378 amdgpu_virt_disable_access_debugfs(adev); 379 return result; 380 } 381 382 /** 383 * amdgpu_debugfs_regs_didt_write - Write to a DIDT register 384 * 385 * @f: open file handle 386 * @buf: User buffer to write data from 387 * @size: Number of bytes to write 388 * @pos: Offset to seek to 389 * 390 * The lower bits are the BYTE offset of the register to write. This 391 * allows writing multiple registers in a single call and having 392 * the returned size reflect that. 393 */ 394 static ssize_t amdgpu_debugfs_regs_didt_write(struct file *f, const char __user *buf, 395 size_t size, loff_t *pos) 396 { 397 struct amdgpu_device *adev = file_inode(f)->i_private; 398 ssize_t result = 0; 399 int r; 400 401 if (size & 0x3 || *pos & 0x3) 402 return -EINVAL; 403 404 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 405 if (r < 0) { 406 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 407 return r; 408 } 409 410 r = amdgpu_virt_enable_access_debugfs(adev); 411 if (r < 0) { 412 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 413 return r; 414 } 415 416 while (size) { 417 uint32_t value; 418 419 r = get_user(value, (uint32_t *)buf); 420 if (r) { 421 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 422 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 423 amdgpu_virt_disable_access_debugfs(adev); 424 return r; 425 } 426 427 WREG32_DIDT(*pos >> 2, value); 428 429 result += 4; 430 buf += 4; 431 *pos += 4; 432 size -= 4; 433 } 434 435 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 436 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 437 438 amdgpu_virt_disable_access_debugfs(adev); 439 return result; 440 } 441 442 /** 443 * amdgpu_debugfs_regs_smc_read - Read from a SMC register 444 * 445 * @f: open file handle 446 * @buf: User buffer to store read data in 447 * @size: Number of bytes to read 448 * @pos: Offset to seek to 449 * 450 * The lower bits are the BYTE offset of the register to read. This 451 * allows reading multiple registers in a single call and having 452 * the returned size reflect that. 453 */ 454 static ssize_t amdgpu_debugfs_regs_smc_read(struct file *f, char __user *buf, 455 size_t size, loff_t *pos) 456 { 457 struct amdgpu_device *adev = file_inode(f)->i_private; 458 ssize_t result = 0; 459 int r; 460 461 if (size & 0x3 || *pos & 0x3) 462 return -EINVAL; 463 464 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 465 if (r < 0) { 466 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 467 return r; 468 } 469 470 r = amdgpu_virt_enable_access_debugfs(adev); 471 if (r < 0) { 472 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 473 return r; 474 } 475 476 while (size) { 477 uint32_t value; 478 479 value = RREG32_SMC(*pos); 480 r = put_user(value, (uint32_t *)buf); 481 if (r) { 482 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 483 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 484 amdgpu_virt_disable_access_debugfs(adev); 485 return r; 486 } 487 488 result += 4; 489 buf += 4; 490 *pos += 4; 491 size -= 4; 492 } 493 494 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 495 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 496 497 amdgpu_virt_disable_access_debugfs(adev); 498 return result; 499 } 500 501 /** 502 * amdgpu_debugfs_regs_smc_write - Write to a SMC register 503 * 504 * @f: open file handle 505 * @buf: User buffer to write data from 506 * @size: Number of bytes to write 507 * @pos: Offset to seek to 508 * 509 * The lower bits are the BYTE offset of the register to write. This 510 * allows writing multiple registers in a single call and having 511 * the returned size reflect that. 512 */ 513 static ssize_t amdgpu_debugfs_regs_smc_write(struct file *f, const char __user *buf, 514 size_t size, loff_t *pos) 515 { 516 struct amdgpu_device *adev = file_inode(f)->i_private; 517 ssize_t result = 0; 518 int r; 519 520 if (size & 0x3 || *pos & 0x3) 521 return -EINVAL; 522 523 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 524 if (r < 0) { 525 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 526 return r; 527 } 528 529 r = amdgpu_virt_enable_access_debugfs(adev); 530 if (r < 0) { 531 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 532 return r; 533 } 534 535 while (size) { 536 uint32_t value; 537 538 r = get_user(value, (uint32_t *)buf); 539 if (r) { 540 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 541 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 542 amdgpu_virt_disable_access_debugfs(adev); 543 return r; 544 } 545 546 WREG32_SMC(*pos, value); 547 548 result += 4; 549 buf += 4; 550 *pos += 4; 551 size -= 4; 552 } 553 554 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 555 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 556 557 amdgpu_virt_disable_access_debugfs(adev); 558 return result; 559 } 560 561 /** 562 * amdgpu_debugfs_gca_config_read - Read from gfx config data 563 * 564 * @f: open file handle 565 * @buf: User buffer to store read data in 566 * @size: Number of bytes to read 567 * @pos: Offset to seek to 568 * 569 * This file is used to access configuration data in a somewhat 570 * stable fashion. The format is a series of DWORDs with the first 571 * indicating which revision it is. New content is appended to the 572 * end so that older software can still read the data. 573 */ 574 575 static ssize_t amdgpu_debugfs_gca_config_read(struct file *f, char __user *buf, 576 size_t size, loff_t *pos) 577 { 578 struct amdgpu_device *adev = file_inode(f)->i_private; 579 ssize_t result = 0; 580 int r; 581 uint32_t *config, no_regs = 0; 582 583 if (size & 0x3 || *pos & 0x3) 584 return -EINVAL; 585 586 config = kmalloc_array(256, sizeof(*config), GFP_KERNEL); 587 if (!config) 588 return -ENOMEM; 589 590 /* version, increment each time something is added */ 591 config[no_regs++] = 3; 592 config[no_regs++] = adev->gfx.config.max_shader_engines; 593 config[no_regs++] = adev->gfx.config.max_tile_pipes; 594 config[no_regs++] = adev->gfx.config.max_cu_per_sh; 595 config[no_regs++] = adev->gfx.config.max_sh_per_se; 596 config[no_regs++] = adev->gfx.config.max_backends_per_se; 597 config[no_regs++] = adev->gfx.config.max_texture_channel_caches; 598 config[no_regs++] = adev->gfx.config.max_gprs; 599 config[no_regs++] = adev->gfx.config.max_gs_threads; 600 config[no_regs++] = adev->gfx.config.max_hw_contexts; 601 config[no_regs++] = adev->gfx.config.sc_prim_fifo_size_frontend; 602 config[no_regs++] = adev->gfx.config.sc_prim_fifo_size_backend; 603 config[no_regs++] = adev->gfx.config.sc_hiz_tile_fifo_size; 604 config[no_regs++] = adev->gfx.config.sc_earlyz_tile_fifo_size; 605 config[no_regs++] = adev->gfx.config.num_tile_pipes; 606 config[no_regs++] = adev->gfx.config.backend_enable_mask; 607 config[no_regs++] = adev->gfx.config.mem_max_burst_length_bytes; 608 config[no_regs++] = adev->gfx.config.mem_row_size_in_kb; 609 config[no_regs++] = adev->gfx.config.shader_engine_tile_size; 610 config[no_regs++] = adev->gfx.config.num_gpus; 611 config[no_regs++] = adev->gfx.config.multi_gpu_tile_size; 612 config[no_regs++] = adev->gfx.config.mc_arb_ramcfg; 613 config[no_regs++] = adev->gfx.config.gb_addr_config; 614 config[no_regs++] = adev->gfx.config.num_rbs; 615 616 /* rev==1 */ 617 config[no_regs++] = adev->rev_id; 618 config[no_regs++] = adev->pg_flags; 619 config[no_regs++] = adev->cg_flags; 620 621 /* rev==2 */ 622 config[no_regs++] = adev->family; 623 config[no_regs++] = adev->external_rev_id; 624 625 /* rev==3 */ 626 config[no_regs++] = adev->pdev->device; 627 config[no_regs++] = adev->pdev->revision; 628 config[no_regs++] = adev->pdev->subsystem_device; 629 config[no_regs++] = adev->pdev->subsystem_vendor; 630 631 while (size && (*pos < no_regs * 4)) { 632 uint32_t value; 633 634 value = config[*pos >> 2]; 635 r = put_user(value, (uint32_t *)buf); 636 if (r) { 637 kfree(config); 638 return r; 639 } 640 641 result += 4; 642 buf += 4; 643 *pos += 4; 644 size -= 4; 645 } 646 647 kfree(config); 648 return result; 649 } 650 651 /** 652 * amdgpu_debugfs_sensor_read - Read from the powerplay sensors 653 * 654 * @f: open file handle 655 * @buf: User buffer to store read data in 656 * @size: Number of bytes to read 657 * @pos: Offset to seek to 658 * 659 * The offset is treated as the BYTE address of one of the sensors 660 * enumerated in amd/include/kgd_pp_interface.h under the 661 * 'amd_pp_sensors' enumeration. For instance to read the UVD VCLK 662 * you would use the offset 3 * 4 = 12. 663 */ 664 static ssize_t amdgpu_debugfs_sensor_read(struct file *f, char __user *buf, 665 size_t size, loff_t *pos) 666 { 667 struct amdgpu_device *adev = file_inode(f)->i_private; 668 int idx, x, outsize, r, valuesize; 669 uint32_t values[16]; 670 671 if (size & 3 || *pos & 0x3) 672 return -EINVAL; 673 674 if (!adev->pm.dpm_enabled) 675 return -EINVAL; 676 677 /* convert offset to sensor number */ 678 idx = *pos >> 2; 679 680 valuesize = sizeof(values); 681 682 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 683 if (r < 0) { 684 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 685 return r; 686 } 687 688 r = amdgpu_virt_enable_access_debugfs(adev); 689 if (r < 0) { 690 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 691 return r; 692 } 693 694 r = amdgpu_dpm_read_sensor(adev, idx, &values[0], &valuesize); 695 696 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 697 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 698 699 if (r) { 700 amdgpu_virt_disable_access_debugfs(adev); 701 return r; 702 } 703 704 if (size > valuesize) { 705 amdgpu_virt_disable_access_debugfs(adev); 706 return -EINVAL; 707 } 708 709 outsize = 0; 710 x = 0; 711 if (!r) { 712 while (size) { 713 r = put_user(values[x++], (int32_t *)buf); 714 buf += 4; 715 size -= 4; 716 outsize += 4; 717 } 718 } 719 720 amdgpu_virt_disable_access_debugfs(adev); 721 return !r ? outsize : r; 722 } 723 724 /** amdgpu_debugfs_wave_read - Read WAVE STATUS data 725 * 726 * @f: open file handle 727 * @buf: User buffer to store read data in 728 * @size: Number of bytes to read 729 * @pos: Offset to seek to 730 * 731 * The offset being sought changes which wave that the status data 732 * will be returned for. The bits are used as follows: 733 * 734 * Bits 0..6: Byte offset into data 735 * Bits 7..14: SE selector 736 * Bits 15..22: SH/SA selector 737 * Bits 23..30: CU/{WGP+SIMD} selector 738 * Bits 31..36: WAVE ID selector 739 * Bits 37..44: SIMD ID selector 740 * 741 * The returned data begins with one DWORD of version information 742 * Followed by WAVE STATUS registers relevant to the GFX IP version 743 * being used. See gfx_v8_0_read_wave_data() for an example output. 744 */ 745 static ssize_t amdgpu_debugfs_wave_read(struct file *f, char __user *buf, 746 size_t size, loff_t *pos) 747 { 748 struct amdgpu_device *adev = f->f_inode->i_private; 749 int r, x; 750 ssize_t result = 0; 751 uint32_t offset, se, sh, cu, wave, simd, data[32]; 752 753 if (size & 3 || *pos & 3) 754 return -EINVAL; 755 756 /* decode offset */ 757 offset = (*pos & GENMASK_ULL(6, 0)); 758 se = (*pos & GENMASK_ULL(14, 7)) >> 7; 759 sh = (*pos & GENMASK_ULL(22, 15)) >> 15; 760 cu = (*pos & GENMASK_ULL(30, 23)) >> 23; 761 wave = (*pos & GENMASK_ULL(36, 31)) >> 31; 762 simd = (*pos & GENMASK_ULL(44, 37)) >> 37; 763 764 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 765 if (r < 0) { 766 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 767 return r; 768 } 769 770 r = amdgpu_virt_enable_access_debugfs(adev); 771 if (r < 0) { 772 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 773 return r; 774 } 775 776 /* switch to the specific se/sh/cu */ 777 mutex_lock(&adev->grbm_idx_mutex); 778 amdgpu_gfx_select_se_sh(adev, se, sh, cu); 779 780 x = 0; 781 if (adev->gfx.funcs->read_wave_data) 782 adev->gfx.funcs->read_wave_data(adev, simd, wave, data, &x); 783 784 amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF); 785 mutex_unlock(&adev->grbm_idx_mutex); 786 787 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 788 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 789 790 if (!x) { 791 amdgpu_virt_disable_access_debugfs(adev); 792 return -EINVAL; 793 } 794 795 while (size && (offset < x * 4)) { 796 uint32_t value; 797 798 value = data[offset >> 2]; 799 r = put_user(value, (uint32_t *)buf); 800 if (r) { 801 amdgpu_virt_disable_access_debugfs(adev); 802 return r; 803 } 804 805 result += 4; 806 buf += 4; 807 offset += 4; 808 size -= 4; 809 } 810 811 amdgpu_virt_disable_access_debugfs(adev); 812 return result; 813 } 814 815 /** amdgpu_debugfs_gpr_read - Read wave gprs 816 * 817 * @f: open file handle 818 * @buf: User buffer to store read data in 819 * @size: Number of bytes to read 820 * @pos: Offset to seek to 821 * 822 * The offset being sought changes which wave that the status data 823 * will be returned for. The bits are used as follows: 824 * 825 * Bits 0..11: Byte offset into data 826 * Bits 12..19: SE selector 827 * Bits 20..27: SH/SA selector 828 * Bits 28..35: CU/{WGP+SIMD} selector 829 * Bits 36..43: WAVE ID selector 830 * Bits 37..44: SIMD ID selector 831 * Bits 52..59: Thread selector 832 * Bits 60..61: Bank selector (VGPR=0,SGPR=1) 833 * 834 * The return data comes from the SGPR or VGPR register bank for 835 * the selected operational unit. 836 */ 837 static ssize_t amdgpu_debugfs_gpr_read(struct file *f, char __user *buf, 838 size_t size, loff_t *pos) 839 { 840 struct amdgpu_device *adev = f->f_inode->i_private; 841 int r; 842 ssize_t result = 0; 843 uint32_t offset, se, sh, cu, wave, simd, thread, bank, *data; 844 845 if (size > 4096 || size & 3 || *pos & 3) 846 return -EINVAL; 847 848 /* decode offset */ 849 offset = (*pos & GENMASK_ULL(11, 0)) >> 2; 850 se = (*pos & GENMASK_ULL(19, 12)) >> 12; 851 sh = (*pos & GENMASK_ULL(27, 20)) >> 20; 852 cu = (*pos & GENMASK_ULL(35, 28)) >> 28; 853 wave = (*pos & GENMASK_ULL(43, 36)) >> 36; 854 simd = (*pos & GENMASK_ULL(51, 44)) >> 44; 855 thread = (*pos & GENMASK_ULL(59, 52)) >> 52; 856 bank = (*pos & GENMASK_ULL(61, 60)) >> 60; 857 858 data = kcalloc(1024, sizeof(*data), GFP_KERNEL); 859 if (!data) 860 return -ENOMEM; 861 862 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 863 if (r < 0) 864 goto err; 865 866 r = amdgpu_virt_enable_access_debugfs(adev); 867 if (r < 0) 868 goto err; 869 870 /* switch to the specific se/sh/cu */ 871 mutex_lock(&adev->grbm_idx_mutex); 872 amdgpu_gfx_select_se_sh(adev, se, sh, cu); 873 874 if (bank == 0) { 875 if (adev->gfx.funcs->read_wave_vgprs) 876 adev->gfx.funcs->read_wave_vgprs(adev, simd, wave, thread, offset, size>>2, data); 877 } else { 878 if (adev->gfx.funcs->read_wave_sgprs) 879 adev->gfx.funcs->read_wave_sgprs(adev, simd, wave, offset, size>>2, data); 880 } 881 882 amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF); 883 mutex_unlock(&adev->grbm_idx_mutex); 884 885 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 886 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 887 888 while (size) { 889 uint32_t value; 890 891 value = data[result >> 2]; 892 r = put_user(value, (uint32_t *)buf); 893 if (r) { 894 amdgpu_virt_disable_access_debugfs(adev); 895 goto err; 896 } 897 898 result += 4; 899 buf += 4; 900 size -= 4; 901 } 902 903 kfree(data); 904 amdgpu_virt_disable_access_debugfs(adev); 905 return result; 906 907 err: 908 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 909 kfree(data); 910 return r; 911 } 912 913 /** 914 * amdgpu_debugfs_gfxoff_write - Enable/disable GFXOFF 915 * 916 * @f: open file handle 917 * @buf: User buffer to write data from 918 * @size: Number of bytes to write 919 * @pos: Offset to seek to 920 * 921 * Write a 32-bit zero to disable or a 32-bit non-zero to enable 922 */ 923 static ssize_t amdgpu_debugfs_gfxoff_write(struct file *f, const char __user *buf, 924 size_t size, loff_t *pos) 925 { 926 struct amdgpu_device *adev = file_inode(f)->i_private; 927 ssize_t result = 0; 928 int r; 929 930 if (size & 0x3 || *pos & 0x3) 931 return -EINVAL; 932 933 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 934 if (r < 0) { 935 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 936 return r; 937 } 938 939 while (size) { 940 uint32_t value; 941 942 r = get_user(value, (uint32_t *)buf); 943 if (r) { 944 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 945 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 946 return r; 947 } 948 949 amdgpu_gfx_off_ctrl(adev, value ? true : false); 950 951 result += 4; 952 buf += 4; 953 *pos += 4; 954 size -= 4; 955 } 956 957 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 958 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 959 960 return result; 961 } 962 963 964 /** 965 * amdgpu_debugfs_gfxoff_read - read gfxoff status 966 * 967 * @f: open file handle 968 * @buf: User buffer to store read data in 969 * @size: Number of bytes to read 970 * @pos: Offset to seek to 971 */ 972 static ssize_t amdgpu_debugfs_gfxoff_read(struct file *f, char __user *buf, 973 size_t size, loff_t *pos) 974 { 975 struct amdgpu_device *adev = file_inode(f)->i_private; 976 ssize_t result = 0; 977 int r; 978 979 if (size & 0x3 || *pos & 0x3) 980 return -EINVAL; 981 982 r = pm_runtime_get_sync(adev_to_drm(adev)->dev); 983 if (r < 0) 984 return r; 985 986 while (size) { 987 uint32_t value; 988 989 r = amdgpu_get_gfx_off_status(adev, &value); 990 if (r) { 991 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 992 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 993 return r; 994 } 995 996 r = put_user(value, (uint32_t *)buf); 997 if (r) { 998 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 999 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1000 return r; 1001 } 1002 1003 result += 4; 1004 buf += 4; 1005 *pos += 4; 1006 size -= 4; 1007 } 1008 1009 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1010 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1011 1012 return result; 1013 } 1014 1015 static const struct file_operations amdgpu_debugfs_regs_fops = { 1016 .owner = THIS_MODULE, 1017 .read = amdgpu_debugfs_regs_read, 1018 .write = amdgpu_debugfs_regs_write, 1019 .llseek = default_llseek 1020 }; 1021 static const struct file_operations amdgpu_debugfs_regs_didt_fops = { 1022 .owner = THIS_MODULE, 1023 .read = amdgpu_debugfs_regs_didt_read, 1024 .write = amdgpu_debugfs_regs_didt_write, 1025 .llseek = default_llseek 1026 }; 1027 static const struct file_operations amdgpu_debugfs_regs_pcie_fops = { 1028 .owner = THIS_MODULE, 1029 .read = amdgpu_debugfs_regs_pcie_read, 1030 .write = amdgpu_debugfs_regs_pcie_write, 1031 .llseek = default_llseek 1032 }; 1033 static const struct file_operations amdgpu_debugfs_regs_smc_fops = { 1034 .owner = THIS_MODULE, 1035 .read = amdgpu_debugfs_regs_smc_read, 1036 .write = amdgpu_debugfs_regs_smc_write, 1037 .llseek = default_llseek 1038 }; 1039 1040 static const struct file_operations amdgpu_debugfs_gca_config_fops = { 1041 .owner = THIS_MODULE, 1042 .read = amdgpu_debugfs_gca_config_read, 1043 .llseek = default_llseek 1044 }; 1045 1046 static const struct file_operations amdgpu_debugfs_sensors_fops = { 1047 .owner = THIS_MODULE, 1048 .read = amdgpu_debugfs_sensor_read, 1049 .llseek = default_llseek 1050 }; 1051 1052 static const struct file_operations amdgpu_debugfs_wave_fops = { 1053 .owner = THIS_MODULE, 1054 .read = amdgpu_debugfs_wave_read, 1055 .llseek = default_llseek 1056 }; 1057 static const struct file_operations amdgpu_debugfs_gpr_fops = { 1058 .owner = THIS_MODULE, 1059 .read = amdgpu_debugfs_gpr_read, 1060 .llseek = default_llseek 1061 }; 1062 1063 static const struct file_operations amdgpu_debugfs_gfxoff_fops = { 1064 .owner = THIS_MODULE, 1065 .read = amdgpu_debugfs_gfxoff_read, 1066 .write = amdgpu_debugfs_gfxoff_write, 1067 .llseek = default_llseek 1068 }; 1069 1070 static const struct file_operations *debugfs_regs[] = { 1071 &amdgpu_debugfs_regs_fops, 1072 &amdgpu_debugfs_regs_didt_fops, 1073 &amdgpu_debugfs_regs_pcie_fops, 1074 &amdgpu_debugfs_regs_smc_fops, 1075 &amdgpu_debugfs_gca_config_fops, 1076 &amdgpu_debugfs_sensors_fops, 1077 &amdgpu_debugfs_wave_fops, 1078 &amdgpu_debugfs_gpr_fops, 1079 &amdgpu_debugfs_gfxoff_fops, 1080 }; 1081 1082 static const char *debugfs_regs_names[] = { 1083 "amdgpu_regs", 1084 "amdgpu_regs_didt", 1085 "amdgpu_regs_pcie", 1086 "amdgpu_regs_smc", 1087 "amdgpu_gca_config", 1088 "amdgpu_sensors", 1089 "amdgpu_wave", 1090 "amdgpu_gpr", 1091 "amdgpu_gfxoff", 1092 }; 1093 1094 /** 1095 * amdgpu_debugfs_regs_init - Initialize debugfs entries that provide 1096 * register access. 1097 * 1098 * @adev: The device to attach the debugfs entries to 1099 */ 1100 int amdgpu_debugfs_regs_init(struct amdgpu_device *adev) 1101 { 1102 struct drm_minor *minor = adev_to_drm(adev)->primary; 1103 struct dentry *ent, *root = minor->debugfs_root; 1104 unsigned int i; 1105 1106 for (i = 0; i < ARRAY_SIZE(debugfs_regs); i++) { 1107 ent = debugfs_create_file(debugfs_regs_names[i], 1108 S_IFREG | S_IRUGO, root, 1109 adev, debugfs_regs[i]); 1110 if (!i && !IS_ERR_OR_NULL(ent)) 1111 i_size_write(ent->d_inode, adev->rmmio_size); 1112 } 1113 1114 return 0; 1115 } 1116 1117 static int amdgpu_debugfs_test_ib_show(struct seq_file *m, void *unused) 1118 { 1119 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 1120 struct drm_device *dev = adev_to_drm(adev); 1121 int r = 0, i; 1122 1123 r = pm_runtime_get_sync(dev->dev); 1124 if (r < 0) { 1125 pm_runtime_put_autosuspend(dev->dev); 1126 return r; 1127 } 1128 1129 /* Avoid accidently unparking the sched thread during GPU reset */ 1130 r = down_read_killable(&adev->reset_sem); 1131 if (r) 1132 return r; 1133 1134 /* hold on the scheduler */ 1135 for (i = 0; i < AMDGPU_MAX_RINGS; i++) { 1136 struct amdgpu_ring *ring = adev->rings[i]; 1137 1138 if (!ring || !ring->sched.thread) 1139 continue; 1140 kthread_park(ring->sched.thread); 1141 } 1142 1143 seq_printf(m, "run ib test:\n"); 1144 r = amdgpu_ib_ring_tests(adev); 1145 if (r) 1146 seq_printf(m, "ib ring tests failed (%d).\n", r); 1147 else 1148 seq_printf(m, "ib ring tests passed.\n"); 1149 1150 /* go on the scheduler */ 1151 for (i = 0; i < AMDGPU_MAX_RINGS; i++) { 1152 struct amdgpu_ring *ring = adev->rings[i]; 1153 1154 if (!ring || !ring->sched.thread) 1155 continue; 1156 kthread_unpark(ring->sched.thread); 1157 } 1158 1159 up_read(&adev->reset_sem); 1160 1161 pm_runtime_mark_last_busy(dev->dev); 1162 pm_runtime_put_autosuspend(dev->dev); 1163 1164 return 0; 1165 } 1166 1167 static int amdgpu_debugfs_evict_vram(void *data, u64 *val) 1168 { 1169 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1170 struct drm_device *dev = adev_to_drm(adev); 1171 int r; 1172 1173 r = pm_runtime_get_sync(dev->dev); 1174 if (r < 0) { 1175 pm_runtime_put_autosuspend(dev->dev); 1176 return r; 1177 } 1178 1179 *val = amdgpu_ttm_evict_resources(adev, TTM_PL_VRAM); 1180 1181 pm_runtime_mark_last_busy(dev->dev); 1182 pm_runtime_put_autosuspend(dev->dev); 1183 1184 return 0; 1185 } 1186 1187 1188 static int amdgpu_debugfs_evict_gtt(void *data, u64 *val) 1189 { 1190 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1191 struct drm_device *dev = adev_to_drm(adev); 1192 int r; 1193 1194 r = pm_runtime_get_sync(dev->dev); 1195 if (r < 0) { 1196 pm_runtime_put_autosuspend(dev->dev); 1197 return r; 1198 } 1199 1200 *val = amdgpu_ttm_evict_resources(adev, TTM_PL_TT); 1201 1202 pm_runtime_mark_last_busy(dev->dev); 1203 pm_runtime_put_autosuspend(dev->dev); 1204 1205 return 0; 1206 } 1207 1208 1209 static int amdgpu_debugfs_vm_info_show(struct seq_file *m, void *unused) 1210 { 1211 struct amdgpu_device *adev = (struct amdgpu_device *)m->private; 1212 struct drm_device *dev = adev_to_drm(adev); 1213 struct drm_file *file; 1214 int r; 1215 1216 r = mutex_lock_interruptible(&dev->filelist_mutex); 1217 if (r) 1218 return r; 1219 1220 list_for_each_entry(file, &dev->filelist, lhead) { 1221 struct amdgpu_fpriv *fpriv = file->driver_priv; 1222 struct amdgpu_vm *vm = &fpriv->vm; 1223 1224 seq_printf(m, "pid:%d\tProcess:%s ----------\n", 1225 vm->task_info.pid, vm->task_info.process_name); 1226 r = amdgpu_bo_reserve(vm->root.bo, true); 1227 if (r) 1228 break; 1229 amdgpu_debugfs_vm_bo_info(vm, m); 1230 amdgpu_bo_unreserve(vm->root.bo); 1231 } 1232 1233 mutex_unlock(&dev->filelist_mutex); 1234 1235 return r; 1236 } 1237 1238 DEFINE_SHOW_ATTRIBUTE(amdgpu_debugfs_test_ib); 1239 DEFINE_SHOW_ATTRIBUTE(amdgpu_debugfs_vm_info); 1240 DEFINE_DEBUGFS_ATTRIBUTE(amdgpu_evict_vram_fops, amdgpu_debugfs_evict_vram, 1241 NULL, "%lld\n"); 1242 DEFINE_DEBUGFS_ATTRIBUTE(amdgpu_evict_gtt_fops, amdgpu_debugfs_evict_gtt, 1243 NULL, "%lld\n"); 1244 1245 static void amdgpu_ib_preempt_fences_swap(struct amdgpu_ring *ring, 1246 struct dma_fence **fences) 1247 { 1248 struct amdgpu_fence_driver *drv = &ring->fence_drv; 1249 uint32_t sync_seq, last_seq; 1250 1251 last_seq = atomic_read(&ring->fence_drv.last_seq); 1252 sync_seq = ring->fence_drv.sync_seq; 1253 1254 last_seq &= drv->num_fences_mask; 1255 sync_seq &= drv->num_fences_mask; 1256 1257 do { 1258 struct dma_fence *fence, **ptr; 1259 1260 ++last_seq; 1261 last_seq &= drv->num_fences_mask; 1262 ptr = &drv->fences[last_seq]; 1263 1264 fence = rcu_dereference_protected(*ptr, 1); 1265 RCU_INIT_POINTER(*ptr, NULL); 1266 1267 if (!fence) 1268 continue; 1269 1270 fences[last_seq] = fence; 1271 1272 } while (last_seq != sync_seq); 1273 } 1274 1275 static void amdgpu_ib_preempt_signal_fences(struct dma_fence **fences, 1276 int length) 1277 { 1278 int i; 1279 struct dma_fence *fence; 1280 1281 for (i = 0; i < length; i++) { 1282 fence = fences[i]; 1283 if (!fence) 1284 continue; 1285 dma_fence_signal(fence); 1286 dma_fence_put(fence); 1287 } 1288 } 1289 1290 static void amdgpu_ib_preempt_job_recovery(struct drm_gpu_scheduler *sched) 1291 { 1292 struct drm_sched_job *s_job; 1293 struct dma_fence *fence; 1294 1295 spin_lock(&sched->job_list_lock); 1296 list_for_each_entry(s_job, &sched->pending_list, list) { 1297 fence = sched->ops->run_job(s_job); 1298 dma_fence_put(fence); 1299 } 1300 spin_unlock(&sched->job_list_lock); 1301 } 1302 1303 static void amdgpu_ib_preempt_mark_partial_job(struct amdgpu_ring *ring) 1304 { 1305 struct amdgpu_job *job; 1306 struct drm_sched_job *s_job, *tmp; 1307 uint32_t preempt_seq; 1308 struct dma_fence *fence, **ptr; 1309 struct amdgpu_fence_driver *drv = &ring->fence_drv; 1310 struct drm_gpu_scheduler *sched = &ring->sched; 1311 bool preempted = true; 1312 1313 if (ring->funcs->type != AMDGPU_RING_TYPE_GFX) 1314 return; 1315 1316 preempt_seq = le32_to_cpu(*(drv->cpu_addr + 2)); 1317 if (preempt_seq <= atomic_read(&drv->last_seq)) { 1318 preempted = false; 1319 goto no_preempt; 1320 } 1321 1322 preempt_seq &= drv->num_fences_mask; 1323 ptr = &drv->fences[preempt_seq]; 1324 fence = rcu_dereference_protected(*ptr, 1); 1325 1326 no_preempt: 1327 spin_lock(&sched->job_list_lock); 1328 list_for_each_entry_safe(s_job, tmp, &sched->pending_list, list) { 1329 if (dma_fence_is_signaled(&s_job->s_fence->finished)) { 1330 /* remove job from ring_mirror_list */ 1331 list_del_init(&s_job->list); 1332 sched->ops->free_job(s_job); 1333 continue; 1334 } 1335 job = to_amdgpu_job(s_job); 1336 if (preempted && (&job->hw_fence) == fence) 1337 /* mark the job as preempted */ 1338 job->preemption_status |= AMDGPU_IB_PREEMPTED; 1339 } 1340 spin_unlock(&sched->job_list_lock); 1341 } 1342 1343 static int amdgpu_debugfs_ib_preempt(void *data, u64 val) 1344 { 1345 int r, resched, length; 1346 struct amdgpu_ring *ring; 1347 struct dma_fence **fences = NULL; 1348 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1349 1350 if (val >= AMDGPU_MAX_RINGS) 1351 return -EINVAL; 1352 1353 ring = adev->rings[val]; 1354 1355 if (!ring || !ring->funcs->preempt_ib || !ring->sched.thread) 1356 return -EINVAL; 1357 1358 /* the last preemption failed */ 1359 if (ring->trail_seq != le32_to_cpu(*ring->trail_fence_cpu_addr)) 1360 return -EBUSY; 1361 1362 length = ring->fence_drv.num_fences_mask + 1; 1363 fences = kcalloc(length, sizeof(void *), GFP_KERNEL); 1364 if (!fences) 1365 return -ENOMEM; 1366 1367 /* Avoid accidently unparking the sched thread during GPU reset */ 1368 r = down_read_killable(&adev->reset_sem); 1369 if (r) 1370 goto pro_end; 1371 1372 /* stop the scheduler */ 1373 kthread_park(ring->sched.thread); 1374 1375 resched = ttm_bo_lock_delayed_workqueue(&adev->mman.bdev); 1376 1377 /* preempt the IB */ 1378 r = amdgpu_ring_preempt_ib(ring); 1379 if (r) { 1380 DRM_WARN("failed to preempt ring %d\n", ring->idx); 1381 goto failure; 1382 } 1383 1384 amdgpu_fence_process(ring); 1385 1386 if (atomic_read(&ring->fence_drv.last_seq) != 1387 ring->fence_drv.sync_seq) { 1388 DRM_INFO("ring %d was preempted\n", ring->idx); 1389 1390 amdgpu_ib_preempt_mark_partial_job(ring); 1391 1392 /* swap out the old fences */ 1393 amdgpu_ib_preempt_fences_swap(ring, fences); 1394 1395 amdgpu_fence_driver_force_completion(ring); 1396 1397 /* resubmit unfinished jobs */ 1398 amdgpu_ib_preempt_job_recovery(&ring->sched); 1399 1400 /* wait for jobs finished */ 1401 amdgpu_fence_wait_empty(ring); 1402 1403 /* signal the old fences */ 1404 amdgpu_ib_preempt_signal_fences(fences, length); 1405 } 1406 1407 failure: 1408 /* restart the scheduler */ 1409 kthread_unpark(ring->sched.thread); 1410 1411 up_read(&adev->reset_sem); 1412 1413 ttm_bo_unlock_delayed_workqueue(&adev->mman.bdev, resched); 1414 1415 pro_end: 1416 kfree(fences); 1417 1418 return r; 1419 } 1420 1421 static int amdgpu_debugfs_sclk_set(void *data, u64 val) 1422 { 1423 int ret = 0; 1424 uint32_t max_freq, min_freq; 1425 struct amdgpu_device *adev = (struct amdgpu_device *)data; 1426 1427 if (amdgpu_sriov_vf(adev) && !amdgpu_sriov_is_pp_one_vf(adev)) 1428 return -EINVAL; 1429 1430 ret = pm_runtime_get_sync(adev_to_drm(adev)->dev); 1431 if (ret < 0) { 1432 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1433 return ret; 1434 } 1435 1436 if (is_support_sw_smu(adev)) { 1437 ret = smu_get_dpm_freq_range(&adev->smu, SMU_SCLK, &min_freq, &max_freq); 1438 if (ret || val > max_freq || val < min_freq) 1439 return -EINVAL; 1440 ret = smu_set_soft_freq_range(&adev->smu, SMU_SCLK, (uint32_t)val, (uint32_t)val); 1441 } else { 1442 return 0; 1443 } 1444 1445 pm_runtime_mark_last_busy(adev_to_drm(adev)->dev); 1446 pm_runtime_put_autosuspend(adev_to_drm(adev)->dev); 1447 1448 if (ret) 1449 return -EINVAL; 1450 1451 return 0; 1452 } 1453 1454 DEFINE_DEBUGFS_ATTRIBUTE(fops_ib_preempt, NULL, 1455 amdgpu_debugfs_ib_preempt, "%llu\n"); 1456 1457 DEFINE_DEBUGFS_ATTRIBUTE(fops_sclk_set, NULL, 1458 amdgpu_debugfs_sclk_set, "%llu\n"); 1459 1460 int amdgpu_debugfs_init(struct amdgpu_device *adev) 1461 { 1462 struct dentry *root = adev_to_drm(adev)->primary->debugfs_root; 1463 struct dentry *ent; 1464 int r, i; 1465 1466 ent = debugfs_create_file("amdgpu_preempt_ib", 0600, root, adev, 1467 &fops_ib_preempt); 1468 if (IS_ERR(ent)) { 1469 DRM_ERROR("unable to create amdgpu_preempt_ib debugsfs file\n"); 1470 return PTR_ERR(ent); 1471 } 1472 1473 ent = debugfs_create_file("amdgpu_force_sclk", 0200, root, adev, 1474 &fops_sclk_set); 1475 if (IS_ERR(ent)) { 1476 DRM_ERROR("unable to create amdgpu_set_sclk debugsfs file\n"); 1477 return PTR_ERR(ent); 1478 } 1479 1480 /* Register debugfs entries for amdgpu_ttm */ 1481 amdgpu_ttm_debugfs_init(adev); 1482 amdgpu_debugfs_pm_init(adev); 1483 amdgpu_debugfs_sa_init(adev); 1484 amdgpu_debugfs_fence_init(adev); 1485 amdgpu_debugfs_gem_init(adev); 1486 1487 r = amdgpu_debugfs_regs_init(adev); 1488 if (r) 1489 DRM_ERROR("registering register debugfs failed (%d).\n", r); 1490 1491 amdgpu_debugfs_firmware_init(adev); 1492 1493 #if defined(CONFIG_DRM_AMD_DC) 1494 if (amdgpu_device_has_dc_support(adev)) 1495 dtn_debugfs_init(adev); 1496 #endif 1497 1498 for (i = 0; i < AMDGPU_MAX_RINGS; ++i) { 1499 struct amdgpu_ring *ring = adev->rings[i]; 1500 1501 if (!ring) 1502 continue; 1503 1504 if (amdgpu_debugfs_ring_init(adev, ring)) { 1505 DRM_ERROR("Failed to register debugfs file for rings !\n"); 1506 } 1507 } 1508 1509 amdgpu_ras_debugfs_create_all(adev); 1510 amdgpu_rap_debugfs_init(adev); 1511 amdgpu_securedisplay_debugfs_init(adev); 1512 amdgpu_fw_attestation_debugfs_init(adev); 1513 1514 debugfs_create_file("amdgpu_evict_vram", 0444, root, adev, 1515 &amdgpu_evict_vram_fops); 1516 debugfs_create_file("amdgpu_evict_gtt", 0444, root, adev, 1517 &amdgpu_evict_gtt_fops); 1518 debugfs_create_file("amdgpu_test_ib", 0444, root, adev, 1519 &amdgpu_debugfs_test_ib_fops); 1520 debugfs_create_file("amdgpu_vm_info", 0444, root, adev, 1521 &amdgpu_debugfs_vm_info_fops); 1522 1523 adev->debugfs_vbios_blob.data = adev->bios; 1524 adev->debugfs_vbios_blob.size = adev->bios_size; 1525 debugfs_create_blob("amdgpu_vbios", 0444, root, 1526 &adev->debugfs_vbios_blob); 1527 1528 return 0; 1529 } 1530 1531 #else 1532 int amdgpu_debugfs_init(struct amdgpu_device *adev) 1533 { 1534 return 0; 1535 } 1536 int amdgpu_debugfs_regs_init(struct amdgpu_device *adev) 1537 { 1538 return 0; 1539 } 1540 #endif 1541