1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2016-2019 Intel Corporation 4 */ 5 6 #include <linux/types.h> 7 8 #include "gt/intel_gt.h" 9 #include "intel_guc_reg.h" 10 #include "intel_huc.h" 11 #include "intel_huc_print.h" 12 #include "i915_drv.h" 13 #include "i915_reg.h" 14 #include "pxp/intel_pxp_cmd_interface_43.h" 15 16 #include <linux/device/bus.h> 17 #include <linux/mei_aux.h> 18 19 /** 20 * DOC: HuC 21 * 22 * The HuC is a dedicated microcontroller for usage in media HEVC (High 23 * Efficiency Video Coding) operations. Userspace can directly use the firmware 24 * capabilities by adding HuC specific commands to batch buffers. 25 * 26 * The kernel driver is only responsible for loading the HuC firmware and 27 * triggering its security authentication. This is done differently depending 28 * on the platform: 29 * - older platforms (from Gen9 to most Gen12s): the load is performed via DMA 30 * and the authentication via GuC 31 * - DG2: load and authentication are both performed via GSC. 32 * - MTL and newer platforms: the load is performed via DMA (same as with 33 * not-DG2 older platforms), while the authentication is done in 2-steps, 34 * a first auth for clear-media workloads via GuC and a second one for all 35 * workloads via GSC. 36 * On platforms where the GuC does the authentication, to correctly do so the 37 * HuC binary must be loaded before the GuC one. 38 * Loading the HuC is optional; however, not using the HuC might negatively 39 * impact power usage and/or performance of media workloads, depending on the 40 * use-cases. 41 * HuC must be reloaded on events that cause the WOPCM to lose its contents 42 * (S3/S4, FLR); on older platforms the HuC must also be reloaded on GuC/GT 43 * reset, while on newer ones it will survive that. 44 * 45 * See https://github.com/intel/media-driver for the latest details on HuC 46 * functionality. 47 */ 48 49 /** 50 * DOC: HuC Memory Management 51 * 52 * Similarly to the GuC, the HuC can't do any memory allocations on its own, 53 * with the difference being that the allocations for HuC usage are handled by 54 * the userspace driver instead of the kernel one. The HuC accesses the memory 55 * via the PPGTT belonging to the context loaded on the VCS executing the 56 * HuC-specific commands. 57 */ 58 59 /* 60 * MEI-GSC load is an async process. The probing of the exposed aux device 61 * (see intel_gsc.c) usually happens a few seconds after i915 probe, depending 62 * on when the kernel schedules it. Unless something goes terribly wrong, we're 63 * guaranteed for this to happen during boot, so the big timeout is a safety net 64 * that we never expect to need. 65 * MEI-PXP + HuC load usually takes ~300ms, but if the GSC needs to be resumed 66 * and/or reset, this can take longer. Note that the kernel might schedule 67 * other work between the i915 init/resume and the MEI one, which can add to 68 * the delay. 69 */ 70 #define GSC_INIT_TIMEOUT_MS 10000 71 #define PXP_INIT_TIMEOUT_MS 5000 72 73 static int sw_fence_dummy_notify(struct i915_sw_fence *sf, 74 enum i915_sw_fence_notify state) 75 { 76 return NOTIFY_DONE; 77 } 78 79 static void __delayed_huc_load_complete(struct intel_huc *huc) 80 { 81 if (!i915_sw_fence_done(&huc->delayed_load.fence)) 82 i915_sw_fence_complete(&huc->delayed_load.fence); 83 } 84 85 static void delayed_huc_load_complete(struct intel_huc *huc) 86 { 87 hrtimer_cancel(&huc->delayed_load.timer); 88 __delayed_huc_load_complete(huc); 89 } 90 91 static void __gsc_init_error(struct intel_huc *huc) 92 { 93 huc->delayed_load.status = INTEL_HUC_DELAYED_LOAD_ERROR; 94 __delayed_huc_load_complete(huc); 95 } 96 97 static void gsc_init_error(struct intel_huc *huc) 98 { 99 hrtimer_cancel(&huc->delayed_load.timer); 100 __gsc_init_error(huc); 101 } 102 103 static void gsc_init_done(struct intel_huc *huc) 104 { 105 hrtimer_cancel(&huc->delayed_load.timer); 106 107 /* MEI-GSC init is done, now we wait for MEI-PXP to bind */ 108 huc->delayed_load.status = INTEL_HUC_WAITING_ON_PXP; 109 if (!i915_sw_fence_done(&huc->delayed_load.fence)) 110 hrtimer_start(&huc->delayed_load.timer, 111 ms_to_ktime(PXP_INIT_TIMEOUT_MS), 112 HRTIMER_MODE_REL); 113 } 114 115 static enum hrtimer_restart huc_delayed_load_timer_callback(struct hrtimer *hrtimer) 116 { 117 struct intel_huc *huc = container_of(hrtimer, struct intel_huc, delayed_load.timer); 118 119 if (!intel_huc_is_authenticated(huc, INTEL_HUC_AUTH_BY_GSC)) { 120 if (huc->delayed_load.status == INTEL_HUC_WAITING_ON_GSC) 121 huc_notice(huc, "timed out waiting for MEI GSC\n"); 122 else if (huc->delayed_load.status == INTEL_HUC_WAITING_ON_PXP) 123 huc_notice(huc, "timed out waiting for MEI PXP\n"); 124 else 125 MISSING_CASE(huc->delayed_load.status); 126 127 __gsc_init_error(huc); 128 } 129 130 return HRTIMER_NORESTART; 131 } 132 133 static void huc_delayed_load_start(struct intel_huc *huc) 134 { 135 ktime_t delay; 136 137 GEM_BUG_ON(intel_huc_is_authenticated(huc, INTEL_HUC_AUTH_BY_GSC)); 138 139 /* 140 * On resume we don't have to wait for MEI-GSC to be re-probed, but we 141 * do need to wait for MEI-PXP to reset & re-bind 142 */ 143 switch (huc->delayed_load.status) { 144 case INTEL_HUC_WAITING_ON_GSC: 145 delay = ms_to_ktime(GSC_INIT_TIMEOUT_MS); 146 break; 147 case INTEL_HUC_WAITING_ON_PXP: 148 delay = ms_to_ktime(PXP_INIT_TIMEOUT_MS); 149 break; 150 default: 151 gsc_init_error(huc); 152 return; 153 } 154 155 /* 156 * This fence is always complete unless we're waiting for the 157 * GSC device to come up to load the HuC. We arm the fence here 158 * and complete it when we confirm that the HuC is loaded from 159 * the PXP bind callback. 160 */ 161 GEM_BUG_ON(!i915_sw_fence_done(&huc->delayed_load.fence)); 162 i915_sw_fence_fini(&huc->delayed_load.fence); 163 i915_sw_fence_reinit(&huc->delayed_load.fence); 164 i915_sw_fence_await(&huc->delayed_load.fence); 165 i915_sw_fence_commit(&huc->delayed_load.fence); 166 167 hrtimer_start(&huc->delayed_load.timer, delay, HRTIMER_MODE_REL); 168 } 169 170 static int gsc_notifier(struct notifier_block *nb, unsigned long action, void *data) 171 { 172 struct device *dev = data; 173 struct intel_huc *huc = container_of(nb, struct intel_huc, delayed_load.nb); 174 struct intel_gsc_intf *intf = &huc_to_gt(huc)->gsc.intf[0]; 175 176 if (!intf->adev || &intf->adev->aux_dev.dev != dev) 177 return 0; 178 179 switch (action) { 180 case BUS_NOTIFY_BOUND_DRIVER: /* mei driver bound to aux device */ 181 gsc_init_done(huc); 182 break; 183 184 case BUS_NOTIFY_DRIVER_NOT_BOUND: /* mei driver fails to be bound */ 185 case BUS_NOTIFY_UNBIND_DRIVER: /* mei driver about to be unbound */ 186 huc_info(huc, "MEI driver not bound, disabling load\n"); 187 gsc_init_error(huc); 188 break; 189 } 190 191 return 0; 192 } 193 194 void intel_huc_register_gsc_notifier(struct intel_huc *huc, const struct bus_type *bus) 195 { 196 int ret; 197 198 if (!intel_huc_is_loaded_by_gsc(huc)) 199 return; 200 201 huc->delayed_load.nb.notifier_call = gsc_notifier; 202 ret = bus_register_notifier(bus, &huc->delayed_load.nb); 203 if (ret) { 204 huc_err(huc, "failed to register GSC notifier %pe\n", ERR_PTR(ret)); 205 huc->delayed_load.nb.notifier_call = NULL; 206 gsc_init_error(huc); 207 } 208 } 209 210 void intel_huc_unregister_gsc_notifier(struct intel_huc *huc, const struct bus_type *bus) 211 { 212 if (!huc->delayed_load.nb.notifier_call) 213 return; 214 215 delayed_huc_load_complete(huc); 216 217 bus_unregister_notifier(bus, &huc->delayed_load.nb); 218 huc->delayed_load.nb.notifier_call = NULL; 219 } 220 221 static void delayed_huc_load_init(struct intel_huc *huc) 222 { 223 /* 224 * Initialize fence to be complete as this is expected to be complete 225 * unless there is a delayed HuC load in progress. 226 */ 227 i915_sw_fence_init(&huc->delayed_load.fence, 228 sw_fence_dummy_notify); 229 i915_sw_fence_commit(&huc->delayed_load.fence); 230 231 hrtimer_init(&huc->delayed_load.timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL); 232 huc->delayed_load.timer.function = huc_delayed_load_timer_callback; 233 } 234 235 static void delayed_huc_load_fini(struct intel_huc *huc) 236 { 237 /* 238 * the fence is initialized in init_early, so we need to clean it up 239 * even if HuC loading is off. 240 */ 241 delayed_huc_load_complete(huc); 242 i915_sw_fence_fini(&huc->delayed_load.fence); 243 } 244 245 int intel_huc_sanitize(struct intel_huc *huc) 246 { 247 delayed_huc_load_complete(huc); 248 intel_uc_fw_sanitize(&huc->fw); 249 return 0; 250 } 251 252 static bool vcs_supported(struct intel_gt *gt) 253 { 254 intel_engine_mask_t mask = gt->info.engine_mask; 255 256 /* 257 * We reach here from i915_driver_early_probe for the primary GT before 258 * its engine mask is set, so we use the device info engine mask for it; 259 * this means we're not taking VCS fusing into account, but if the 260 * primary GT supports VCS engines we expect at least one of them to 261 * remain unfused so we're fine. 262 * For other GTs we expect the GT-specific mask to be set before we 263 * call this function. 264 */ 265 GEM_BUG_ON(!gt_is_root(gt) && !gt->info.engine_mask); 266 267 if (gt_is_root(gt)) 268 mask = RUNTIME_INFO(gt->i915)->platform_engine_mask; 269 else 270 mask = gt->info.engine_mask; 271 272 return __ENGINE_INSTANCES_MASK(mask, VCS0, I915_MAX_VCS); 273 } 274 275 void intel_huc_init_early(struct intel_huc *huc) 276 { 277 struct drm_i915_private *i915 = huc_to_gt(huc)->i915; 278 struct intel_gt *gt = huc_to_gt(huc); 279 280 intel_uc_fw_init_early(&huc->fw, INTEL_UC_FW_TYPE_HUC, true); 281 282 /* 283 * we always init the fence as already completed, even if HuC is not 284 * supported. This way we don't have to distinguish between HuC not 285 * supported/disabled or already loaded, and can focus on if the load 286 * is currently in progress (fence not complete) or not, which is what 287 * we care about for stalling userspace submissions. 288 */ 289 delayed_huc_load_init(huc); 290 291 if (!vcs_supported(gt)) { 292 intel_uc_fw_change_status(&huc->fw, INTEL_UC_FIRMWARE_NOT_SUPPORTED); 293 return; 294 } 295 296 if (GRAPHICS_VER(i915) >= 11) { 297 huc->status[INTEL_HUC_AUTH_BY_GUC].reg = GEN11_HUC_KERNEL_LOAD_INFO; 298 huc->status[INTEL_HUC_AUTH_BY_GUC].mask = HUC_LOAD_SUCCESSFUL; 299 huc->status[INTEL_HUC_AUTH_BY_GUC].value = HUC_LOAD_SUCCESSFUL; 300 } else { 301 huc->status[INTEL_HUC_AUTH_BY_GUC].reg = HUC_STATUS2; 302 huc->status[INTEL_HUC_AUTH_BY_GUC].mask = HUC_FW_VERIFIED; 303 huc->status[INTEL_HUC_AUTH_BY_GUC].value = HUC_FW_VERIFIED; 304 } 305 306 if (IS_DG2(i915)) { 307 huc->status[INTEL_HUC_AUTH_BY_GSC].reg = GEN11_HUC_KERNEL_LOAD_INFO; 308 huc->status[INTEL_HUC_AUTH_BY_GSC].mask = HUC_LOAD_SUCCESSFUL; 309 huc->status[INTEL_HUC_AUTH_BY_GSC].value = HUC_LOAD_SUCCESSFUL; 310 } else { 311 huc->status[INTEL_HUC_AUTH_BY_GSC].reg = HECI_FWSTS5(MTL_GSC_HECI1_BASE); 312 huc->status[INTEL_HUC_AUTH_BY_GSC].mask = HECI_FWSTS5_HUC_AUTH_DONE; 313 huc->status[INTEL_HUC_AUTH_BY_GSC].value = HECI_FWSTS5_HUC_AUTH_DONE; 314 } 315 } 316 317 #define HUC_LOAD_MODE_STRING(x) (x ? "GSC" : "legacy") 318 static int check_huc_loading_mode(struct intel_huc *huc) 319 { 320 struct intel_gt *gt = huc_to_gt(huc); 321 bool gsc_enabled = huc->fw.has_gsc_headers; 322 323 /* 324 * The fuse for HuC load via GSC is only valid on platforms that have 325 * GuC deprivilege. 326 */ 327 if (HAS_GUC_DEPRIVILEGE(gt->i915)) 328 huc->loaded_via_gsc = intel_uncore_read(gt->uncore, GUC_SHIM_CONTROL2) & 329 GSC_LOADS_HUC; 330 331 if (huc->loaded_via_gsc && !gsc_enabled) { 332 huc_err(huc, "HW requires a GSC-enabled blob, but we found a legacy one\n"); 333 return -ENOEXEC; 334 } 335 336 /* 337 * On newer platforms we have GSC-enabled binaries but we load the HuC 338 * via DMA. To do so we need to find the location of the legacy-style 339 * binary inside the GSC-enabled one, which we do at fetch time. Make 340 * sure that we were able to do so if the fuse says we need to load via 341 * DMA and the binary is GSC-enabled. 342 */ 343 if (!huc->loaded_via_gsc && gsc_enabled && !huc->fw.dma_start_offset) { 344 huc_err(huc, "HW in DMA mode, but we have an incompatible GSC-enabled blob\n"); 345 return -ENOEXEC; 346 } 347 348 /* 349 * If the HuC is loaded via GSC, we need to be able to access the GSC. 350 * On DG2 this is done via the mei components, while on newer platforms 351 * it is done via the GSCCS, 352 */ 353 if (huc->loaded_via_gsc) { 354 if (IS_DG2(gt->i915)) { 355 if (!IS_ENABLED(CONFIG_INTEL_MEI_PXP) || 356 !IS_ENABLED(CONFIG_INTEL_MEI_GSC)) { 357 huc_info(huc, "can't load due to missing mei modules\n"); 358 return -EIO; 359 } 360 } else { 361 if (!HAS_ENGINE(gt, GSC0)) { 362 huc_info(huc, "can't load due to missing GSCCS\n"); 363 return -EIO; 364 } 365 } 366 } 367 368 huc_dbg(huc, "loaded by GSC = %s\n", str_yes_no(huc->loaded_via_gsc)); 369 370 return 0; 371 } 372 373 int intel_huc_init(struct intel_huc *huc) 374 { 375 struct intel_gt *gt = huc_to_gt(huc); 376 int err; 377 378 err = check_huc_loading_mode(huc); 379 if (err) 380 goto out; 381 382 if (HAS_ENGINE(gt, GSC0)) { 383 struct i915_vma *vma; 384 385 vma = intel_guc_allocate_vma(>->uc.guc, PXP43_HUC_AUTH_INOUT_SIZE * 2); 386 if (IS_ERR(vma)) { 387 err = PTR_ERR(vma); 388 huc_info(huc, "Failed to allocate heci pkt\n"); 389 goto out; 390 } 391 392 huc->heci_pkt = vma; 393 } 394 395 err = intel_uc_fw_init(&huc->fw); 396 if (err) 397 goto out_pkt; 398 399 intel_uc_fw_change_status(&huc->fw, INTEL_UC_FIRMWARE_LOADABLE); 400 401 return 0; 402 403 out_pkt: 404 if (huc->heci_pkt) 405 i915_vma_unpin_and_release(&huc->heci_pkt, 0); 406 out: 407 intel_uc_fw_change_status(&huc->fw, INTEL_UC_FIRMWARE_INIT_FAIL); 408 huc_info(huc, "initialization failed %pe\n", ERR_PTR(err)); 409 return err; 410 } 411 412 void intel_huc_fini(struct intel_huc *huc) 413 { 414 /* 415 * the fence is initialized in init_early, so we need to clean it up 416 * even if HuC loading is off. 417 */ 418 delayed_huc_load_fini(huc); 419 420 if (huc->heci_pkt) 421 i915_vma_unpin_and_release(&huc->heci_pkt, 0); 422 423 if (intel_uc_fw_is_loadable(&huc->fw)) 424 intel_uc_fw_fini(&huc->fw); 425 } 426 427 void intel_huc_suspend(struct intel_huc *huc) 428 { 429 if (!intel_uc_fw_is_loadable(&huc->fw)) 430 return; 431 432 /* 433 * in the unlikely case that we're suspending before the GSC has 434 * completed its loading sequence, just stop waiting. We'll restart 435 * on resume. 436 */ 437 delayed_huc_load_complete(huc); 438 } 439 440 static const char *auth_mode_string(struct intel_huc *huc, 441 enum intel_huc_authentication_type type) 442 { 443 bool partial = huc->fw.has_gsc_headers && type == INTEL_HUC_AUTH_BY_GUC; 444 445 return partial ? "clear media" : "all workloads"; 446 } 447 448 int intel_huc_wait_for_auth_complete(struct intel_huc *huc, 449 enum intel_huc_authentication_type type) 450 { 451 struct intel_gt *gt = huc_to_gt(huc); 452 int ret; 453 454 ret = __intel_wait_for_register(gt->uncore, 455 huc->status[type].reg, 456 huc->status[type].mask, 457 huc->status[type].value, 458 2, 50, NULL); 459 460 /* mark the load process as complete even if the wait failed */ 461 delayed_huc_load_complete(huc); 462 463 if (ret) { 464 huc_err(huc, "firmware not verified for %s: %pe\n", 465 auth_mode_string(huc, type), ERR_PTR(ret)); 466 intel_uc_fw_change_status(&huc->fw, INTEL_UC_FIRMWARE_LOAD_FAIL); 467 return ret; 468 } 469 470 intel_uc_fw_change_status(&huc->fw, INTEL_UC_FIRMWARE_RUNNING); 471 huc_info(huc, "authenticated for %s\n", auth_mode_string(huc, type)); 472 return 0; 473 } 474 475 /** 476 * intel_huc_auth() - Authenticate HuC uCode 477 * @huc: intel_huc structure 478 * @type: authentication type (via GuC or via GSC) 479 * 480 * Called after HuC and GuC firmware loading during intel_uc_init_hw(). 481 * 482 * This function invokes the GuC action to authenticate the HuC firmware, 483 * passing the offset of the RSA signature to intel_guc_auth_huc(). It then 484 * waits for up to 50ms for firmware verification ACK. 485 */ 486 int intel_huc_auth(struct intel_huc *huc, enum intel_huc_authentication_type type) 487 { 488 struct intel_gt *gt = huc_to_gt(huc); 489 struct intel_guc *guc = >->uc.guc; 490 int ret; 491 492 if (!intel_uc_fw_is_loaded(&huc->fw)) 493 return -ENOEXEC; 494 495 /* GSC will do the auth with the load */ 496 if (intel_huc_is_loaded_by_gsc(huc)) 497 return -ENODEV; 498 499 if (intel_huc_is_authenticated(huc, type)) 500 return -EEXIST; 501 502 ret = i915_inject_probe_error(gt->i915, -ENXIO); 503 if (ret) 504 goto fail; 505 506 switch (type) { 507 case INTEL_HUC_AUTH_BY_GUC: 508 ret = intel_guc_auth_huc(guc, intel_guc_ggtt_offset(guc, huc->fw.rsa_data)); 509 break; 510 case INTEL_HUC_AUTH_BY_GSC: 511 ret = intel_huc_fw_auth_via_gsccs(huc); 512 break; 513 default: 514 MISSING_CASE(type); 515 ret = -EINVAL; 516 } 517 if (ret) 518 goto fail; 519 520 /* Check authentication status, it should be done by now */ 521 ret = intel_huc_wait_for_auth_complete(huc, type); 522 if (ret) 523 goto fail; 524 525 return 0; 526 527 fail: 528 huc_probe_error(huc, "%s authentication failed %pe\n", 529 auth_mode_string(huc, type), ERR_PTR(ret)); 530 return ret; 531 } 532 533 bool intel_huc_is_authenticated(struct intel_huc *huc, 534 enum intel_huc_authentication_type type) 535 { 536 struct intel_gt *gt = huc_to_gt(huc); 537 intel_wakeref_t wakeref; 538 u32 status = 0; 539 540 with_intel_runtime_pm(gt->uncore->rpm, wakeref) 541 status = intel_uncore_read(gt->uncore, huc->status[type].reg); 542 543 return (status & huc->status[type].mask) == huc->status[type].value; 544 } 545 546 static bool huc_is_fully_authenticated(struct intel_huc *huc) 547 { 548 struct intel_uc_fw *huc_fw = &huc->fw; 549 550 if (!huc_fw->has_gsc_headers) 551 return intel_huc_is_authenticated(huc, INTEL_HUC_AUTH_BY_GUC); 552 else if (intel_huc_is_loaded_by_gsc(huc) || HAS_ENGINE(huc_to_gt(huc), GSC0)) 553 return intel_huc_is_authenticated(huc, INTEL_HUC_AUTH_BY_GSC); 554 else 555 return false; 556 } 557 558 /** 559 * intel_huc_check_status() - check HuC status 560 * @huc: intel_huc structure 561 * 562 * This function reads status register to verify if HuC 563 * firmware was successfully loaded. 564 * 565 * The return values match what is expected for the I915_PARAM_HUC_STATUS 566 * getparam. 567 */ 568 int intel_huc_check_status(struct intel_huc *huc) 569 { 570 struct intel_uc_fw *huc_fw = &huc->fw; 571 572 switch (__intel_uc_fw_status(huc_fw)) { 573 case INTEL_UC_FIRMWARE_NOT_SUPPORTED: 574 return -ENODEV; 575 case INTEL_UC_FIRMWARE_DISABLED: 576 return -EOPNOTSUPP; 577 case INTEL_UC_FIRMWARE_MISSING: 578 return -ENOPKG; 579 case INTEL_UC_FIRMWARE_ERROR: 580 return -ENOEXEC; 581 case INTEL_UC_FIRMWARE_INIT_FAIL: 582 return -ENOMEM; 583 case INTEL_UC_FIRMWARE_LOAD_FAIL: 584 return -EIO; 585 default: 586 break; 587 } 588 589 /* 590 * GSC-enabled binaries loaded via DMA are first partially 591 * authenticated by GuC and then fully authenticated by GSC 592 */ 593 if (huc_is_fully_authenticated(huc)) 594 return 1; /* full auth */ 595 else if (huc_fw->has_gsc_headers && !intel_huc_is_loaded_by_gsc(huc) && 596 intel_huc_is_authenticated(huc, INTEL_HUC_AUTH_BY_GUC)) 597 return 2; /* clear media only */ 598 else 599 return 0; 600 } 601 602 static bool huc_has_delayed_load(struct intel_huc *huc) 603 { 604 return intel_huc_is_loaded_by_gsc(huc) && 605 (huc->delayed_load.status != INTEL_HUC_DELAYED_LOAD_ERROR); 606 } 607 608 void intel_huc_update_auth_status(struct intel_huc *huc) 609 { 610 if (!intel_uc_fw_is_loadable(&huc->fw)) 611 return; 612 613 if (!huc->fw.has_gsc_headers) 614 return; 615 616 if (huc_is_fully_authenticated(huc)) 617 intel_uc_fw_change_status(&huc->fw, 618 INTEL_UC_FIRMWARE_RUNNING); 619 else if (huc_has_delayed_load(huc)) 620 huc_delayed_load_start(huc); 621 } 622 623 /** 624 * intel_huc_load_status - dump information about HuC load status 625 * @huc: the HuC 626 * @p: the &drm_printer 627 * 628 * Pretty printer for HuC load status. 629 */ 630 void intel_huc_load_status(struct intel_huc *huc, struct drm_printer *p) 631 { 632 struct intel_gt *gt = huc_to_gt(huc); 633 intel_wakeref_t wakeref; 634 635 if (!intel_huc_is_supported(huc)) { 636 drm_printf(p, "HuC not supported\n"); 637 return; 638 } 639 640 if (!intel_huc_is_wanted(huc)) { 641 drm_printf(p, "HuC disabled\n"); 642 return; 643 } 644 645 intel_uc_fw_dump(&huc->fw, p); 646 647 with_intel_runtime_pm(gt->uncore->rpm, wakeref) 648 drm_printf(p, "HuC status: 0x%08x\n", 649 intel_uncore_read(gt->uncore, huc->status[INTEL_HUC_AUTH_BY_GUC].reg)); 650 } 651