/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_crat.c | 1526 kfd_fill_gpu_cache_info_from_gfx_config(kdev->kfd, *pcache_info); in kfd_get_gpu_cache_info() 1990 bool use_ta_info = kdev->kfd->num_nodes == 1; in kfd_fill_gpu_xgmi_link_to_gpu() 2016 bool is_single_hop = kdev->kfd == peer_kdev->kfd; in kfd_fill_gpu_xgmi_link_to_gpu() 2091 (cu_info.cu_active_number / kdev->kfd->num_nodes); in kfd_create_vcrat_image_gpu() 2166 if (kdev->kfd->hive_id) { in kfd_create_vcrat_image_gpu() 2171 if (peer_dev->gpu->kfd->hive_id != kdev->kfd->hive_id) in kfd_create_vcrat_image_gpu()
|
H A D | kfd_mqd_manager_v9.c | 44 if (mm->dev->kfd->cwsr_enabled && in mqd_stride_v9() 132 if (node->kfd->cwsr_enabled && (q->type == KFD_QUEUE_TYPE_COMPUTE)) { in allocate_mqd() 207 if (mm->dev->kfd->cwsr_enabled && q->ctx_save_restore_area_address) { in init_mqd() 299 if (mm->dev->kfd->cwsr_enabled && q->ctx_save_restore_area_address) in update_mqd() 636 if (mm->dev->kfd->cwsr_enabled && in init_mqd_v9_4_3()
|
H A D | kfd_packet_manager_vi.c | 80 struct kfd_node *kfd = pm->dqm->dev; in pm_runlist_vi() local 95 kfd->max_proc_per_quantum); in pm_runlist_vi()
|
H A D | kfd_int_process_v9.c | 262 dev->kfd->device_info.ih_ring_entry_size); in event_interrupt_isr_v9() 291 if (context_id == 0 && context_id_expected(dev->kfd)) in event_interrupt_isr_v9()
|
H A D | kfd_mqd_manager.h | 71 struct kfd_mem_obj* (*allocate_mqd)(struct kfd_node *kfd,
|
H A D | kfd_mqd_manager_cik.c | 76 static struct kfd_mem_obj *allocate_mqd(struct kfd_node *kfd, in allocate_mqd() argument 81 if (kfd_gtt_sa_allocate(kfd, sizeof(struct cik_mqd), in allocate_mqd()
|
H A D | kfd_svm.h | 202 #define KFD_IS_SVM_API_SUPPORTED(adev) ((adev)->kfd.pgmap.type != 0 ||\
|
H A D | kfd_chardev.c | 337 err = kfd_alloc_process_doorbells(dev->kfd, pdd); in kfd_ioctl_create_queue() 347 if (dev->kfd->shared_resources.enable_mes && in kfd_ioctl_create_queue() 1034 dev->kfd->adev->gmc.is_app_apu) { in kfd_dev_is_large_bar() 1126 if (args->size != kfd_doorbell_process_slice(dev->kfd)) { in kfd_ioctl_alloc_memory_of_gpu() 1426 flush_tlb = kfd_flush_tlb_after_unmap(pdd->dev->kfd); in kfd_ioctl_unmap_memory_from_gpu() 1539 args->gpu_id = dmabuf_adev->kfd.dev->nodes[xcp_id]->id; in kfd_ioctl_get_dmabuf_info() 2268 ret = kfd_alloc_process_doorbells(dev->kfd, pdd); in criu_restore_devices() 2297 kfd_doorbell_process_slice(pdd->dev->kfd)) in criu_restore_memory_of_gpu() 2768 if (pdd->dev->kfd->shared_resources.enable_mes) in runtime_enable() 2860 if (!pdd->dev->kfd->shared_resources.enable_mes) in runtime_disable()
|
H A D | kfd_migrate.c | 209 return (addr + adev->kfd.pgmap.range.start) >> PAGE_SHIFT; in svm_migrate_addr_to_pfn() 239 return (addr - adev->kfd.pgmap.range.start); in svm_migrate_addr() 1011 struct amdgpu_kfd_dev *kfddev = &adev->kfd; in kgd2kfd_init_zone_device()
|
H A D | kfd_events.c | 351 struct kfd_node *kfd; in kfd_kmap_event_page() local 367 kfd = pdd->dev; in kfd_kmap_event_page() 369 pdd = kfd_bind_process_to_device(kfd, p); in kfd_kmap_event_page()
|
H A D | kfd_svm.c | 187 bo_adev->kfd.pgmap.range.start; in svm_range_dma_map_dev() 1997 for (i = 0; i < adev->kfd.dev->num_nodes; i++) { in svm_range_set_max_pages() 1998 if (adev->kfd.dev->nodes[i]->xcp) in svm_range_set_max_pages() 1999 id = adev->kfd.dev->nodes[i]->xcp->id; in svm_range_set_max_pages()
|
/openbmc/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_umc.c | 88 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_umc_do_page_retirement() 178 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_umc_poison_handler()
|
H A D | amdgpu_amdkfd_gpuvm.c | 215 (adev && xcp_id >= 0 && adev->kfd.vram_used[xcp_id] + vram_needed > in amdgpu_amdkfd_reserve_mem_limit() 227 adev->kfd.vram_used[xcp_id] += vram_needed; in amdgpu_amdkfd_reserve_mem_limit() 228 adev->kfd.vram_used_aligned[xcp_id] += adev->gmc.is_app_apu ? in amdgpu_amdkfd_reserve_mem_limit() 255 adev->kfd.vram_used[xcp_id] -= size; in amdgpu_amdkfd_unreserve_mem_limit() 257 adev->kfd.vram_used_aligned[xcp_id] -= size; in amdgpu_amdkfd_unreserve_mem_limit() 261 adev->kfd.vram_used_aligned[xcp_id] -= in amdgpu_amdkfd_unreserve_mem_limit() 273 WARN_ONCE(adev && xcp_id >= 0 && adev->kfd.vram_used[xcp_id] < 0, in amdgpu_amdkfd_unreserve_mem_limit() 1604 - adev->kfd.vram_used_aligned[xcp_id] in amdgpu_amdkfd_get_available_memory()
|
H A D | aqua_vanjaram.c | 484 if (!xcp_mgr->adev->kfd.init_complete) in __aqua_vanjaram_post_partition_switch() 510 if (adev->kfd.init_complete) in aqua_vanjaram_switch_partition_mode()
|
H A D | amdgpu_sdma.c | 127 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_sdma_process_ras_data_cb()
|
H A D | amdgpu_gfx.c | 873 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_gfx_process_ras_data_cb()
|
H A D | amdgpu.h | 959 struct amdgpu_kfd_dev kfd; member
|
H A D | amdgpu_device.c | 5161 if (adev->kfd.dev) in amdgpu_device_stop_pending_resets() 5162 cancel_work(&adev->kfd.reset_work); in amdgpu_device_stop_pending_resets() 5406 if (!adev->kfd.init_complete) in amdgpu_device_gpu_recover()
|
H A D | amdgpu_drv.c | 2375 if (!adev->kfd.init_complete) in amdgpu_drv_delayed_reset_work_handler()
|
/openbmc/linux/tools/perf/util/ |
H A D | probe-event.c | 3718 int ret, ret2, ufd = -1, kfd = -1; in del_perf_probe_events() local 3725 ret = probe_file__open_both(&kfd, &ufd, PF_FL_RW); in del_perf_probe_events() 3729 ret = probe_file__del_events(kfd, filter); in del_perf_probe_events() 3741 if (kfd >= 0) in del_perf_probe_events() 3742 close(kfd); in del_perf_probe_events()
|
/openbmc/linux/drivers/gpu/drm/amd/pm/swsmu/smu13/ |
H A D | aldebaran_ppt.c | 1711 kgd2kfd_smi_event_throttle(smu->adev->kfd.dev, in aldebaran_log_thermal_throttling_event()
|
H A D | smu_v13_0_6_ppt.c | 1922 smu->adev->kfd.dev, in smu_v13_0_6_log_thermal_throttling_event()
|
/openbmc/linux/drivers/gpu/drm/amd/pm/swsmu/smu11/ |
H A D | arcturus_ppt.c | 2347 kgd2kfd_smi_event_throttle(smu->adev->kfd.dev, in arcturus_log_thermal_throttling_event()
|