Home
last modified time | relevance | path

Searched refs:kfd (Results 26 – 48 of 48) sorted by relevance

12

/openbmc/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_crat.c1526 kfd_fill_gpu_cache_info_from_gfx_config(kdev->kfd, *pcache_info); in kfd_get_gpu_cache_info()
1990 bool use_ta_info = kdev->kfd->num_nodes == 1; in kfd_fill_gpu_xgmi_link_to_gpu()
2016 bool is_single_hop = kdev->kfd == peer_kdev->kfd; in kfd_fill_gpu_xgmi_link_to_gpu()
2091 (cu_info.cu_active_number / kdev->kfd->num_nodes); in kfd_create_vcrat_image_gpu()
2166 if (kdev->kfd->hive_id) { in kfd_create_vcrat_image_gpu()
2171 if (peer_dev->gpu->kfd->hive_id != kdev->kfd->hive_id) in kfd_create_vcrat_image_gpu()
H A Dkfd_mqd_manager_v9.c44 if (mm->dev->kfd->cwsr_enabled && in mqd_stride_v9()
132 if (node->kfd->cwsr_enabled && (q->type == KFD_QUEUE_TYPE_COMPUTE)) { in allocate_mqd()
207 if (mm->dev->kfd->cwsr_enabled && q->ctx_save_restore_area_address) { in init_mqd()
299 if (mm->dev->kfd->cwsr_enabled && q->ctx_save_restore_area_address) in update_mqd()
636 if (mm->dev->kfd->cwsr_enabled && in init_mqd_v9_4_3()
H A Dkfd_packet_manager_vi.c80 struct kfd_node *kfd = pm->dqm->dev; in pm_runlist_vi() local
95 kfd->max_proc_per_quantum); in pm_runlist_vi()
H A Dkfd_int_process_v9.c262 dev->kfd->device_info.ih_ring_entry_size); in event_interrupt_isr_v9()
291 if (context_id == 0 && context_id_expected(dev->kfd)) in event_interrupt_isr_v9()
H A Dkfd_mqd_manager.h71 struct kfd_mem_obj* (*allocate_mqd)(struct kfd_node *kfd,
H A Dkfd_mqd_manager_cik.c76 static struct kfd_mem_obj *allocate_mqd(struct kfd_node *kfd, in allocate_mqd() argument
81 if (kfd_gtt_sa_allocate(kfd, sizeof(struct cik_mqd), in allocate_mqd()
H A Dkfd_svm.h202 #define KFD_IS_SVM_API_SUPPORTED(adev) ((adev)->kfd.pgmap.type != 0 ||\
H A Dkfd_chardev.c337 err = kfd_alloc_process_doorbells(dev->kfd, pdd); in kfd_ioctl_create_queue()
347 if (dev->kfd->shared_resources.enable_mes && in kfd_ioctl_create_queue()
1034 dev->kfd->adev->gmc.is_app_apu) { in kfd_dev_is_large_bar()
1126 if (args->size != kfd_doorbell_process_slice(dev->kfd)) { in kfd_ioctl_alloc_memory_of_gpu()
1426 flush_tlb = kfd_flush_tlb_after_unmap(pdd->dev->kfd); in kfd_ioctl_unmap_memory_from_gpu()
1539 args->gpu_id = dmabuf_adev->kfd.dev->nodes[xcp_id]->id; in kfd_ioctl_get_dmabuf_info()
2268 ret = kfd_alloc_process_doorbells(dev->kfd, pdd); in criu_restore_devices()
2297 kfd_doorbell_process_slice(pdd->dev->kfd)) in criu_restore_memory_of_gpu()
2768 if (pdd->dev->kfd->shared_resources.enable_mes) in runtime_enable()
2860 if (!pdd->dev->kfd->shared_resources.enable_mes) in runtime_disable()
H A Dkfd_migrate.c209 return (addr + adev->kfd.pgmap.range.start) >> PAGE_SHIFT; in svm_migrate_addr_to_pfn()
239 return (addr - adev->kfd.pgmap.range.start); in svm_migrate_addr()
1011 struct amdgpu_kfd_dev *kfddev = &adev->kfd; in kgd2kfd_init_zone_device()
H A Dkfd_events.c351 struct kfd_node *kfd; in kfd_kmap_event_page() local
367 kfd = pdd->dev; in kfd_kmap_event_page()
369 pdd = kfd_bind_process_to_device(kfd, p); in kfd_kmap_event_page()
H A Dkfd_svm.c187 bo_adev->kfd.pgmap.range.start; in svm_range_dma_map_dev()
1997 for (i = 0; i < adev->kfd.dev->num_nodes; i++) { in svm_range_set_max_pages()
1998 if (adev->kfd.dev->nodes[i]->xcp) in svm_range_set_max_pages()
1999 id = adev->kfd.dev->nodes[i]->xcp->id; in svm_range_set_max_pages()
/openbmc/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_umc.c88 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_umc_do_page_retirement()
178 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_umc_poison_handler()
H A Damdgpu_amdkfd_gpuvm.c215 (adev && xcp_id >= 0 && adev->kfd.vram_used[xcp_id] + vram_needed > in amdgpu_amdkfd_reserve_mem_limit()
227 adev->kfd.vram_used[xcp_id] += vram_needed; in amdgpu_amdkfd_reserve_mem_limit()
228 adev->kfd.vram_used_aligned[xcp_id] += adev->gmc.is_app_apu ? in amdgpu_amdkfd_reserve_mem_limit()
255 adev->kfd.vram_used[xcp_id] -= size; in amdgpu_amdkfd_unreserve_mem_limit()
257 adev->kfd.vram_used_aligned[xcp_id] -= size; in amdgpu_amdkfd_unreserve_mem_limit()
261 adev->kfd.vram_used_aligned[xcp_id] -= in amdgpu_amdkfd_unreserve_mem_limit()
273 WARN_ONCE(adev && xcp_id >= 0 && adev->kfd.vram_used[xcp_id] < 0, in amdgpu_amdkfd_unreserve_mem_limit()
1604 - adev->kfd.vram_used_aligned[xcp_id] in amdgpu_amdkfd_get_available_memory()
H A Daqua_vanjaram.c484 if (!xcp_mgr->adev->kfd.init_complete) in __aqua_vanjaram_post_partition_switch()
510 if (adev->kfd.init_complete) in aqua_vanjaram_switch_partition_mode()
H A Damdgpu_sdma.c127 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_sdma_process_ras_data_cb()
H A Damdgpu_gfx.c873 kgd2kfd_set_sram_ecc_flag(adev->kfd.dev); in amdgpu_gfx_process_ras_data_cb()
H A Damdgpu.h959 struct amdgpu_kfd_dev kfd; member
H A Damdgpu_device.c5161 if (adev->kfd.dev) in amdgpu_device_stop_pending_resets()
5162 cancel_work(&adev->kfd.reset_work); in amdgpu_device_stop_pending_resets()
5406 if (!adev->kfd.init_complete) in amdgpu_device_gpu_recover()
H A Damdgpu_drv.c2375 if (!adev->kfd.init_complete) in amdgpu_drv_delayed_reset_work_handler()
/openbmc/linux/tools/perf/util/
H A Dprobe-event.c3718 int ret, ret2, ufd = -1, kfd = -1; in del_perf_probe_events() local
3725 ret = probe_file__open_both(&kfd, &ufd, PF_FL_RW); in del_perf_probe_events()
3729 ret = probe_file__del_events(kfd, filter); in del_perf_probe_events()
3741 if (kfd >= 0) in del_perf_probe_events()
3742 close(kfd); in del_perf_probe_events()
/openbmc/linux/drivers/gpu/drm/amd/pm/swsmu/smu13/
H A Daldebaran_ppt.c1711 kgd2kfd_smi_event_throttle(smu->adev->kfd.dev, in aldebaran_log_thermal_throttling_event()
H A Dsmu_v13_0_6_ppt.c1922 smu->adev->kfd.dev, in smu_v13_0_6_log_thermal_throttling_event()
/openbmc/linux/drivers/gpu/drm/amd/pm/swsmu/smu11/
H A Darcturus_ppt.c2347 kgd2kfd_smi_event_throttle(smu->adev->kfd.dev, in arcturus_log_thermal_throttling_event()

12