Searched refs:hive (Results 1 – 10 of 10) sorted by relevance
286 struct amdgpu_hive_info *hive = container_of( in amdgpu_xgmi_show_attrs() local290 return snprintf(buf, PAGE_SIZE, "%llu\n", hive->hive_id); in amdgpu_xgmi_show_attrs()297 struct amdgpu_hive_info *hive = container_of( in amdgpu_xgmi_hive_release() local300 amdgpu_reset_put_reset_domain(hive->reset_domain); in amdgpu_xgmi_hive_release()301 hive->reset_domain = NULL; in amdgpu_xgmi_hive_release()303 mutex_destroy(&hive->hive_lock); in amdgpu_xgmi_hive_release()304 kfree(hive); in amdgpu_xgmi_hive_release()398 struct amdgpu_hive_info *hive) in amdgpu_xgmi_sysfs_add_dev_info() argument426 if (hive->kobj.parent != (&adev->dev->kobj)) { in amdgpu_xgmi_sysfs_add_dev_info()427 ret = sysfs_create_link(&adev->dev->kobj, &hive->kobj, in amdgpu_xgmi_sysfs_add_dev_info()[all …]
57 void amdgpu_put_xgmi_hive(struct amdgpu_hive_info *hive);58 int amdgpu_xgmi_update_topology(struct amdgpu_hive_info *hive, struct amdgpu_device *adev);
161 reset_context->hive == NULL) { in aldebaran_mode2_perform_reset()344 reset_context->hive == NULL) { in aldebaran_mode2_restore_hwcontext()389 if (reset_context->hive && in aldebaran_mode2_restore_hwcontext()391 r = amdgpu_xgmi_update_topology(reset_context->hive, in aldebaran_mode2_restore_hwcontext()
2434 struct amdgpu_hive_info *hive = amdgpu_get_xgmi_hive(adev); in amdgpu_device_ip_init() local2436 if (WARN_ON(!hive)) { in amdgpu_device_ip_init()2441 if (!hive->reset_domain || in amdgpu_device_ip_init()2442 !amdgpu_reset_get_reset_domain(hive->reset_domain)) { in amdgpu_device_ip_init()2444 amdgpu_put_xgmi_hive(hive); in amdgpu_device_ip_init()2450 adev->reset_domain = hive->reset_domain; in amdgpu_device_ip_init()2451 amdgpu_put_xgmi_hive(hive); in amdgpu_device_ip_init()3347 struct amdgpu_hive_info *hive = amdgpu_get_xgmi_hive(adev); in amdgpu_device_xgmi_reset_func() local3350 if (WARN_ON(!hive)) in amdgpu_device_xgmi_reset_func()3361 task_barrier_enter(&hive->tb); in amdgpu_device_xgmi_reset_func()[all …]
40 struct amdgpu_hive_info *hive; member
2251 struct amdgpu_hive_info *hive; in amdgpu_pci_remove() local2253 hive = amdgpu_get_xgmi_hive(adev); in amdgpu_pci_remove()2254 if (hive->device_remove_count == 0) in amdgpu_pci_remove()2256 hive->device_remove_count++; in amdgpu_pci_remove()2257 amdgpu_put_xgmi_hive(hive); in amdgpu_pci_remove()
2035 struct amdgpu_hive_info *hive = amdgpu_get_xgmi_hive(adev); in amdgpu_ras_do_recovery() local2038 if (hive && adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_do_recovery()2039 device_list_handle = &hive->device_list; in amdgpu_ras_do_recovery()2052 amdgpu_put_xgmi_hive(hive); in amdgpu_ras_do_recovery()
1332 struct amdgpu_hive_info *hive; in psp_xgmi_reflect_topology_info() local1338 hive = amdgpu_get_xgmi_hive(psp->adev); in psp_xgmi_reflect_topology_info()1339 if (WARN_ON(!hive)) in psp_xgmi_reflect_topology_info()1342 list_for_each_entry(mirror_adev, &hive->device_list, gmc.xgmi.head) { in psp_xgmi_reflect_topology_info()1369 amdgpu_put_xgmi_hive(hive); in psp_xgmi_reflect_topology_info()
776 struct amdgpu_hive_info *hive; member
40 ((adev)->hive ? (void *)(adev)->hive : (void *)(adev))