Home
last modified time | relevance | path

Searched refs:adev (Results 1 – 25 of 861) sorted by relevance

12345678910>>...35

/openbmc/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_device.c415 ret = adev->pcie_rreg(adev, reg * 4); in amdgpu_device_rreg()
499 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_device_wreg()
527 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_mm_wreg_mmio_rlc()
755 return adev->nbio.funcs->get_rev_id(adev); in amdgpu_device_get_rev_id()
1040 &adev->wb.wb_obj, &adev->wb.gpu_addr, in amdgpu_device_wb_init()
2488 if (memcmp(adev->gart.ptr, adev->reset_magic, in amdgpu_device_check_vram_lost()
2612 adev = gpu_ins->adev; in amdgpu_device_enable_mgpu_fan_boost()
3393 adev->sdma_timeout = adev->video_timeout = adev->gfx_timeout; in amdgpu_device_get_job_timeout_settings()
3440 adev->sdma_timeout = adev->video_timeout = adev->gfx_timeout; in amdgpu_device_get_job_timeout_settings()
3629 adev->rmmio = ioremap(adev->rmmio_base, adev->rmmio_size); in amdgpu_device_init()
[all …]
H A Dgmc_v10_0.c334 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v10_0_flush_gpu_tlb()
524 struct amdgpu_device *adev = ring->adev; in gmc_v10_0_emit_pasid_mapping() local
819 adev->gmc.aper_base = adev->gfxhub.funcs->get_mc_fb_offset(adev); in gmc_v10_0_mc_init()
874 adev->gfxhub.funcs->init(adev); in gmc_v10_0_sw_init()
876 adev->mmhub.funcs->init(adev); in gmc_v10_0_sw_init()
1059 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v10_0_gart_enable()
1063 adev->hdp.funcs->init_registers(adev); in gmc_v10_0_gart_enable()
1066 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v10_0_gart_enable()
1111 adev->umc.funcs->init_registers(adev); in gmc_v10_0_hw_init()
1126 adev->gfxhub.funcs->gart_disable(adev); in gmc_v10_0_gart_disable()
[all …]
H A Dgmc_v9_0.c1037 struct amdgpu_device *adev = ring->adev; in gmc_v9_0_emit_flush_gpu_tlb() local
1085 struct amdgpu_device *adev = ring->adev; in gmc_v9_0_emit_pasid_mapping() local
1595 adev->smuio.funcs->get_pkg_type(adev); in gmc_v9_0_early_init()
2020 adev->gfxhub.funcs->init(adev); in gmc_v9_0_sw_init()
2022 adev->mmhub.funcs->init(adev); in gmc_v9_0_sw_init()
2312 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v9_0_gart_enable()
2346 adev->hdp.funcs->init_registers(adev); in gmc_v9_0_hw_init()
2349 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v9_0_hw_init()
2368 adev->umc.funcs->init_registers(adev); in gmc_v9_0_hw_init()
2390 adev->gfxhub.funcs->gart_disable(adev); in gmc_v9_0_gart_disable()
[all …]
H A Dgmc_v11_0.c292 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v11_0_flush_gpu_tlb()
432 struct amdgpu_device *adev = ring->adev; in gmc_v11_0_emit_pasid_mapping() local
690 adev->vm_manager.vram_base_offset = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v11_0_vram_gtt_location()
721 adev->gmc.aper_base = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v11_0_mc_init()
767 adev->mmhub.funcs->init(adev); in gmc_v11_0_sw_init()
916 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v11_0_gart_enable()
921 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v11_0_gart_enable()
949 adev->umc.funcs->init_registers(adev); in gmc_v11_0_hw_init()
963 adev->mmhub.funcs->gart_disable(adev); in gmc_v11_0_gart_disable()
976 amdgpu_irq_put(adev, &adev->gmc.vm_fault, 0); in gmc_v11_0_hw_fini()
[all …]
H A Dsoc21.c222 return adev->nbio.funcs->get_memsize(adev); in soc21_get_config_memsize()
437 adev->nbio.funcs->program_aspm(adev); in soc21_program_aspm()
574 adev->rev_id = amdgpu_device_get_rev_id(adev); in soc21_common_early_init()
620 adev->external_rev_id = adev->rev_id + 0x10; in soc21_common_early_init()
646 adev->external_rev_id = adev->rev_id + 0x1; in soc21_common_early_init()
661 adev->external_rev_id = adev->rev_id + 0x20; in soc21_common_early_init()
686 adev->external_rev_id = adev->rev_id + 0x80; in soc21_common_early_init()
762 adev->nbio.funcs->init_registers(adev); in soc21_common_hw_init()
768 adev->nbio.funcs->remap_hdp_registers(adev); in soc21_common_hw_init()
813 if (!(adev->flags & AMD_IS_APU) && adev->in_s3 && in soc21_need_reset_on_resume()
[all …]
H A Dsoc15.c320 return adev->nbio.funcs->get_memsize(adev); in soc15_get_config_memsize()
587 if (adev->flags & AMD_IS_APU && adev->in_s3 && in soc15_need_reset_on_resume()
674 adev->nbio.funcs->program_aspm(adev); in soc15_program_aspm()
1008 adev->external_rev_id = adev->rev_id + 0x14; in soc15_common_early_init()
1031 adev->external_rev_id = adev->rev_id + 0x28; in soc15_common_early_init()
1229 adev->df.funcs->sw_init(adev); in soc15_common_sw_init()
1240 adev->df.funcs->sw_fini(adev); in soc15_common_sw_fini()
1265 adev->nbio.funcs->init_registers(adev); in soc15_common_hw_init()
1408 adev->hdp.funcs->update_clock_gating(adev, in soc15_common_set_clockgating_state()
1426 adev->hdp.funcs->update_clock_gating(adev, in soc15_common_set_clockgating_state()
[all …]
H A Damdgpu_rlc.c44 if (!adev->gfx.rlc.funcs->is_rlc_enabled(adev)) in amdgpu_gfx_rlc_enter_safe_mode()
47 if (adev->cg_flags & in amdgpu_gfx_rlc_enter_safe_mode()
50 adev->gfx.rlc.funcs->set_safe_mode(adev, xcc_id); in amdgpu_gfx_rlc_enter_safe_mode()
69 if (!adev->gfx.rlc.funcs->is_rlc_enabled(adev)) in amdgpu_gfx_rlc_exit_safe_mode()
72 if (adev->cg_flags & in amdgpu_gfx_rlc_exit_safe_mode()
75 adev->gfx.rlc.funcs->unset_safe_mode(adev, xcc_id); in amdgpu_gfx_rlc_exit_safe_mode()
105 amdgpu_gfx_rlc_fini(adev); in amdgpu_gfx_rlc_init_sr()
134 adev->gfx.rlc.clear_state_size = dws = adev->gfx.rlc.funcs->get_csb_size(adev); in amdgpu_gfx_rlc_init_csb()
162 r = amdgpu_bo_create_reserved(adev, adev->gfx.rlc.cp_table_size, in amdgpu_gfx_rlc_init_cpt()
197 max_me = adev->gfx.rlc.funcs->get_cp_table_num(adev); in amdgpu_gfx_rlc_setup_cp_table()
[all …]
H A Dnv.c309 return adev->nbio.funcs->get_memsize(adev); in nv_get_config_memsize()
521 adev->nbio.funcs->program_aspm(adev); in nv_program_aspm()
615 adev->nbio.funcs->enable_aspm(adev, !enter); in nv_update_umd_stable_pstate()
693 adev->external_rev_id = adev->rev_id + 0x1; in nv_common_early_init()
714 adev->external_rev_id = adev->rev_id + 20; in nv_common_early_init()
743 adev->external_rev_id = adev->rev_id + 0xa; in nv_common_early_init()
767 adev->external_rev_id = adev->rev_id + 0x28; in nv_common_early_init()
786 adev->external_rev_id = adev->rev_id + 0x32; in nv_common_early_init()
828 adev->external_rev_id = adev->rev_id + 0x3c; in nv_common_early_init()
1011 adev->nbio.funcs->init_registers(adev); in nv_common_hw_init()
[all …]
H A Damdgpu_discovery.c372 adev, adev->mman.discovery_bin); in amdgpu_discovery_init()
379 dev_err(adev->dev, in amdgpu_discovery_init()
662 adev->umc.active_mask = ((1 << adev->umc.node_inst_num) - 1) & in amdgpu_discovery_read_from_harvest_table()
910 struct amdgpu_device *adev = ip_top->adev; in ip_disc_release() local
1110 adev->ip_top = kzalloc(sizeof(*adev->ip_top), GFP_KERNEL); in amdgpu_discovery_sysfs_init()
1111 if (!adev->ip_top) in amdgpu_discovery_sysfs_init()
1114 adev->ip_top->adev = adev; in amdgpu_discovery_sysfs_init()
1275 adev->vcn.vcn_config[adev->vcn.num_vcn_inst] = in amdgpu_discovery_reg_base_init()
1553 adev->gmc.mall_size = mall_size_per_umc * adev->gmc.num_umc; in amdgpu_discovery_get_mall_info()
2029 if (!(adev->asic_type == CHIP_VEGA20 && amdgpu_sriov_vf(adev))) in amdgpu_discovery_set_mm_ip_blocks()
[all …]
H A Damdgpu_acp.c105 adev->acp.parent = adev->dev; in acp_sw_init()
126 void *adev; member
136 adev = apd->adev; in acp_poweroff()
153 adev = apd->adev; in acp_poweron()
251 if (adev->rmmio_size == 0 || adev->rmmio_size < 0x5289) in acp_hw_init()
262 adev->acp.acp_genpd->adev = adev; in acp_hw_init()
308 adev->acp.acp_res[2].end = adev->acp.acp_res[2].start; in acp_hw_init()
312 adev->acp.acp_cell[0].resources = &adev->acp.acp_res[0]; in acp_hw_init()
313 adev->acp.acp_cell[0].platform_data = &adev->asic_type; in acp_hw_init()
416 adev->acp.acp_res[4].end = adev->acp.acp_res[4].start; in acp_hw_init()
[all …]
H A Damdgpu_virt.c66 adev->cg_flags = 0; in amdgpu_virt_init_setting()
67 adev->pg_flags = 0; in amdgpu_virt_init_setting()
232 if (!amdgpu_sriov_vf(adev) || adev->virt.mm_table.gpu_addr) in amdgpu_virt_alloc_mm_table()
260 if (!amdgpu_sriov_vf(adev) || !adev->virt.mm_table.gpu_addr) in amdgpu_virt_free_mm_table()
526 adev->unique_id = in amdgpu_virt_read_pf2vf_data()
652 if (adev->mman.fw_vram_usage_va && adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange()
654 } else if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange()
677 if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_exchange_data()
750 if (amdgpu_sriov_vf(adev) && adev->asic_type == CHIP_SIENNA_CICHLID) in amdgpu_detect_virtualization()
991 dev_err(adev->dev, in amdgpu_virt_rlcg_reg_rw()
[all …]
H A Damdgpu_ras.c141 if (adev && amdgpu_ras_get_context(adev)) in amdgpu_ras_set_error_query_ready()
147 if (adev && amdgpu_ras_get_context(adev)) in amdgpu_ras_get_error_query_ready()
663 obj->adev = adev; in amdgpu_ras_create_obj()
1068 adev->smuio.funcs->get_die_id(adev), in amdgpu_ras_query_error_status()
1088 adev->smuio.funcs->get_die_id(adev), in amdgpu_ras_query_error_status()
1340 struct amdgpu_device *adev = con->adev; in amdgpu_ras_sysfs_badpages_read() local
1669 struct amdgpu_device *adev = obj->adev; in amdgpu_ras_interrupt_poison_consumption_handler() local
2028 struct amdgpu_device *adev = ras->adev; in amdgpu_ras_do_recovery() local
2312 con->adev = adev; in amdgpu_ras_recovery_init()
2541 struct amdgpu_device *adev = con->adev; in amdgpu_ras_counte_dw() local
[all …]
H A Damdgpu_gfx.c214 int num_xcc = adev->gfx.xcc_mask ? NUM_XCC(adev->gfx.xcc_mask) : 1; in amdgpu_gfx_compute_queue_acquire()
315 ring->adev = NULL; in amdgpu_gfx_kiq_init_ring()
414 dev_warn(adev->dev, in amdgpu_gfx_mqd_sw_init()
559 if (adev->gfx.kiq[0].ring.sched.ready && !adev->job_hang) in amdgpu_gfx_disable_kgq()
804 r = amdgpu_irq_get(adev, &adev->gfx.cp_ecc_error_irq, 0); in amdgpu_gfx_ras_late_init()
826 if (!adev->gfx.ras) in amdgpu_gfx_ras_sw_init()
829 ras = adev->gfx.ras; in amdgpu_gfx_ras_sw_init()
856 if (adev->gfx.ras && adev->gfx.ras->poison_consumption_handler) in amdgpu_gfx_poison_consumption_handler()
857 return adev->gfx.ras->poison_consumption_handler(adev, entry); in amdgpu_gfx_poison_consumption_handler()
874 if (adev->gfx.ras && adev->gfx.ras->ras_block.hw_ops && in amdgpu_gfx_process_ras_data_cb()
[all …]
H A Damdgpu_bios.c105 adev->bios = NULL; in igp_read_bios_from_vram()
112 if (!adev->bios) { in igp_read_bios_from_vram()
133 adev->bios = NULL; in amdgpu_read_bios()
161 if (!adev->asic_funcs || !adev->asic_funcs->read_bios_from_rom) in amdgpu_read_bios_from_rom()
179 if (!adev->bios) { in amdgpu_read_bios_from_rom()
186 amdgpu_asic_read_bios_from_rom(adev, adev->bios, len); in amdgpu_read_bios_from_rom()
208 if (!adev->bios) in amdgpu_read_platform_bios()
468 adev->is_atom_fw = adev->asic_type >= CHIP_VEGA10; in amdgpu_get_bios()
498 adev->smuio.funcs->get_rom_index_offset(adev); in amdgpu_soc15_read_bios_from_rom()
500 adev->smuio.funcs->get_rom_data_offset(adev); in amdgpu_soc15_read_bios_from_rom()
[all …]
H A Damdgpu_amdkfd.c76 adev->kfd.dev = kgd2kfd_probe(adev, vf); in amdgpu_amdkfd_device_probe()
198 adev->kfd.init_complete = kgd2kfd_device_init(adev->kfd.dev, in amdgpu_amdkfd_device_init()
219 if (adev->kfd.dev) in amdgpu_amdkfd_interrupt()
225 if (adev->kfd.dev) in amdgpu_amdkfd_suspend()
233 if (adev->kfd.dev) in amdgpu_amdkfd_resume()
243 if (adev->kfd.dev) in amdgpu_amdkfd_pre_reset()
253 if (adev->kfd.dev) in amdgpu_amdkfd_post_reset()
427 if (adev->gmc.real_vram_size == adev->gmc.visible_vram_size) in amdgpu_amdkfd_get_local_mem_info()
457 return adev->gfx.funcs->get_gpu_clock_counter(adev); in amdgpu_amdkfd_get_gpu_clock_counter()
789 if (adev->gfx.ras && adev->gfx.ras->query_utcl2_poison_status) in amdgpu_amdkfd_ras_query_utcl2_poison_status()
[all …]
H A Dvi.c1271 !(ASICID_IS_P23(adev->pdev->device, adev->pdev->revision))) || in vi_program_aspm()
1272 ASIC_IS_P22(adev->asic_type, adev->external_rev_id)) { in vi_program_aspm()
1481 adev->rev_id = vi_get_rev_id(adev); in vi_common_early_init()
1508 adev->external_rev_id = adev->rev_id + 0x3c; in vi_common_early_init()
1525 adev->external_rev_id = adev->rev_id + 0x14; in vi_common_early_init()
1548 adev->external_rev_id = adev->rev_id + 0x5A; in vi_common_early_init()
1571 adev->external_rev_id = adev->rev_id + 0x50; in vi_common_early_init()
1594 adev->external_rev_id = adev->rev_id + 0x64; in vi_common_early_init()
1618 adev->external_rev_id = adev->rev_id + 0x6E; in vi_common_early_init()
1645 adev->external_rev_id = adev->rev_id + 0x1; in vi_common_early_init()
[all …]
H A Dmxgpu_nv.c148 r = xgpu_nv_poll_ack(adev); in xgpu_nv_mailbox_trans_msg()
348 if (amdgpu_sriov_runtime(adev) && !amdgpu_in_reset(adev)) in xgpu_nv_mailbox_rcv_irq()
350 &adev->virt.flr_work), in xgpu_nv_mailbox_rcv_irq()
390 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_nv_mailbox_add_irq_id()
394 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 138, &adev->virt.ack_irq); in xgpu_nv_mailbox_add_irq_id()
396 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_add_irq_id()
407 r = amdgpu_irq_get(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_get_irq()
410 r = amdgpu_irq_get(adev, &adev->virt.ack_irq, 0); in xgpu_nv_mailbox_get_irq()
412 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_get_irq()
423 amdgpu_irq_put(adev, &adev->virt.ack_irq, 0); in xgpu_nv_mailbox_put_irq()
[all …]
H A Damdgpu_irq.c170 ret = amdgpu_ih_process(adev, &adev->irq.ih); in amdgpu_irq_handler()
191 amdgpu_ih_process(adev, &adev->irq.ih1); in amdgpu_irq_handle_ih1()
206 amdgpu_ih_process(adev, &adev->irq.ih2); in amdgpu_irq_handle_ih2()
221 amdgpu_ih_process(adev, &adev->irq.ih_soft); in amdgpu_irq_handle_ih_soft()
314 adev->irq.irq = irq; in amdgpu_irq_init()
325 free_irq(adev->irq.irq, adev_to_drm(adev)); in amdgpu_irq_fini_hw()
331 amdgpu_ih_ring_fini(adev, &adev->irq.ih_soft); in amdgpu_irq_fini_hw()
332 amdgpu_ih_ring_fini(adev, &adev->irq.ih); in amdgpu_irq_fini_hw()
333 amdgpu_ih_ring_fini(adev, &adev->irq.ih1); in amdgpu_irq_fini_hw()
334 amdgpu_ih_ring_fini(adev, &adev->irq.ih2); in amdgpu_irq_fini_hw()
[all …]
H A Damdgpu_umc.c37 dev_warn(adev->dev, in amdgpu_umc_convert_error_address()
55 dev_warn(adev->dev, in amdgpu_umc_page_retirement_mca()
91 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops && in amdgpu_umc_do_page_retirement()
95 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops && in amdgpu_umc_do_page_retirement()
115 if (adev->umc.ras && in amdgpu_umc_do_page_retirement()
198 if (adev->virt.ops && adev->virt.ops->ras_poison_handler) in amdgpu_umc_poison_handler()
199 adev->virt.ops->ras_poison_handler(adev); in amdgpu_umc_poison_handler()
220 if (!adev->umc.ras) in amdgpu_umc_ras_sw_init()
223 ras = adev->umc.ras; in amdgpu_umc_ras_sw_init()
254 r = amdgpu_irq_get(adev, &adev->gmc.ecc_irq, 0); in amdgpu_umc_ras_late_init()
[all …]
H A Dvega20_reg_init.c29 int vega20_reg_base_init(struct amdgpu_device *adev) in vega20_reg_base_init() argument
58 void vega20_doorbell_index_init(struct amdgpu_device *adev) in vega20_doorbell_index_init() argument
60 adev->doorbell_index.kiq = AMDGPU_VEGA20_DOORBELL_KIQ; in vega20_doorbell_index_init()
61 adev->doorbell_index.mec_ring0 = AMDGPU_VEGA20_DOORBELL_MEC_RING0; in vega20_doorbell_index_init()
62 adev->doorbell_index.mec_ring1 = AMDGPU_VEGA20_DOORBELL_MEC_RING1; in vega20_doorbell_index_init()
63 adev->doorbell_index.mec_ring2 = AMDGPU_VEGA20_DOORBELL_MEC_RING2; in vega20_doorbell_index_init()
64 adev->doorbell_index.mec_ring3 = AMDGPU_VEGA20_DOORBELL_MEC_RING3; in vega20_doorbell_index_init()
65 adev->doorbell_index.mec_ring4 = AMDGPU_VEGA20_DOORBELL_MEC_RING4; in vega20_doorbell_index_init()
66 adev->doorbell_index.mec_ring5 = AMDGPU_VEGA20_DOORBELL_MEC_RING5; in vega20_doorbell_index_init()
80 adev->doorbell_index.ih = AMDGPU_VEGA20_DOORBELL_IH; in vega20_doorbell_index_init()
[all …]
H A Daqua_vanjaram.c66 return (adev->xcp_mgr->num_xcps > adev->vcn.num_vcn_inst); in aqua_vanjaram_xcp_vcn_shared()
309 struct amdgpu_device *adev = xcp_mgr->adev; in aqua_vanjaram_query_partition_mode() local
312 mode = adev->nbio.funcs->get_compute_partition_mode(adev); in aqua_vanjaram_query_partition_mode()
348 struct amdgpu_device *adev = xcp_mgr->adev; in __aqua_vanjaram_get_xcp_ip_info() local
411 struct amdgpu_device *adev = xcp_mgr->adev; in __aqua_vanjaram_get_auto_mode() local
435 struct amdgpu_device *adev = xcp_mgr->adev; in __aqua_vanjaram_is_valid_mode() local
498 adev = xcp_mgr->adev; in aqua_vanjaram_switch_partition_mode()
525 adev->gfx.funcs->switch_partition_mode(xcp_mgr->adev, in aqua_vanjaram_switch_partition_mode()
558 adev = xcp_mgr->adev; in aqua_vanjaram_get_xcp_mem_id()
637 adev->sdma.num_instances = NUM_SDMA(adev->sdma.sdma_mask); in aqua_vanjaram_init_soc_config()
[all …]
H A Dsienna_cichlid.c40 adev->pm.fw_version >= 0x3a5500 && !amdgpu_sriov_vf(adev)) in sienna_cichlid_is_mode2_default()
85 r = adev->ip_blocks[i].version->funcs->suspend(adev); in sienna_cichlid_mode2_suspend_ip()
108 adev->gfxhub.funcs->mode2_save_regs(adev); in sienna_cichlid_mode2_prepare_hwcontext()
110 adev->gfxhub.funcs->halt(adev); in sienna_cichlid_mode2_prepare_hwcontext()
150 dev_err(adev->dev, in sienna_cichlid_mode2_perform_reset()
169 adev->gfxhub.funcs->mode2_restore_regs(adev); in sienna_cichlid_mode2_restore_ip()
170 adev->gfxhub.funcs->init(adev); in sienna_cichlid_mode2_restore_ip()
171 r = adev->gfxhub.funcs->gart_enable(adev); in sienna_cichlid_mode2_restore_ip()
179 r = adev->ip_blocks[i].version->funcs->resume(adev); in sienna_cichlid_mode2_restore_ip()
197 r = adev->ip_blocks[i].version->funcs->resume(adev); in sienna_cichlid_mode2_restore_ip()
[all …]
H A Damdgpu_doorbell_mgr.c38 if (amdgpu_device_skip_hw_access(adev)) in amdgpu_mm_rdoorbell()
60 if (amdgpu_device_skip_hw_access(adev)) in amdgpu_mm_wdoorbell()
80 if (amdgpu_device_skip_hw_access(adev)) in amdgpu_mm_rdoorbell64()
161 r = amdgpu_bo_create_kernel(adev, in amdgpu_doorbell_create_kernel_doorbells()
192 if (adev->asic_type < CHIP_BONAIRE) { in amdgpu_doorbell_init()
193 adev->doorbell.base = 0; in amdgpu_doorbell_init()
194 adev->doorbell.size = 0; in amdgpu_doorbell_init()
205 adev->doorbell.base = pci_resource_start(adev->pdev, 2); in amdgpu_doorbell_init()
206 adev->doorbell.size = pci_resource_len(adev->pdev, 2); in amdgpu_doorbell_init()
208 adev->doorbell.num_kernel_doorbells = in amdgpu_doorbell_init()
[all …]
H A Dmxgpu_ai.c133 trn = xgpu_ai_peek_ack(adev); in xgpu_ai_mailbox_trans_msg()
156 r = xgpu_ai_poll_ack(adev); in xgpu_ai_mailbox_trans_msg()
320 if (amdgpu_sriov_runtime(adev) && !amdgpu_in_reset(adev)) in xgpu_ai_mailbox_rcv_irq()
365 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_ai_mailbox_add_irq_id()
369 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 138, &adev->virt.ack_irq); in xgpu_ai_mailbox_add_irq_id()
371 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_ai_mailbox_add_irq_id()
382 r = amdgpu_irq_get(adev, &adev->virt.rcv_irq, 0); in xgpu_ai_mailbox_get_irq()
385 r = amdgpu_irq_get(adev, &adev->virt.ack_irq, 0); in xgpu_ai_mailbox_get_irq()
387 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_ai_mailbox_get_irq()
398 amdgpu_irq_put(adev, &adev->virt.ack_irq, 0); in xgpu_ai_mailbox_put_irq()
[all …]
/openbmc/linux/drivers/gpu/drm/amd/pm/
H A Damdgpu_dpm.c37 ((adev)->powerplay.pp_funcs->enable_bapm((adev)->powerplay.pp_handle, (e)))
39 #define amdgpu_dpm_is_legacy_dpm(adev) ((adev)->powerplay.pp_handle == (adev)) argument
49 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_get_sclk()
65 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_get_mclk()
85 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_set_powergating_by_smu()
117 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_set_gfx_power_up_by_imu()
201 if (adev->in_s3) in amdgpu_dpm_is_baco_supported()
309 if (amdgpu_sriov_vf(adev)) in amdgpu_dpm_switch_power_profile()
431 amdgpu_dpm_enable_bapm(adev, adev->pm.ac_power); in amdgpu_pm_acpi_event_handler()
1179 adev->pm.dpm.current_ps = adev->pm.dpm.boot_ps; in amdgpu_dpm_set_sclk_od()
[all …]

12345678910>>...35