Home
last modified time | relevance | path

Searched refs:adev (Results 1 – 25 of 471) sorted by relevance

12345678910>>...19

/openbsd/sys/dev/pci/drm/amd/amdgpu/
H A Damdgpu_device.c415 ret = adev->pcie_rreg(adev, reg * 4); in amdgpu_device_rreg()
499 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_device_wreg()
527 adev->pcie_wreg(adev, reg * 4, v); in amdgpu_mm_wreg_mmio_rlc()
755 return adev->nbio.funcs->get_rev_id(adev); in amdgpu_device_get_rev_id()
1044 &adev->wb.wb_obj, &adev->wb.gpu_addr, in amdgpu_device_wb_init()
2516 if (memcmp(adev->gart.ptr, adev->reset_magic, in amdgpu_device_check_vram_lost()
2640 adev = gpu_ins->adev; in amdgpu_device_enable_mgpu_fan_boost()
3421 adev->sdma_timeout = adev->video_timeout = adev->gfx_timeout; in amdgpu_device_get_job_timeout_settings()
3469 adev->sdma_timeout = adev->video_timeout = adev->gfx_timeout; in amdgpu_device_get_job_timeout_settings()
3664 adev->rmmio = ioremap(adev->rmmio_base, adev->rmmio_size); in amdgpu_device_init()
[all …]
H A Dgmc_v10_0.c334 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v10_0_flush_gpu_tlb()
524 struct amdgpu_device *adev = ring->adev; in gmc_v10_0_emit_pasid_mapping() local
815 adev->gmc.aper_size = adev->fb_aper_size; in gmc_v10_0_mc_init()
819 adev->gmc.aper_base = adev->gfxhub.funcs->get_mc_fb_offset(adev); in gmc_v10_0_mc_init()
874 adev->gfxhub.funcs->init(adev); in gmc_v10_0_sw_init()
876 adev->mmhub.funcs->init(adev); in gmc_v10_0_sw_init()
1063 adev->hdp.funcs->init_registers(adev); in gmc_v10_0_gart_enable()
1066 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v10_0_gart_enable()
1111 adev->umc.funcs->init_registers(adev); in gmc_v10_0_hw_init()
1126 adev->gfxhub.funcs->gart_disable(adev); in gmc_v10_0_gart_disable()
[all …]
H A Dgmc_v9_0.c1037 struct amdgpu_device *adev = ring->adev; in gmc_v9_0_emit_flush_gpu_tlb() local
1085 struct amdgpu_device *adev = ring->adev; in gmc_v9_0_emit_pasid_mapping() local
1597 adev->smuio.funcs->get_pkg_type(adev); in gmc_v9_0_early_init()
1726 adev->gmc.aper_size = adev->fb_aper_size; in gmc_v9_0_mc_init()
2027 adev->gfxhub.funcs->init(adev); in gmc_v9_0_sw_init()
2029 adev->mmhub.funcs->init(adev); in gmc_v9_0_sw_init()
2353 adev->hdp.funcs->init_registers(adev); in gmc_v9_0_hw_init()
2356 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v9_0_hw_init()
2375 adev->umc.funcs->init_registers(adev); in gmc_v9_0_hw_init()
2397 adev->gfxhub.funcs->gart_disable(adev); in gmc_v9_0_gart_disable()
[all …]
H A Dgmc_v11_0.c292 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v11_0_flush_gpu_tlb()
432 struct amdgpu_device *adev = ring->adev; in gmc_v11_0_emit_pasid_mapping() local
690 adev->vm_manager.vram_base_offset = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v11_0_vram_gtt_location()
716 adev->gmc.aper_base = adev->fb_aper_offset; in gmc_v11_0_mc_init()
717 adev->gmc.aper_size = adev->fb_aper_size; in gmc_v11_0_mc_init()
721 adev->gmc.aper_base = adev->mmhub.funcs->get_mc_fb_offset(adev); in gmc_v11_0_mc_init()
767 adev->mmhub.funcs->init(adev); in gmc_v11_0_sw_init()
916 r = adev->mmhub.funcs->gart_enable(adev); in gmc_v11_0_gart_enable()
921 adev->hdp.funcs->flush_hdp(adev, NULL); in gmc_v11_0_gart_enable()
949 adev->umc.funcs->init_registers(adev); in gmc_v11_0_hw_init()
[all …]
H A Dsoc21.c222 return adev->nbio.funcs->get_memsize(adev); in soc21_get_config_memsize()
437 adev->nbio.funcs->program_aspm(adev); in soc21_program_aspm()
574 adev->rev_id = amdgpu_device_get_rev_id(adev); in soc21_common_early_init()
620 adev->external_rev_id = adev->rev_id + 0x10; in soc21_common_early_init()
646 adev->external_rev_id = adev->rev_id + 0x1; in soc21_common_early_init()
661 adev->external_rev_id = adev->rev_id + 0x20; in soc21_common_early_init()
686 adev->external_rev_id = adev->rev_id + 0x80; in soc21_common_early_init()
762 adev->nbio.funcs->init_registers(adev); in soc21_common_hw_init()
768 adev->nbio.funcs->remap_hdp_registers(adev); in soc21_common_hw_init()
813 if (!(adev->flags & AMD_IS_APU) && adev->in_s3 && in soc21_need_reset_on_resume()
[all …]
H A Dsoc15.c320 return adev->nbio.funcs->get_memsize(adev); in soc15_get_config_memsize()
587 if (adev->flags & AMD_IS_APU && adev->in_s3 && in soc15_need_reset_on_resume()
674 adev->nbio.funcs->program_aspm(adev); in soc15_program_aspm()
1008 adev->external_rev_id = adev->rev_id + 0x14; in soc15_common_early_init()
1031 adev->external_rev_id = adev->rev_id + 0x28; in soc15_common_early_init()
1229 adev->df.funcs->sw_init(adev); in soc15_common_sw_init()
1240 adev->df.funcs->sw_fini(adev); in soc15_common_sw_fini()
1265 adev->nbio.funcs->init_registers(adev); in soc15_common_hw_init()
1408 adev->hdp.funcs->update_clock_gating(adev, in soc15_common_set_clockgating_state()
1426 adev->hdp.funcs->update_clock_gating(adev, in soc15_common_set_clockgating_state()
[all …]
H A Damdgpu_rlc.c44 if (!adev->gfx.rlc.funcs->is_rlc_enabled(adev)) in amdgpu_gfx_rlc_enter_safe_mode()
47 if (adev->cg_flags & in amdgpu_gfx_rlc_enter_safe_mode()
50 adev->gfx.rlc.funcs->set_safe_mode(adev, xcc_id); in amdgpu_gfx_rlc_enter_safe_mode()
69 if (!adev->gfx.rlc.funcs->is_rlc_enabled(adev)) in amdgpu_gfx_rlc_exit_safe_mode()
72 if (adev->cg_flags & in amdgpu_gfx_rlc_exit_safe_mode()
75 adev->gfx.rlc.funcs->unset_safe_mode(adev, xcc_id); in amdgpu_gfx_rlc_exit_safe_mode()
105 amdgpu_gfx_rlc_fini(adev); in amdgpu_gfx_rlc_init_sr()
134 adev->gfx.rlc.clear_state_size = dws = adev->gfx.rlc.funcs->get_csb_size(adev); in amdgpu_gfx_rlc_init_csb()
162 r = amdgpu_bo_create_reserved(adev, adev->gfx.rlc.cp_table_size, in amdgpu_gfx_rlc_init_cpt()
197 max_me = adev->gfx.rlc.funcs->get_cp_table_num(adev); in amdgpu_gfx_rlc_setup_cp_table()
[all …]
H A Dnv.c309 return adev->nbio.funcs->get_memsize(adev); in nv_get_config_memsize()
521 adev->nbio.funcs->program_aspm(adev); in nv_program_aspm()
615 adev->nbio.funcs->enable_aspm(adev, !enter); in nv_update_umd_stable_pstate()
693 adev->external_rev_id = adev->rev_id + 0x1; in nv_common_early_init()
714 adev->external_rev_id = adev->rev_id + 20; in nv_common_early_init()
743 adev->external_rev_id = adev->rev_id + 0xa; in nv_common_early_init()
767 adev->external_rev_id = adev->rev_id + 0x28; in nv_common_early_init()
786 adev->external_rev_id = adev->rev_id + 0x32; in nv_common_early_init()
828 adev->external_rev_id = adev->rev_id + 0x3c; in nv_common_early_init()
1011 adev->nbio.funcs->init_registers(adev); in nv_common_hw_init()
[all …]
H A Damdgpu_discovery.c376 adev, adev->mman.discovery_bin); in amdgpu_discovery_init()
383 dev_err(adev->dev, in amdgpu_discovery_init()
666 adev->umc.active_mask = ((1 << adev->umc.node_inst_num) - 1) & in amdgpu_discovery_read_from_harvest_table()
916 struct amdgpu_device *adev = ip_top->adev; in ip_disc_release() local
1120 adev->ip_top = kzalloc(sizeof(*adev->ip_top), GFP_KERNEL); in amdgpu_discovery_sysfs_init()
1121 if (!adev->ip_top) in amdgpu_discovery_sysfs_init()
1124 adev->ip_top->adev = adev; in amdgpu_discovery_sysfs_init()
1292 adev->vcn.vcn_config[adev->vcn.num_vcn_inst] = in amdgpu_discovery_reg_base_init()
1570 adev->gmc.mall_size = mall_size_per_umc * adev->gmc.num_umc; in amdgpu_discovery_get_mall_info()
2046 if (!(adev->asic_type == CHIP_VEGA20 && amdgpu_sriov_vf(adev))) in amdgpu_discovery_set_mm_ip_blocks()
[all …]
H A Damdgpu_bios.c214 address = pci_conf_read(adev->pc, adev->pa_tag, PCI_ROM_REG); in amdgpu_read_bios()
216 mask = pci_conf_read(adev->pc, adev->pa_tag, PCI_ROM_REG); in amdgpu_read_bios()
218 pci_conf_write(adev->pc, adev->pa_tag, PCI_ROM_REG, address); in amdgpu_read_bios()
231 bus_space_read_region_1(adev->memt, romh, 0, adev->bios, size); in amdgpu_read_bios()
248 if (!adev->asic_funcs || !adev->asic_funcs->read_bios_from_rom) in amdgpu_read_bios_from_rom()
273 amdgpu_asic_read_bios_from_rom(adev, adev->bios, len); in amdgpu_read_bios_from_rom()
296 if (!adev->bios) in amdgpu_read_platform_bios()
599 adev->is_atom_fw = adev->asic_type >= CHIP_VEGA10; in amdgpu_get_bios()
629 adev->smuio.funcs->get_rom_index_offset(adev); in amdgpu_soc15_read_bios_from_rom()
631 adev->smuio.funcs->get_rom_data_offset(adev); in amdgpu_soc15_read_bios_from_rom()
[all …]
H A Damdgpu_acp.c105 adev->acp.parent = adev->dev; in acp_sw_init()
126 void *adev; member
136 adev = apd->adev; in acp_poweroff()
153 adev = apd->adev; in acp_poweron()
251 if (adev->rmmio_size == 0 || adev->rmmio_size < 0x5289) in acp_hw_init()
262 adev->acp.acp_genpd->adev = adev; in acp_hw_init()
308 adev->acp.acp_res[2].end = adev->acp.acp_res[2].start; in acp_hw_init()
312 adev->acp.acp_cell[0].resources = &adev->acp.acp_res[0]; in acp_hw_init()
313 adev->acp.acp_cell[0].platform_data = &adev->asic_type; in acp_hw_init()
416 adev->acp.acp_res[4].end = adev->acp.acp_res[4].start; in acp_hw_init()
[all …]
H A Damdgpu_virt.c66 adev->cg_flags = 0; in amdgpu_virt_init_setting()
67 adev->pg_flags = 0; in amdgpu_virt_init_setting()
232 if (!amdgpu_sriov_vf(adev) || adev->virt.mm_table.gpu_addr) in amdgpu_virt_alloc_mm_table()
260 if (!amdgpu_sriov_vf(adev) || !adev->virt.mm_table.gpu_addr) in amdgpu_virt_free_mm_table()
526 adev->unique_id = in amdgpu_virt_read_pf2vf_data()
653 if (adev->mman.fw_vram_usage_va && adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange()
655 } else if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange()
678 if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_exchange_data()
751 if (amdgpu_sriov_vf(adev) && adev->asic_type == CHIP_SIENNA_CICHLID) in amdgpu_detect_virtualization()
992 dev_err(adev->dev, in amdgpu_virt_rlcg_reg_rw()
[all …]
H A Damdgpu_ras.c141 if (adev && amdgpu_ras_get_context(adev)) in amdgpu_ras_set_error_query_ready()
147 if (adev && amdgpu_ras_get_context(adev)) in amdgpu_ras_get_error_query_ready()
667 obj->adev = adev; in amdgpu_ras_create_obj()
1072 adev->smuio.funcs->get_die_id(adev), in amdgpu_ras_query_error_status()
1092 adev->smuio.funcs->get_die_id(adev), in amdgpu_ras_query_error_status()
1345 struct amdgpu_device *adev = con->adev; in amdgpu_ras_sysfs_badpages_read() local
1689 struct amdgpu_device *adev = obj->adev; in amdgpu_ras_interrupt_poison_consumption_handler() local
2048 struct amdgpu_device *adev = ras->adev; in amdgpu_ras_do_recovery() local
2332 con->adev = adev; in amdgpu_ras_recovery_init()
2561 struct amdgpu_device *adev = con->adev; in amdgpu_ras_counte_dw() local
[all …]
H A Damdgpu_gfx.c216 int num_xcc = adev->gfx.xcc_mask ? NUM_XCC(adev->gfx.xcc_mask) : 1; in amdgpu_gfx_compute_queue_acquire()
317 ring->adev = NULL; in amdgpu_gfx_kiq_init_ring()
416 dev_warn(adev->dev, in amdgpu_gfx_mqd_sw_init()
561 if (adev->gfx.kiq[0].ring.sched.ready && !adev->job_hang) in amdgpu_gfx_disable_kgq()
806 r = amdgpu_irq_get(adev, &adev->gfx.cp_ecc_error_irq, 0); in amdgpu_gfx_ras_late_init()
828 if (!adev->gfx.ras) in amdgpu_gfx_ras_sw_init()
831 ras = adev->gfx.ras; in amdgpu_gfx_ras_sw_init()
859 if (adev->gfx.ras && adev->gfx.ras->poison_consumption_handler) in amdgpu_gfx_poison_consumption_handler()
860 return adev->gfx.ras->poison_consumption_handler(adev, entry); in amdgpu_gfx_poison_consumption_handler()
877 if (adev->gfx.ras && adev->gfx.ras->ras_block.hw_ops && in amdgpu_gfx_process_ras_data_cb()
[all …]
H A Damdgpu_amdkfd.c81 adev->kfd.dev = kgd2kfd_probe(adev, vf); in amdgpu_amdkfd_device_probe()
203 adev->kfd.init_complete = kgd2kfd_device_init(adev->kfd.dev, in amdgpu_amdkfd_device_init()
224 if (adev->kfd.dev) in amdgpu_amdkfd_interrupt()
230 if (adev->kfd.dev) in amdgpu_amdkfd_suspend()
238 if (adev->kfd.dev) in amdgpu_amdkfd_resume()
248 if (adev->kfd.dev) in amdgpu_amdkfd_pre_reset()
258 if (adev->kfd.dev) in amdgpu_amdkfd_post_reset()
432 if (adev->gmc.real_vram_size == adev->gmc.visible_vram_size) in amdgpu_amdkfd_get_local_mem_info()
462 return adev->gfx.funcs->get_gpu_clock_counter(adev); in amdgpu_amdkfd_get_gpu_clock_counter()
794 if (adev->gfx.ras && adev->gfx.ras->query_utcl2_poison_status) in amdgpu_amdkfd_ras_query_utcl2_poison_status()
[all …]
H A Dvi.c1271 !(ASICID_IS_P23(adev->pdev->device, adev->pdev->revision))) || in vi_program_aspm()
1272 ASIC_IS_P22(adev->asic_type, adev->external_rev_id)) { in vi_program_aspm()
1481 adev->rev_id = vi_get_rev_id(adev); in vi_common_early_init()
1508 adev->external_rev_id = adev->rev_id + 0x3c; in vi_common_early_init()
1525 adev->external_rev_id = adev->rev_id + 0x14; in vi_common_early_init()
1548 adev->external_rev_id = adev->rev_id + 0x5A; in vi_common_early_init()
1571 adev->external_rev_id = adev->rev_id + 0x50; in vi_common_early_init()
1594 adev->external_rev_id = adev->rev_id + 0x64; in vi_common_early_init()
1618 adev->external_rev_id = adev->rev_id + 0x6E; in vi_common_early_init()
1645 adev->external_rev_id = adev->rev_id + 0x1; in vi_common_early_init()
[all …]
H A Damdgpu_irq.c173 ret = amdgpu_ih_process(adev, &adev->irq.ih); in amdgpu_irq_handler()
194 amdgpu_ih_process(adev, &adev->irq.ih1); in amdgpu_irq_handle_ih1()
209 amdgpu_ih_process(adev, &adev->irq.ih2); in amdgpu_irq_handle_ih2()
224 amdgpu_ih_process(adev, &adev->irq.ih_soft); in amdgpu_irq_handle_ih_soft()
322 adev->irq.irq = irq; in amdgpu_irq_init()
333 free_irq(adev->irq.irq, adev_to_drm(adev)); in amdgpu_irq_fini_hw()
339 amdgpu_ih_ring_fini(adev, &adev->irq.ih_soft); in amdgpu_irq_fini_hw()
340 amdgpu_ih_ring_fini(adev, &adev->irq.ih); in amdgpu_irq_fini_hw()
341 amdgpu_ih_ring_fini(adev, &adev->irq.ih1); in amdgpu_irq_fini_hw()
342 amdgpu_ih_ring_fini(adev, &adev->irq.ih2); in amdgpu_irq_fini_hw()
[all …]
H A Daqua_vanjaram.c66 return (adev->xcp_mgr->num_xcps > adev->vcn.num_vcn_inst); in aqua_vanjaram_xcp_vcn_shared()
309 struct amdgpu_device *adev = xcp_mgr->adev; in aqua_vanjaram_query_partition_mode() local
312 mode = adev->nbio.funcs->get_compute_partition_mode(adev); in aqua_vanjaram_query_partition_mode()
348 struct amdgpu_device *adev = xcp_mgr->adev; in __aqua_vanjaram_get_xcp_ip_info() local
411 struct amdgpu_device *adev = xcp_mgr->adev; in __aqua_vanjaram_get_auto_mode() local
435 struct amdgpu_device *adev = xcp_mgr->adev; in __aqua_vanjaram_is_valid_mode() local
498 adev = xcp_mgr->adev; in aqua_vanjaram_switch_partition_mode()
525 adev->gfx.funcs->switch_partition_mode(xcp_mgr->adev, in aqua_vanjaram_switch_partition_mode()
558 adev = xcp_mgr->adev; in aqua_vanjaram_get_xcp_mem_id()
637 adev->sdma.num_instances = NUM_SDMA(adev->sdma.sdma_mask); in aqua_vanjaram_init_soc_config()
[all …]
H A Dmxgpu_nv.c148 r = xgpu_nv_poll_ack(adev); in xgpu_nv_mailbox_trans_msg()
348 if (amdgpu_sriov_runtime(adev) && !amdgpu_in_reset(adev)) in xgpu_nv_mailbox_rcv_irq()
350 &adev->virt.flr_work), in xgpu_nv_mailbox_rcv_irq()
390 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_nv_mailbox_add_irq_id()
394 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 138, &adev->virt.ack_irq); in xgpu_nv_mailbox_add_irq_id()
396 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_add_irq_id()
407 r = amdgpu_irq_get(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_get_irq()
410 r = amdgpu_irq_get(adev, &adev->virt.ack_irq, 0); in xgpu_nv_mailbox_get_irq()
412 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_nv_mailbox_get_irq()
423 amdgpu_irq_put(adev, &adev->virt.ack_irq, 0); in xgpu_nv_mailbox_put_irq()
[all …]
H A Damdgpu_umc.c37 dev_warn(adev->dev, in amdgpu_umc_convert_error_address()
55 dev_warn(adev->dev, in amdgpu_umc_page_retirement_mca()
91 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops && in amdgpu_umc_do_page_retirement()
95 if (adev->umc.ras && adev->umc.ras->ras_block.hw_ops && in amdgpu_umc_do_page_retirement()
115 if (adev->umc.ras && in amdgpu_umc_do_page_retirement()
198 if (adev->virt.ops && adev->virt.ops->ras_poison_handler) in amdgpu_umc_poison_handler()
199 adev->virt.ops->ras_poison_handler(adev); in amdgpu_umc_poison_handler()
220 if (!adev->umc.ras) in amdgpu_umc_ras_sw_init()
223 ras = adev->umc.ras; in amdgpu_umc_ras_sw_init()
255 r = amdgpu_irq_get(adev, &adev->gmc.ecc_irq, 0); in amdgpu_umc_ras_late_init()
[all …]
H A Damdgpu_doorbell_mgr.c38 if (amdgpu_device_skip_hw_access(adev)) in amdgpu_mm_rdoorbell()
60 if (amdgpu_device_skip_hw_access(adev)) in amdgpu_mm_wdoorbell()
161 r = amdgpu_bo_create_kernel(adev, in amdgpu_doorbell_create_kernel_doorbells()
193 adev->doorbell.base = 0; in amdgpu_doorbell_init()
194 adev->doorbell.size = 0; in amdgpu_doorbell_init()
208 adev->doorbell.base = pci_resource_start(adev->pdev, 2); in amdgpu_doorbell_init()
209 adev->doorbell.size = pci_resource_len(adev->pdev, 2); in amdgpu_doorbell_init()
213 mtype = pci_mapreg_type(adev->pdev->pc, adev->pdev->tag, 0x18); in amdgpu_doorbell_init()
214 if (pci_mapreg_info(adev->pdev->pc, adev->pdev->tag, 0x18, in amdgpu_doorbell_init()
215 mtype, &adev->doorbell.base, &adev->doorbell.size, NULL)) in amdgpu_doorbell_init()
[all …]
H A Dvega20_reg_init.c29 int vega20_reg_base_init(struct amdgpu_device *adev) in vega20_reg_base_init() argument
58 void vega20_doorbell_index_init(struct amdgpu_device *adev) in vega20_doorbell_index_init() argument
60 adev->doorbell_index.kiq = AMDGPU_VEGA20_DOORBELL_KIQ; in vega20_doorbell_index_init()
61 adev->doorbell_index.mec_ring0 = AMDGPU_VEGA20_DOORBELL_MEC_RING0; in vega20_doorbell_index_init()
62 adev->doorbell_index.mec_ring1 = AMDGPU_VEGA20_DOORBELL_MEC_RING1; in vega20_doorbell_index_init()
63 adev->doorbell_index.mec_ring2 = AMDGPU_VEGA20_DOORBELL_MEC_RING2; in vega20_doorbell_index_init()
64 adev->doorbell_index.mec_ring3 = AMDGPU_VEGA20_DOORBELL_MEC_RING3; in vega20_doorbell_index_init()
65 adev->doorbell_index.mec_ring4 = AMDGPU_VEGA20_DOORBELL_MEC_RING4; in vega20_doorbell_index_init()
66 adev->doorbell_index.mec_ring5 = AMDGPU_VEGA20_DOORBELL_MEC_RING5; in vega20_doorbell_index_init()
80 adev->doorbell_index.ih = AMDGPU_VEGA20_DOORBELL_IH; in vega20_doorbell_index_init()
[all …]
H A Dsienna_cichlid.c40 adev->pm.fw_version >= 0x3a5500 && !amdgpu_sriov_vf(adev)) in sienna_cichlid_is_mode2_default()
85 r = adev->ip_blocks[i].version->funcs->suspend(adev); in sienna_cichlid_mode2_suspend_ip()
108 adev->gfxhub.funcs->mode2_save_regs(adev); in sienna_cichlid_mode2_prepare_hwcontext()
110 adev->gfxhub.funcs->halt(adev); in sienna_cichlid_mode2_prepare_hwcontext()
150 dev_err(adev->dev, in sienna_cichlid_mode2_perform_reset()
169 adev->gfxhub.funcs->mode2_restore_regs(adev); in sienna_cichlid_mode2_restore_ip()
170 adev->gfxhub.funcs->init(adev); in sienna_cichlid_mode2_restore_ip()
171 r = adev->gfxhub.funcs->gart_enable(adev); in sienna_cichlid_mode2_restore_ip()
179 r = adev->ip_blocks[i].version->funcs->resume(adev); in sienna_cichlid_mode2_restore_ip()
197 r = adev->ip_blocks[i].version->funcs->resume(adev); in sienna_cichlid_mode2_restore_ip()
[all …]
H A Dgmc_v6_0.c135 err = amdgpu_ucode_request(adev, &adev->gmc.fw, fw_name); in gmc_v6_0_init_microcode()
137 dev_err(adev->dev, in gmc_v6_0_init_microcode()
153 if (!adev->gmc.fw) in gmc_v6_0_mc_load_microcode()
319 adev->gmc.aper_base = adev->fb_aper_offset; in gmc_v6_0_mc_init()
320 adev->gmc.aper_size = adev->fb_aper_size; in gmc_v6_0_mc_init()
321 adev->gmc.visible_vram_size = adev->gmc.aper_size; in gmc_v6_0_mc_init()
341 adev->gmc.gart_size += adev->pm.smu_prv_buffer_size; in gmc_v6_0_mc_init()
342 gmc_v6_0_vram_gtt_location(adev, &adev->gmc); in gmc_v6_0_mc_init()
565 adev->gart.table_size = adev->gart.num_gpu_pages * 8; in gmc_v6_0_gart_init()
784 return amdgpu_irq_get(adev, &adev->gmc.vm_fault, 0); in gmc_v6_0_late_init()
[all …]
/openbsd/sys/dev/pci/drm/amd/pm/
H A Damdgpu_dpm.c37 ((adev)->powerplay.pp_funcs->enable_bapm((adev)->powerplay.pp_handle, (e)))
39 #define amdgpu_dpm_is_legacy_dpm(adev) ((adev)->powerplay.pp_handle == (adev)) argument
49 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_get_sclk()
65 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_get_mclk()
85 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_set_powergating_by_smu()
117 mutex_lock(&adev->pm.mutex); in amdgpu_dpm_set_gfx_power_up_by_imu()
201 if (adev->in_s3) in amdgpu_dpm_is_baco_supported()
309 if (amdgpu_sriov_vf(adev)) in amdgpu_dpm_switch_power_profile()
431 amdgpu_dpm_enable_bapm(adev, adev->pm.ac_power); in amdgpu_pm_acpi_event_handler()
1179 adev->pm.dpm.current_ps = adev->pm.dpm.boot_ps; in amdgpu_dpm_set_sclk_od()
[all …]

12345678910>>...19