/dragonfly/sys/dev/drm/i915/ |
H A D | intel_guc_log.c | 190 relay_reserve(guc->log.runtime.relay_chan, guc->log.vma->obj->base.size); 337 guc_move_to_next_buf(guc); 349 struct intel_guc *guc = 352 guc_log_capture_logs(guc); 501 guc_log_flush_complete(guc); 522 guc_log_flush(guc); 525 guc_log_capture_logs(guc); 537 GEM_BUG_ON(guc->log.vma); in intel_guc_log_create() 563 guc->log.vma = vma; in intel_guc_log_create() 605 struct intel_guc *guc = &dev_priv->guc; [all …]
|
H A D | intel_guc.c | 37 GEM_BUG_ON(!guc->send_regs.base); in guc_send_reg() 38 GEM_BUG_ON(!guc->send_regs.count); in guc_send_reg() 55 guc_send_reg(guc, i), in intel_guc_init_send_regs() 66 guc->send = intel_guc_send_nop; in intel_guc_init_early() 67 guc->notify = gen8_guc_raise_irq; in intel_guc_init_early() 187 mutex_lock(&guc->send_mutex); in intel_guc_send_mmio() 195 intel_guc_notify(guc); in intel_guc_send_mmio() 202 guc_send_reg(guc, 0), in intel_guc_send_mmio() 221 mutex_unlock(&guc->send_mutex); in intel_guc_send_mmio() 270 struct intel_guc *guc = &dev_priv->guc; in intel_guc_suspend() local [all …]
|
H A D | intel_uc.c | 119 if (!guc->load_err_log) in guc_capture_load_err_log() 120 guc->load_err_log = i915_gem_object_get(guc->log.vma->obj); in guc_capture_load_err_log() 127 if (guc->load_err_log) in guc_free_load_err_log() 147 intel_guc_disable_ct(guc); in guc_disable_communication() 149 guc->send = intel_guc_send_nop; in guc_disable_communication() 154 struct intel_guc *guc = &dev_priv->guc; in intel_uc_init_hw() local 160 guc_disable_communication(guc); in intel_uc_init_hw() 198 intel_guc_init_params(guc); in intel_uc_init_hw() 228 guc->fw.path, in intel_uc_init_hw() 229 guc->fw.major_ver_found, guc->fw.minor_ver_found); in intel_uc_init_hw() [all …]
|
H A D | i915_guc_submission.c | 496 struct intel_guc *guc = &dev_priv->guc; in i915_guc_submit() local 766 struct intel_guc *guc = &dev_priv->guc; in guc_client_alloc() local 775 client->guc = guc; in guc_client_alloc() 912 GEM_BUG_ON(guc->ads_vma); in guc_ads_create() 918 guc->ads_vma = vma; in guc_ads_create() 976 struct intel_guc *guc = &dev_priv->guc; in i915_guc_submission_init() local 981 if (guc->stage_desc_pool) in i915_guc_submission_init() 1023 struct intel_guc *guc = &dev_priv->guc; in i915_guc_submission_fini() local 1026 guc_ads_destroy(guc); in i915_guc_submission_fini() 1102 struct intel_guc *guc = &dev_priv->guc; in i915_guc_submission_enable() local [all …]
|
H A D | intel_guc_ct.c | 198 err = ctch_init(guc, ctch); in ctch_open() 235 guc_action_deregister_ct_buffer(guc, in ctch_open() 239 ctch_fini(guc, ctch); in ctch_open() 256 ctch_fini(guc, ctch); in ctch_close() 385 intel_guc_notify(guc); in ctch_send() 404 mutex_lock(&guc->send_mutex); in intel_guc_send_ct() 412 mutex_unlock(&guc->send_mutex); in intel_guc_send_ct() 431 err = ctch_open(guc, ctch); in intel_guc_enable_ct() 436 guc->send = intel_guc_send_ct; in intel_guc_enable_ct() 456 ctch_close(guc, ctch); in intel_guc_disable_ct() [all …]
|
H A D | intel_guc.h | 75 int (*send)(struct intel_guc *guc, const u32 *data, u32 len); 78 void (*notify)(struct intel_guc *guc); 84 return guc->send(guc, action, len); in intel_guc_send() 87 static inline void intel_guc_notify(struct intel_guc *guc) in intel_guc_notify() argument 89 guc->notify(guc); in intel_guc_notify() 108 void intel_guc_init_early(struct intel_guc *guc); 109 void intel_guc_init_send_regs(struct intel_guc *guc); 110 void intel_guc_init_params(struct intel_guc *guc); 113 int intel_guc_sample_forcewake(struct intel_guc *guc); 114 int intel_guc_auth_huc(struct intel_guc *guc, u32 rsa_offset); [all …]
|
H A D | intel_guc_fw.c | 69 guc->fw.major_ver_wanted = 0; in intel_guc_fw_select() 70 guc->fw.minor_ver_wanted = 0; in intel_guc_fw_select() 72 guc->fw.path = I915_SKL_GUC_UCODE; in intel_guc_fw_select() 73 guc->fw.major_ver_wanted = SKL_FW_MAJOR; in intel_guc_fw_select() 74 guc->fw.minor_ver_wanted = SKL_FW_MINOR; in intel_guc_fw_select() 76 guc->fw.path = I915_BXT_GUC_UCODE; in intel_guc_fw_select() 77 guc->fw.major_ver_wanted = BXT_FW_MAJOR; in intel_guc_fw_select() 78 guc->fw.minor_ver_wanted = BXT_FW_MINOR; in intel_guc_fw_select() 80 guc->fw.path = I915_KBL_GUC_UCODE; in intel_guc_fw_select() 81 guc->fw.major_ver_wanted = KBL_FW_MAJOR; in intel_guc_fw_select() [all …]
|
H A D | intel_guc_fw.h | 30 int intel_guc_fw_select(struct intel_guc *guc); 31 int intel_guc_fw_upload(struct intel_guc *guc);
|
H A D | intel_guc_ct.h | 83 int intel_guc_enable_ct(struct intel_guc *guc); 84 void intel_guc_disable_ct(struct intel_guc *guc);
|
H A D | intel_guc_log.h | 53 int intel_guc_log_create(struct intel_guc *guc); 54 void intel_guc_log_destroy(struct intel_guc *guc);
|
H A D | intel_huc.c | 186 struct intel_guc *guc = &i915->guc; in intel_huc_auth() local 201 ret = intel_guc_auth_huc(guc, in intel_huc_auth()
|
H A D | i915_guc_submission.h | 59 struct intel_guc *guc; member
|
H A D | intel_uc_fw.c | 134 uc_fw->major_ver_found = css->guc.sw_version >> 16; in intel_uc_fw_fetch() 135 uc_fw->minor_ver_found = css->guc.sw_version & 0xFFFF; in intel_uc_fw_fetch()
|
H A D | intel_guc_fwif.h | 215 } guc; member
|
H A D | i915_irq.c | 463 if (!dev_priv->guc.interrupts_enabled) { in gen9_enable_guc_interrupts() 466 dev_priv->guc.interrupts_enabled = true; in gen9_enable_guc_interrupts() 475 dev_priv->guc.interrupts_enabled = false; in gen9_disable_guc_interrupts() 1779 queue_work(dev_priv->guc.log.runtime.flush_wq, in gen9_guc_irq_handler() 1780 &dev_priv->guc.log.runtime.flush_work); in gen9_guc_irq_handler() 1782 dev_priv->guc.log.flush_interrupt_count++; in gen9_guc_irq_handler()
|
H A D | i915_drv.h | 2254 struct intel_guc guc; member 2774 static inline struct drm_i915_private *guc_to_i915(struct intel_guc *guc) in guc_to_i915() argument 2776 return container_of(guc, struct drm_i915_private, guc); in guc_to_i915()
|
H A D | intel_uncore.c | 1811 !dev_priv->guc.execbuf_client && in intel_has_reset_engine()
|