Home
last modified time | relevance | path

Searched refs:context (Results 1 – 25 of 2221) sorted by relevance

12345678910>>...89

/linux/tools/tracing/rtla/src/
H A Dosnoise.c93 if (!strcmp(context->orig_cpus, context->curr_cpus)) in osnoise_restore_cpus()
475 if (context->orig_stop_us == context->stop_us) in osnoise_restore_stop_us()
554 if (context->orig_stop_total_us == context->stop_total_us) in osnoise_restore_stop_total_us()
634 if (context->orig_print_stack == context->print_stack) in osnoise_restore_print_stack()
713 if (context->orig_tracing_thresh == context->tracing_thresh) in osnoise_restore_tracing_thresh()
823 if (context->orig_opt_irq_disable == context->opt_irq_disable) in osnoise_restore_irq_disable()
884 if (context->orig_opt_workload == context->opt_workload) in osnoise_restore_workload()
953 context = calloc(1, sizeof(*context)); in osnoise_context_alloc()
954 if (!context) in osnoise_context_alloc()
977 return context; in osnoise_context_alloc()
[all …]
H A Dosnoise.h59 int osnoise_get_context(struct osnoise_context *context);
60 void osnoise_put_context(struct osnoise_context *context);
62 int osnoise_set_cpus(struct osnoise_context *context, char *cpus);
63 void osnoise_restore_cpus(struct osnoise_context *context);
65 int osnoise_set_runtime_period(struct osnoise_context *context,
70 int osnoise_set_stop_us(struct osnoise_context *context,
72 void osnoise_restore_stop_us(struct osnoise_context *context);
74 int osnoise_set_stop_total_us(struct osnoise_context *context,
82 int osnoise_set_tracing_thresh(struct osnoise_context *context,
87 int osnoise_set_print_stack(struct osnoise_context *context,
[all …]
/linux/drivers/misc/vmw_vmci/
H A Dvmci_context.c107 context = kzalloc(sizeof(*context), GFP_KERNEL); in vmci_ctx_create()
108 if (!context) { in vmci_ctx_create()
173 return context; in vmci_ctx_create()
180 kfree(context); in vmci_ctx_create()
311 if (!context) { in vmci_ctx_enqueue_datagram()
409 context = c; in vmci_ctx_get()
416 return context; in vmci_ctx_get()
435 ctx_fire_notification(context->cid, context->priv_flags); in ctx_free_ctx()
606 if (!context) in vmci_ctx_add_notification()
672 if (!context) in vmci_ctx_remove_notification()
[all …]
H A Dvmci_route.c42 if (VMCI_INVALID_ID == dst->context) in vmci_route()
75 if (VMCI_INVALID_ID == src->context && in vmci_route()
77 src->context = vmci_get_context_id(); in vmci_route()
113 if (VMCI_INVALID_ID == src->context) in vmci_route()
130 if (VMCI_INVALID_ID == src->context) { in vmci_route()
139 src->context = VMCI_HOST_CONTEXT_ID; in vmci_route()
153 if (vmci_ctx_exists(dst->context)) { in vmci_route()
165 src->context = VMCI_HOST_CONTEXT_ID; in vmci_route()
167 src->context != dst->context) { in vmci_route()
209 if (VMCI_INVALID_ID == src->context) in vmci_route()
[all …]
/linux/security/selinux/ss/
H A Dmls.h27 int mls_compute_context_len(struct policydb *p, struct context *context);
28 void mls_sid_to_context(struct policydb *p, struct context *context,
35 struct context *context, struct sidtab *s, u32 def_sid);
40 int mls_range_set(struct context *context, struct mls_range *range);
43 struct context *oldc, struct context *newc);
57 int mls_export_netlbl_cat(struct policydb *p, struct context *context,
59 int mls_import_netlbl_cat(struct policydb *p, struct context *context,
63 struct context *context, in mls_export_netlbl_lvl() argument
69 struct context *context, in mls_import_netlbl_lvl() argument
75 struct context *context, in mls_export_netlbl_cat() argument
[all …]
H A Dmls.c32 int mls_compute_context_len(struct policydb *p, struct context *context) in mls_compute_context_len() argument
86 void mls_sid_to_context(struct policydb *p, struct context *context, in mls_sid_to_context() argument
231 struct context *context, struct sidtab *s, u32 def_sid) in mls_context_to_sid() argument
339 context->range.level[1].sens = context->range.level[0].sens; in mls_context_to_sid()
379 int mls_range_set(struct context *context, struct mls_range *range) in mls_range_set() argument
440 struct context *oldc, struct context *newc) in mls_convert_context()
553 void mls_export_netlbl_lvl(struct policydb *p, struct context *context, in mls_export_netlbl_lvl() argument
581 context->range.level[1].sens = context->range.level[0].sens; in mls_import_netlbl_lvl()
595 int mls_export_netlbl_cat(struct policydb *p, struct context *context, in mls_export_netlbl_cat() argument
624 int mls_import_netlbl_cat(struct policydb *p, struct context *context, in mls_import_netlbl_cat() argument
[all …]
/linux/drivers/gpu/drm/etnaviv/
H A Detnaviv_mmu.c29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap()
55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map()
79 if (!context || !sgt) in etnaviv_iommu_map()
140 struct etnaviv_iommu_context *context = mapping->context; in etnaviv_iommu_reap_mapping() local
298 mapping->context = etnaviv_iommu_context_get(context); in etnaviv_iommu_map_gem()
325 mapping->context = etnaviv_iommu_context_get(context); in etnaviv_iommu_map_gem()
360 etnaviv_cmdbuf_suballoc_unmap(context, &context->cmdbuf_mapping); in etnaviv_iommu_context_free()
362 context->global->ops->free(context); in etnaviv_iommu_context_free()
408 context->global->ops->restore(gpu, context); in etnaviv_iommu_restore()
482 return context->global->ops->dump_size(context); in etnaviv_iommu_dump_size()
[all …]
H A Detnaviv_iommu.c37 drm_mm_takedown(&context->mm); in etnaviv_iommuv1_free()
140 context = global->v1.shared_context; in etnaviv_iommuv1_context_alloc()
141 etnaviv_iommu_context_get(context); in etnaviv_iommuv1_context_alloc()
143 return context; in etnaviv_iommuv1_context_alloc()
160 context = &v1_context->base; in etnaviv_iommuv1_context_alloc()
161 context->global = global; in etnaviv_iommuv1_context_alloc()
162 kref_init(&context->refcount); in etnaviv_iommuv1_context_alloc()
163 mutex_init(&context->lock); in etnaviv_iommuv1_context_alloc()
164 INIT_LIST_HEAD(&context->mappings); in etnaviv_iommuv1_context_alloc()
166 context->global->v1.shared_context = context; in etnaviv_iommuv1_context_alloc()
[all …]
H A Detnaviv_iommu_v2.c52 drm_mm_takedown(&context->mm); in etnaviv_iommuv2_free()
166 struct etnaviv_iommu_context *context) in etnaviv_iommuv2_restore_nonsec() argument
181 (u32)context->global->bad_page_dma); in etnaviv_iommuv2_restore_nonsec()
190 struct etnaviv_iommu_context *context) in etnaviv_iommuv2_restore_sec() argument
273 struct etnaviv_iommu_context *context; in etnaviv_iommuv2_context_alloc() local
300 context = &v2_context->base; in etnaviv_iommuv2_context_alloc()
301 context->global = global; in etnaviv_iommuv2_context_alloc()
302 kref_init(&context->refcount); in etnaviv_iommuv2_context_alloc()
303 mutex_init(&context->lock); in etnaviv_iommuv2_context_alloc()
304 INIT_LIST_HEAD(&context->mappings); in etnaviv_iommuv2_context_alloc()
[all …]
/linux/fs/xfs/
H A Dxfs_attr_list.c90 context->put_listent(context, in xfs_attr_shortform_list()
99 if (context->seen_enough) in xfs_attr_shortform_list()
108 if (context->bufsize == 0) in xfs_attr_shortform_list()
186 context->put_listent(context, in xfs_attr_shortform_list()
419 if (context->resynch) { in xfs_attr3_leaf_list_int()
424 context->dupcnt = 0; in xfs_attr3_leaf_list_int()
427 context->dupcnt++; in xfs_attr3_leaf_list_int()
430 context->dupcnt = 0; in xfs_attr3_leaf_list_int()
442 context->resynch = 0; in xfs_attr3_leaf_list_int()
481 context->put_listent(context, entry->flags, in xfs_attr3_leaf_list_int()
[all …]
H A Dxfs_xattr.c210 if (context->count < 0 || context->seen_enough) in __xfs_xattr_put_listent()
213 if (!context->buffer) in __xfs_xattr_put_listent()
219 context->seen_enough = 1; in __xfs_xattr_put_listent()
222 offset = context->buffer + context->count; in __xfs_xattr_put_listent()
306 memset(&context, 0, sizeof(context)); in xfs_vn_listxattr()
307 context.dp = XFS_I(inode); in xfs_vn_listxattr()
308 context.resynch = 1; in xfs_vn_listxattr()
310 context.bufsize = size; in xfs_vn_listxattr()
311 context.firstu = context.bufsize; in xfs_vn_listxattr()
317 if (context.count < 0) in xfs_vn_listxattr()
[all …]
/linux/kernel/
H A Dauditsc.c1036 context->context = AUDIT_CTX_UNUSED; in audit_alloc_context()
1688 switch (context->context) { in audit_log_exit()
1694 context->arch, context->major); in audit_log_exit()
1760 context->fds[0], context->fds[1]); in audit_log_exit()
1810 if (context->context == AUDIT_CTX_SYSCALL) in audit_log_exit()
2022 if (context->context != AUDIT_CTX_UNUSED || context->name_count) { in __audit_syscall_entry()
2044 context->context = AUDIT_CTX_SYSCALL; in __audit_syscall_entry()
2064 if (!context || context->dummy || in __audit_syscall_exit()
2234 if (context->context == AUDIT_CTX_UNUSED) in __audit_getname()
2306 if (context->context == AUDIT_CTX_UNUSED) in __audit_inode()
[all …]
/linux/drivers/gpu/drm/tegra/
H A Duapi.c46 kfree(context); in tegra_drm_channel_context_close()
89 context = kzalloc(sizeof(*context), GFP_KERNEL); in tegra_drm_ioctl_channel_open()
90 if (!context) in tegra_drm_ioctl_channel_open()
135 err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX), in tegra_drm_ioctl_channel_open()
157 kfree(context); in tegra_drm_ioctl_channel_open()
170 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close()
171 if (!context) { in tegra_drm_ioctl_channel_close()
200 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map()
201 if (!context) { in tegra_drm_ioctl_channel_map()
281 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
[all …]
H A Dsubmit.c26 #define SUBMIT_ERR(context, fmt, ...) \ argument
150 xa_lock(&context->mappings); in tegra_drm_mapping_get()
156 xa_unlock(&context->mappings); in tegra_drm_mapping_get()
243 SUBMIT_ERR(context, in submit_write_reloc()
403 class = context->client->base.class; in submit_create_job()
516 struct tegra_drm_context *context; in tegra_drm_ioctl_channel_submit() local
524 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_submit()
525 if (!context) { in tegra_drm_ioctl_channel_submit()
528 current->comm, args->context); in tegra_drm_ioctl_channel_submit()
598 if (context->memory_context && context->client->ops->can_use_memory_ctx) { in tegra_drm_ioctl_channel_submit()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_resources.c47 memset(context, 0, sizeof(*context)); in mlx4_en_fill_qp_context()
49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context()
50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context()
59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context()
63 context->local_qpn = cpu_to_be32(qpn); in mlx4_en_fill_qp_context()
64 context->pri_path.ackto = 1 & 0x07; in mlx4_en_fill_qp_context()
68 context->pri_path.sched_queue |= user_prio << 3; in mlx4_en_fill_qp_context()
72 context->cqn_send = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
73 context->cqn_recv = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
76 context->pri_path.counter_index != in mlx4_en_fill_qp_context()
[all …]
/linux/drivers/platform/x86/intel/int1092/
H A Dintel_sar.c41 &context->config_data[context->reg_value]; in update_sar_data()
135 out = acpi_evaluate_dsm_typed(context->handle, &context->guid, rev, in sar_get_device_mode()
143 update_sar_data(context); in sar_get_device_mode()
189 context->reg_value = value; in intc_reg_store()
190 update_sar_data(context); in intc_reg_store()
223 out = acpi_evaluate_dsm_typed(context->handle, &context->guid, rev, in sar_get_data()
251 context = kzalloc(sizeof(*context), GFP_KERNEL); in sar_probe()
252 if (!context) in sar_probe()
266 sar_get_data(reg, context); in sar_probe()
291 kfree(context); in sar_probe()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn32/
H A Ddcn32_fpu.c183 static bool dcn32_apply_merge_split_flags_helper(struct dc *dc, struct dc_state *context, in dcn32_build_wm_range_table_fpu()
276 struct dc_state *context, in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
282 struct vba_vars_st *vba = &context->bw_ctx.dml.vba; in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
284 enum clock_change_support temp_clock_change_support = vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb]; in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
290 vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb] = temp_clock_change_support; in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
291 context->bw_ctx.dml.soc.dram_clock_change_latency_us = in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
293 dcn32_internal_validate_bw(dc, context, pipes, &pipe_cnt, &vlevel, false); in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
297 dcn32_subvp_in_use(dc, context)) in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
298 vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb] = temp_clock_change_support; in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
300 if (vlevel < context in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
273 dcn32_find_dummy_latency_index_for_fw_based_mclk_switch(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int pipe_cnt,int vlevel) dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() argument
331 dcn32_helper_populate_phantom_dlg_params(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int pipe_cnt) dcn32_helper_populate_phantom_dlg_params() argument
464 dcn32_set_phantom_stream_timing(struct dc * dc,struct dc_state * context,struct pipe_ctx * ref_pipe,struct dc_stream_state * phantom_stream,display_e2e_pipe_params_st * pipes,unsigned int pipe_cnt,unsigned int dc_pipe_idx) dcn32_set_phantom_stream_timing() argument
555 dcn32_get_num_free_pipes(struct dc * dc,struct dc_state * context) dcn32_get_num_free_pipes() argument
597 dcn32_assign_subvp_pipe(struct dc * dc,struct dc_state * context,unsigned int * index) dcn32_assign_subvp_pipe() argument
675 dcn32_enough_pipes_for_subvp(struct dc * dc,struct dc_state * context) dcn32_enough_pipes_for_subvp() argument
721 subvp_subvp_schedulable(struct dc * dc,struct dc_state * context) subvp_subvp_schedulable() argument
792 subvp_drr_schedulable(struct dc * dc,struct dc_state * context) subvp_drr_schedulable() argument
892 subvp_vblank_schedulable(struct dc * dc,struct dc_state * context) subvp_vblank_schedulable() argument
980 subvp_subvp_admissable(struct dc * dc,struct dc_state * context) subvp_subvp_admissable() argument
1032 subvp_validate_static_schedulability(struct dc * dc,struct dc_state * context,int vlevel) subvp_validate_static_schedulability() argument
1087 assign_subvp_index(struct dc * dc,struct dc_state * context) assign_subvp_index() argument
1163 init_pipe_slice_table_from_context(struct pipe_slice_table * table,struct dc_state * context) init_pipe_slice_table_from_context() argument
1192 update_pipe_slice_table_with_split_flags(struct pipe_slice_table * table,struct dc * dc,struct dc_state * context,struct vba_vars_st * vba,int split[MAX_PIPES],bool merge[MAX_PIPES]) update_pipe_slice_table_with_split_flags() argument
1258 update_pipes_with_slice_table(struct dc * dc,struct dc_state * context,struct pipe_slice_table * table) update_pipes_with_slice_table() argument
1276 update_pipes_with_split_flags(struct dc * dc,struct dc_state * context,struct vba_vars_st * vba,int split[MAX_PIPES],bool merge[MAX_PIPES]) update_pipes_with_split_flags() argument
1292 should_apply_odm_power_optimization(struct dc * dc,struct dc_state * context,struct vba_vars_st * v,int * split,bool * merge) should_apply_odm_power_optimization() argument
1386 try_odm_power_optimization_and_revalidate(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int * split,bool * merge,unsigned int * vlevel,int pipe_cnt) try_odm_power_optimization_and_revalidate() argument
1416 is_test_pattern_enabled(struct dc_state * context) is_test_pattern_enabled() argument
1429 dcn32_full_validate_bw_helper(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int * vlevel,int * split,bool * merge,int * pipe_cnt) dcn32_full_validate_bw_helper() argument
1595 is_dtbclk_required(struct dc * dc,struct dc_state * context) is_dtbclk_required() argument
1640 dcn32_calculate_dlg_params(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int pipe_cnt,int vlevel) dcn32_calculate_dlg_params() argument
1806 dcn32_find_split_pipe(struct dc * dc,struct dc_state * context,int old_index) dcn32_find_split_pipe() argument
1933 dcn32_internal_validate_bw(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int * pipe_cnt_out,int * vlevel_out,bool fast_validate) dcn32_internal_validate_bw() argument
2267 dcn32_calculate_wm_and_dlg_fpu(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int pipe_cnt,int vlevel) dcn32_calculate_wm_and_dlg_fpu() argument
3377 dcn32_allow_subvp_high_refresh_rate(struct dc * dc,struct dc_state * context,struct pipe_ctx * pipe) dcn32_allow_subvp_high_refresh_rate() argument
3433 dcn32_determine_max_vratio_prefetch(struct dc * dc,struct dc_state * context) dcn32_determine_max_vratio_prefetch() argument
3467 dcn32_assign_fpo_vactive_candidate(struct dc * dc,const struct dc_state * context,struct dc_stream_state ** fpo_candidate_stream) dcn32_assign_fpo_vactive_candidate() argument
3503 dcn32_find_vactive_pipe(struct dc * dc,const struct dc_state * context,uint32_t vactive_margin_req_us) dcn32_find_vactive_pipe() argument
3534 dcn32_override_min_req_memclk(struct dc * dc,struct dc_state * context) dcn32_override_min_req_memclk() argument
[all...]
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_nvmetcp_fw_funcs.c145 struct e5_nvmetcp_task_context *context = task_params->context; in init_default_nvmetcp_task() local
149 memset(context, 0, sizeof(*context)); in init_default_nvmetcp_task()
150 init_nvmetcp_task_params(context, task_params, in init_default_nvmetcp_task()
210 SET_FIELD(context->ustorm_st_context.flags, in set_local_completion_context()
221 struct e5_nvmetcp_task_context *context = task_params->context; in init_rw_nvmetcp_task() local
247 &context->mstorm_st_context.data_desc, in init_rw_nvmetcp_task()
260 &context->ustorm_ag_context, in init_rw_nvmetcp_task()
279 context->ustorm_ag_context.exp_cont_len = 0; in init_rw_nvmetcp_task()
327 struct e5_nvmetcp_task_context *context = task_params->context; in init_common_login_request_task() local
334 &context->ustorm_ag_context, in init_common_login_request_task()
[all …]
/linux/arch/s390/include/asm/
H A Dmmu_context.h24 spin_lock_init(&mm->context.lock); in init_new_context()
25 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context()
29 mm->context.gmap_asce = 0; in init_new_context()
30 mm->context.flush_mm = 0; in init_new_context()
35 mm->context.has_pgste = 0; in init_new_context()
36 mm->context.uses_skeys = 0; in init_new_context()
37 mm->context.uses_cmm = 0; in init_new_context()
38 mm->context.allow_cow_sharing = 1; in init_new_context()
39 mm->context.allow_gmap_hpage_1m = 0; in init_new_context()
41 switch (mm->context.asce_limit) { in init_new_context()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn30/
H A Ddcn30_fpu.c365 void dcn30_fpu_update_soc_for_wm_a(struct dc *dc, struct dc_state *context) in dcn30_fpu_update_soc_for_wm_a() argument
371 if (!context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching || in dcn30_fpu_update_soc_for_wm_a()
372 context->bw_ctx.dml.soc.dram_clock_change_latency_us == 0) in dcn30_fpu_update_soc_for_wm_a()
373 context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.pstate_latency_us; in dcn30_fpu_update_soc_for_wm_a()
374 context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.sr_enter_plus_exit_time_us; in dcn30_fpu_update_soc_for_wm_a()
375 context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_input.sr_exit_time_us; in dcn30_fpu_update_soc_for_wm_a()
380 struct dc *dc, struct dc_state *context, in dcn30_fpu_calculate_wm_and_dlg() argument
385 int maxMpcComb = context->bw_ctx.dml.vba.maxMpcComb; in dcn30_fpu_calculate_wm_and_dlg()
387 double dcfclk = context->bw_ctx.dml.vba.DCFCLKState[vlevel][maxMpcComb]; in dcn30_fpu_calculate_wm_and_dlg()
388 bool pstate_en = context in dcn30_fpu_calculate_wm_and_dlg()
692 dcn30_find_dummy_latency_index_for_fw_based_mclk_switch(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes,int pipe_cnt,int vlevel) dcn30_find_dummy_latency_index_for_fw_based_mclk_switch() argument
[all...]
/linux/arch/sparc/mm/
H A Dtsb.c545 spin_lock_init(&mm->context.lock); in init_new_context()
547 mm->context.sparc64_ctx_val = 0UL; in init_new_context()
549 mm->context.tag_store = NULL; in init_new_context()
559 mm->context.hugetlb_pte_count = 0; in init_new_context()
560 mm->context.thp_pte_count = 0; in init_new_context()
570 mm->context.tsb_block[i].tsb = NULL; in init_new_context()
611 if (CTX_VALID(mm->context)) { in destroy_context()
619 if (mm->context.tag_store) { in destroy_context()
624 tag_desc = mm->context.tag_store; in destroy_context()
632 kfree(mm->context.tag_store); in destroy_context()
[all …]
/linux/arch/powerpc/mm/book3s64/
H A Dmmu_context.c101 if (!mm->context.hash_context) in hash__init_new_context()
118 if (mm->context.id == 0) { in hash__init_new_context()
142 kfree(mm->context.hash_context); in hash__init_new_context()
189 mm->context.hash_context = NULL; in radix__init_new_context()
207 mm->context.id = index; in init_new_context()
209 mm->context.pte_frag = NULL; in init_new_context()
210 mm->context.pmd_frag = NULL; in init_new_context()
265 frag = mm->context.pte_frag; in destroy_pagetable_cache()
269 frag = mm->context.pmd_frag; in destroy_pagetable_cache()
297 destroy_contexts(&mm->context); in destroy_context()
[all …]
/linux/drivers/usb/image/
H A Dmicrotek.c384 context in mts_int_submit_urb()
402 context->final_callback(context->srb); in mts_transfer_cleanup()
410 context->srb->result |= (unsigned)(*context->scsi_status)<<1; in mts_transfer_done()
436 scsi_set_resid(context->srb, context->data_length - in mts_data_done()
477 context->data, in mts_command_done()
500 context->curr_sg = sg_next(context->curr_sg); in mts_do_sg()
502 context->data_pipe, in mts_do_sg()
529 desc->context.srb = srb; in mts_build_transfer_context()
537 desc->context.data = sg_virt(desc->context.curr_sg); in mts_build_transfer_context()
538 desc->context.data_length = desc->context.curr_sg->length; in mts_build_transfer_context()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dcn32/
H A Ddcn32_resource_helpers.c85 * @context: new dc state in dcn32_helper_calculate_mall_bytes_for_cursor()
91 struct dc_state *context) in dcn32_helper_calculate_mall_bytes_for_cursor()
93 if (context->bw_ctx.bw.dcn.mall_subvp_size_bytes > 0) { in dcn32_helper_calculate_mall_bytes_for_cursor()
97 return dc->res_pool->funcs->calculate_mall_ways_from_bytes(dc, context->bw_ctx.bw.dcn.mall_subvp_size_bytes);
107 struct dc_state *context)
113 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; in dcn32_helper_calculate_num_ways_for_subvp()
131 dcn20_release_dsc(&context->res_ctx, dc->res_pool, &pipe->stream_res.dsc); in dcn32_merge_pipes_for_subvp()
153 struct dc_state *context) in dcn32_merge_pipes_for_subvp()
158 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; in dcn32_merge_pipes_for_subvp()
170 struct dc_state *context) in dcn32_all_pipes_have_stream_and_plane() argument
110 dcn32_helper_calculate_num_ways_for_subvp(struct dc * dc,struct dc_state * context) dcn32_helper_calculate_num_ways_for_subvp() argument
124 dcn32_merge_pipes_for_subvp(struct dc * dc,struct dc_state * context) dcn32_merge_pipes_for_subvp() argument
187 dcn32_subvp_in_use(struct dc * dc,struct dc_state * context) dcn32_subvp_in_use() argument
200 dcn32_mpo_in_use(struct dc_state * context) dcn32_mpo_in_use() argument
212 dcn32_any_surfaces_rotated(struct dc * dc,struct dc_state * context) dcn32_any_surfaces_rotated() argument
259 override_det_for_subvp(struct dc * dc,struct dc_state * context,uint8_t pipe_segments[]) override_det_for_subvp() argument
328 dcn32_determine_det_override(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes) dcn32_determine_det_override() argument
397 dcn32_set_det_allocations(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes) dcn32_set_det_allocations() argument
527 dcn32_can_support_mclk_switch_using_fw_based_vblank_stretch(struct dc * dc,struct dc_state * context) dcn32_can_support_mclk_switch_using_fw_based_vblank_stretch() argument
654 dcn32_subvp_drr_admissable(struct dc * dc,struct dc_state * context) dcn32_subvp_drr_admissable() argument
714 dcn32_subvp_vblank_admissable(struct dc * dc,struct dc_state * context,int vlevel) dcn32_subvp_vblank_admissable() argument
760 dcn32_update_dml_pipes_odm_policy_based_on_context(struct dc * dc,struct dc_state * context,display_e2e_pipe_params_st * pipes) dcn32_update_dml_pipes_odm_policy_based_on_context() argument
786 dcn32_override_min_req_dcfclk(struct dc * dc,struct dc_state * context) dcn32_override_min_req_dcfclk() argument
[all...]
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_cmd.c99 token, context->token); in hns_roce_cmd_event()
104 context->out_param = out_param; in hns_roce_cmd_event()
105 complete(&context->done); in hns_roce_cmd_event()
120 context = &cmd->context[cmd->free_head]; in __hns_roce_cmd_mbox_wait()
122 } while (context->busy); in __hns_roce_cmd_mbox_wait()
124 context->busy = 1; in __hns_roce_cmd_mbox_wait()
148 ret = context->result; in __hns_roce_cmd_mbox_wait()
154 context->busy = 0; in __hns_roce_cmd_mbox_wait()
221 hr_cmd->context = in hns_roce_cmd_use_events()
223 if (!hr_cmd->context) { in hns_roce_cmd_use_events()
[all …]

12345678910>>...89