Home
last modified time | relevance | path

Searched refs:context (Results 1 – 25 of 2364) sorted by relevance

12345678910>>...95

/linux/tools/tracing/rtla/src/
H A Dosnoise.c93 if (!strcmp(context->orig_cpus, context->curr_cpus)) in osnoise_restore_cpus()
475 if (context->orig_stop_us == context->stop_us) in osnoise_restore_stop_us()
554 if (context->orig_stop_total_us == context->stop_total_us) in osnoise_restore_stop_total_us()
634 if (context->orig_print_stack == context->print_stack) in osnoise_restore_print_stack()
713 if (context->orig_tracing_thresh == context->tracing_thresh) in osnoise_restore_tracing_thresh()
823 if (context->orig_opt_irq_disable == context->opt_irq_disable) in osnoise_restore_irq_disable()
884 if (context->orig_opt_workload == context->opt_workload) in osnoise_restore_workload()
953 context = calloc(1, sizeof(*context)); in osnoise_context_alloc()
954 if (!context) in osnoise_context_alloc()
977 return context; in osnoise_context_alloc()
[all …]
H A Dosnoise.h59 int osnoise_get_context(struct osnoise_context *context);
60 void osnoise_put_context(struct osnoise_context *context);
62 int osnoise_set_cpus(struct osnoise_context *context, char *cpus);
63 void osnoise_restore_cpus(struct osnoise_context *context);
65 int osnoise_set_runtime_period(struct osnoise_context *context,
70 int osnoise_set_stop_us(struct osnoise_context *context,
72 void osnoise_restore_stop_us(struct osnoise_context *context);
74 int osnoise_set_stop_total_us(struct osnoise_context *context,
82 int osnoise_set_tracing_thresh(struct osnoise_context *context,
87 int osnoise_set_print_stack(struct osnoise_context *context,
[all …]
/linux/drivers/accel/qaic/
H A Dsahara.c121 if (image_id >= context->table_size || !context->image_table[image_id]) { in sahara_find_image()
159 ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0], in sahara_send_reset()
200 context->tx[0]->hello_resp.mode = context->rx->hello_resp.mode; in sahara_hello()
202 ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0], in sahara_hello()
270 memcpy(context->tx[i], &context->firmware->data[data_offset], pkt_data_len); in sahara_read_data()
317 ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0], in sahara_end_of_image()
347 ret = mhi_queue_buf(context->mhi_dev, DMA_FROM_DEVICE, context->rx, in sahara_processing()
359 context = devm_kzalloc(&mhi_dev->dev, sizeof(*context), GFP_KERNEL); in sahara_mhi_probe()
360 if (!context) in sahara_mhi_probe()
364 if (!context->rx) in sahara_mhi_probe()
[all …]
/linux/drivers/misc/vmw_vmci/
H A Dvmci_context.c107 context = kzalloc(sizeof(*context), GFP_KERNEL); in vmci_ctx_create()
108 if (!context) { in vmci_ctx_create()
173 return context; in vmci_ctx_create()
180 kfree(context); in vmci_ctx_create()
311 if (!context) { in vmci_ctx_enqueue_datagram()
409 context = c; in vmci_ctx_get()
416 return context; in vmci_ctx_get()
435 ctx_fire_notification(context->cid, context->priv_flags); in ctx_free_ctx()
606 if (!context) in vmci_ctx_add_notification()
672 if (!context) in vmci_ctx_remove_notification()
[all …]
H A Dvmci_route.c42 if (VMCI_INVALID_ID == dst->context) in vmci_route()
75 if (VMCI_INVALID_ID == src->context && in vmci_route()
77 src->context = vmci_get_context_id(); in vmci_route()
113 if (VMCI_INVALID_ID == src->context) in vmci_route()
130 if (VMCI_INVALID_ID == src->context) { in vmci_route()
139 src->context = VMCI_HOST_CONTEXT_ID; in vmci_route()
153 if (vmci_ctx_exists(dst->context)) { in vmci_route()
165 src->context = VMCI_HOST_CONTEXT_ID; in vmci_route()
167 src->context != dst->context) { in vmci_route()
209 if (VMCI_INVALID_ID == src->context) in vmci_route()
[all …]
/linux/security/selinux/ss/
H A Dmls.h27 int mls_compute_context_len(struct policydb *p, struct context *context);
28 void mls_sid_to_context(struct policydb *p, struct context *context,
35 struct context *context, struct sidtab *s, u32 def_sid);
40 int mls_range_set(struct context *context, struct mls_range *range);
43 struct context *oldc, struct context *newc);
57 int mls_export_netlbl_cat(struct policydb *p, struct context *context,
59 int mls_import_netlbl_cat(struct policydb *p, struct context *context,
63 struct context *context, in mls_export_netlbl_lvl() argument
69 struct context *context, in mls_import_netlbl_lvl() argument
75 struct context *context, in mls_export_netlbl_cat() argument
[all …]
H A Dmls.c32 int mls_compute_context_len(struct policydb *p, struct context *context) in mls_compute_context_len() argument
86 void mls_sid_to_context(struct policydb *p, struct context *context, in mls_sid_to_context() argument
231 struct context *context, struct sidtab *s, u32 def_sid) in mls_context_to_sid() argument
339 context->range.level[1].sens = context->range.level[0].sens; in mls_context_to_sid()
379 int mls_range_set(struct context *context, struct mls_range *range) in mls_range_set() argument
440 struct context *oldc, struct context *newc) in mls_convert_context()
553 void mls_export_netlbl_lvl(struct policydb *p, struct context *context, in mls_export_netlbl_lvl() argument
581 context->range.level[1].sens = context->range.level[0].sens; in mls_import_netlbl_lvl()
595 int mls_export_netlbl_cat(struct policydb *p, struct context *context, in mls_export_netlbl_cat() argument
624 int mls_import_netlbl_cat(struct policydb *p, struct context *context, in mls_import_netlbl_cat() argument
[all …]
/linux/drivers/gpu/drm/etnaviv/
H A Detnaviv_mmu.c29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap()
55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map()
79 if (!context || !sgt) in etnaviv_iommu_map()
140 struct etnaviv_iommu_context *context = mapping->context; in etnaviv_iommu_reap_mapping() local
298 mapping->context = etnaviv_iommu_context_get(context); in etnaviv_iommu_map_gem()
325 mapping->context = etnaviv_iommu_context_get(context); in etnaviv_iommu_map_gem()
360 etnaviv_cmdbuf_suballoc_unmap(context, &context->cmdbuf_mapping); in etnaviv_iommu_context_free()
362 context->global->ops->free(context); in etnaviv_iommu_context_free()
408 context->global->ops->restore(gpu, context); in etnaviv_iommu_restore()
482 return context->global->ops->dump_size(context); in etnaviv_iommu_dump_size()
[all …]
H A Detnaviv_iommu.c37 drm_mm_takedown(&context->mm); in etnaviv_iommuv1_free()
140 context = global->v1.shared_context; in etnaviv_iommuv1_context_alloc()
141 etnaviv_iommu_context_get(context); in etnaviv_iommuv1_context_alloc()
143 return context; in etnaviv_iommuv1_context_alloc()
160 context = &v1_context->base; in etnaviv_iommuv1_context_alloc()
161 context->global = global; in etnaviv_iommuv1_context_alloc()
162 kref_init(&context->refcount); in etnaviv_iommuv1_context_alloc()
163 mutex_init(&context->lock); in etnaviv_iommuv1_context_alloc()
164 INIT_LIST_HEAD(&context->mappings); in etnaviv_iommuv1_context_alloc()
166 context->global->v1.shared_context = context; in etnaviv_iommuv1_context_alloc()
[all …]
H A Detnaviv_iommu_v2.c52 drm_mm_takedown(&context->mm); in etnaviv_iommuv2_free()
166 struct etnaviv_iommu_context *context) in etnaviv_iommuv2_restore_nonsec() argument
181 (u32)context->global->bad_page_dma); in etnaviv_iommuv2_restore_nonsec()
190 struct etnaviv_iommu_context *context) in etnaviv_iommuv2_restore_sec() argument
273 struct etnaviv_iommu_context *context; in etnaviv_iommuv2_context_alloc() local
300 context = &v2_context->base; in etnaviv_iommuv2_context_alloc()
301 context->global = global; in etnaviv_iommuv2_context_alloc()
302 kref_init(&context->refcount); in etnaviv_iommuv2_context_alloc()
303 mutex_init(&context->lock); in etnaviv_iommuv2_context_alloc()
304 INIT_LIST_HEAD(&context->mappings); in etnaviv_iommuv2_context_alloc()
[all …]
/linux/kernel/
H A Dauditsc.c1036 context->context = AUDIT_CTX_UNUSED; in audit_alloc_context()
1688 switch (context->context) { in audit_log_exit()
1694 context->arch, context->major); in audit_log_exit()
1760 context->fds[0], context->fds[1]); in audit_log_exit()
1810 if (context->context == AUDIT_CTX_SYSCALL) in audit_log_exit()
2022 if (context->context != AUDIT_CTX_UNUSED || context->name_count) { in __audit_syscall_entry()
2044 context->context = AUDIT_CTX_SYSCALL; in __audit_syscall_entry()
2064 if (!context || context->dummy || in __audit_syscall_exit()
2234 if (context->context == AUDIT_CTX_UNUSED) in __audit_getname()
2306 if (context->context == AUDIT_CTX_UNUSED) in __audit_inode()
[all …]
/linux/fs/xfs/
H A Dxfs_attr_list.c91 context->put_listent(context, in xfs_attr_shortform_list()
101 if (context->seen_enough) in xfs_attr_shortform_list()
110 if (context->bufsize == 0) in xfs_attr_shortform_list()
193 context->put_listent(context, in xfs_attr_shortform_list()
455 if (context->resynch) { in xfs_attr3_leaf_list_int()
460 context->dupcnt = 0; in xfs_attr3_leaf_list_int()
463 context->dupcnt++; in xfs_attr3_leaf_list_int()
466 context->dupcnt = 0; in xfs_attr3_leaf_list_int()
478 context->resynch = 0; in xfs_attr3_leaf_list_int()
521 context->put_listent(context, entry->flags, in xfs_attr3_leaf_list_int()
[all …]
H A Dxfs_xattr.c215 if (context->count < 0 || context->seen_enough) in __xfs_xattr_put_listent()
218 if (!context->buffer) in __xfs_xattr_put_listent()
224 context->seen_enough = 1; in __xfs_xattr_put_listent()
227 offset = context->buffer + context->count; in __xfs_xattr_put_listent()
316 memset(&context, 0, sizeof(context)); in xfs_vn_listxattr()
317 context.dp = XFS_I(inode); in xfs_vn_listxattr()
318 context.resynch = 1; in xfs_vn_listxattr()
320 context.bufsize = size; in xfs_vn_listxattr()
321 context.firstu = context.bufsize; in xfs_vn_listxattr()
327 if (context.count < 0) in xfs_vn_listxattr()
[all …]
/linux/drivers/gpu/drm/tegra/
H A Duapi.c46 kfree(context); in tegra_drm_channel_context_close()
89 context = kzalloc(sizeof(*context), GFP_KERNEL); in tegra_drm_ioctl_channel_open()
90 if (!context) in tegra_drm_ioctl_channel_open()
135 err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX), in tegra_drm_ioctl_channel_open()
157 kfree(context); in tegra_drm_ioctl_channel_open()
170 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close()
171 if (!context) { in tegra_drm_ioctl_channel_close()
200 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map()
201 if (!context) { in tegra_drm_ioctl_channel_map()
281 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
[all …]
H A Dsubmit.c26 #define SUBMIT_ERR(context, fmt, ...) \ argument
150 xa_lock(&context->mappings); in tegra_drm_mapping_get()
156 xa_unlock(&context->mappings); in tegra_drm_mapping_get()
243 SUBMIT_ERR(context, in submit_write_reloc()
403 class = context->client->base.class; in submit_create_job()
516 struct tegra_drm_context *context; in tegra_drm_ioctl_channel_submit() local
524 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_submit()
525 if (!context) { in tegra_drm_ioctl_channel_submit()
528 current->comm, args->context); in tegra_drm_ioctl_channel_submit()
598 if (context->memory_context && context->client->ops->can_use_memory_ctx) { in tegra_drm_ioctl_channel_submit()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_resources.c47 memset(context, 0, sizeof(*context)); in mlx4_en_fill_qp_context()
49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context()
50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context()
59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context()
63 context->local_qpn = cpu_to_be32(qpn); in mlx4_en_fill_qp_context()
64 context->pri_path.ackto = 1 & 0x07; in mlx4_en_fill_qp_context()
68 context->pri_path.sched_queue |= user_prio << 3; in mlx4_en_fill_qp_context()
72 context->cqn_send = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
73 context->cqn_recv = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
76 context->pri_path.counter_index != in mlx4_en_fill_qp_context()
[all …]
/linux/drivers/platform/x86/intel/int1092/
H A Dintel_sar.c41 &context->config_data[context->reg_value]; in update_sar_data()
135 out = acpi_evaluate_dsm_typed(context->handle, &context->guid, rev, in sar_get_device_mode()
143 update_sar_data(context); in sar_get_device_mode()
189 context->reg_value = value; in intc_reg_store()
190 update_sar_data(context); in intc_reg_store()
223 out = acpi_evaluate_dsm_typed(context->handle, &context->guid, rev, in sar_get_data()
251 context = kzalloc(sizeof(*context), GFP_KERNEL); in sar_probe()
252 if (!context) in sar_probe()
266 sar_get_data(reg, context); in sar_probe()
291 kfree(context); in sar_probe()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn32/
H A Ddcn32_fpu.c1482 …!dcn32_mpo_in_use(context) && !dcn32_any_surfaces_rotated(dc, context) && !is_test_pattern_enabled… in dcn32_full_validate_bw_helper()
1651 context->bw_ctx.bw.dcn.clk.dispclk_khz = context->bw_ctx.dml.vba.DISPCLK * 1000; in dcn32_calculate_dlg_params()
1652 context->bw_ctx.bw.dcn.clk.dcfclk_khz = context->bw_ctx.dml.vba.DCFCLK * 1000; in dcn32_calculate_dlg_params()
1653 context->bw_ctx.bw.dcn.clk.socclk_khz = context->bw_ctx.dml.vba.SOCCLK * 1000; in dcn32_calculate_dlg_params()
1656 context->bw_ctx.bw.dcn.clk.fclk_khz = context->bw_ctx.dml.vba.FabricClock * 1000; in dcn32_calculate_dlg_params()
1667 context->bw_ctx.bw.dcn.clk.dtbclk_en = is_dtbclk_required(dc, context); in dcn32_calculate_dlg_params()
1791 context->bw_ctx.dml.funcs.rq_dlg_get_dlg_reg_v2(&context->bw_ctx.dml, in dcn32_calculate_dlg_params()
2540 context->bw_ctx.bw.dcn.watermarks.a = context->bw_ctx.bw.dcn.watermarks.c; in dcn32_calculate_wm_and_dlg_fpu()
2572 context->bw_ctx.bw.dcn.watermarks.d = context->bw_ctx.bw.dcn.watermarks.a; in dcn32_calculate_wm_and_dlg_fpu()
2593 context->perf_params.stutter_period_us = context->bw_ctx.dml.vba.StutterPeriod; in dcn32_calculate_wm_and_dlg_fpu()
[all …]
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_nvmetcp_fw_funcs.c145 struct e5_nvmetcp_task_context *context = task_params->context; in init_default_nvmetcp_task() local
149 memset(context, 0, sizeof(*context)); in init_default_nvmetcp_task()
150 init_nvmetcp_task_params(context, task_params, in init_default_nvmetcp_task()
210 SET_FIELD(context->ustorm_st_context.flags, in set_local_completion_context()
221 struct e5_nvmetcp_task_context *context = task_params->context; in init_rw_nvmetcp_task() local
247 &context->mstorm_st_context.data_desc, in init_rw_nvmetcp_task()
260 &context->ustorm_ag_context, in init_rw_nvmetcp_task()
279 context->ustorm_ag_context.exp_cont_len = 0; in init_rw_nvmetcp_task()
327 struct e5_nvmetcp_task_context *context = task_params->context; in init_common_login_request_task() local
334 &context->ustorm_ag_context, in init_common_login_request_task()
[all …]
/linux/arch/s390/include/asm/
H A Dmmu_context.h24 spin_lock_init(&mm->context.lock); in init_new_context()
25 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context()
29 mm->context.gmap_asce = 0; in init_new_context()
30 mm->context.flush_mm = 0; in init_new_context()
35 mm->context.has_pgste = 0; in init_new_context()
36 mm->context.uses_skeys = 0; in init_new_context()
37 mm->context.uses_cmm = 0; in init_new_context()
38 mm->context.allow_cow_sharing = 1; in init_new_context()
39 mm->context.allow_gmap_hpage_1m = 0; in init_new_context()
41 switch (mm->context.asce_limit) { in init_new_context()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dcn32/
H A Ddcn32_resource_helpers.c91 struct dc_state *context) in dcn32_helper_calculate_num_ways_for_subvp() argument
107 struct dc_state *context) in dcn32_merge_pipes_for_subvp() argument
153 struct dc_state *context) in dcn32_all_pipes_have_stream_and_plane() argument
170 struct dc_state *context) in dcn32_subvp_in_use() argument
254 if (dc_state_get_stream_subvp_type(context, context->streams[i]) != SUBVP_PHANTOM) in override_det_for_subvp()
311 struct dc_state *context, in dcn32_determine_det_override() argument
323 if (dc_state_get_stream_subvp_type(context, context->streams[i]) != SUBVP_PHANTOM) in dcn32_determine_det_override()
339 if (context->res_ctx.pipe_ctx[j].stream == context->streams[i] && in dcn32_determine_det_override()
349 if (k != j && context->res_ctx.pipe_ctx[k].stream == context->streams[i] && in dcn32_determine_det_override()
358 if (k != j && context->res_ctx.pipe_ctx[k].stream == context->streams[i] && in dcn32_determine_det_override()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn30/
H A Ddcn30_fpu.c380 struct dc *dc, struct dc_state *context, in dcn30_fpu_calculate_wm_and_dlg() argument
395 if (context->streams[i]) in dcn30_fpu_calculate_wm_and_dlg()
406 context, pipes, pipe_cnt, vlevel); in dcn30_fpu_calculate_wm_and_dlg()
416 dcfclk = context->bw_ctx.dml.vba.DCFCLKState[vlevel][context->bw_ctx.dml.vba.maxMpcComb]; in dcn30_fpu_calculate_wm_and_dlg()
441context->bw_ctx.bw.dcn.watermarks.b.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cn… in dcn30_fpu_calculate_wm_and_dlg()
511context->bw_ctx.bw.dcn.watermarks.c.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cn… in dcn30_fpu_calculate_wm_and_dlg()
524 context->bw_ctx.bw.dcn.watermarks.a = context->bw_ctx.bw.dcn.watermarks.c; in dcn30_fpu_calculate_wm_and_dlg()
534context->bw_ctx.bw.dcn.watermarks.a.urgent_ns = get_wm_urgent(&context->bw_ctx.dml, pipes, pipe_cn… in dcn30_fpu_calculate_wm_and_dlg()
544 context->perf_params.stutter_period_us = context->bw_ctx.dml.vba.StutterPeriod; in dcn30_fpu_calculate_wm_and_dlg()
547 context->bw_ctx.bw.dcn.watermarks.d = context->bw_ctx.bw.dcn.watermarks.a; in dcn30_fpu_calculate_wm_and_dlg()
[all …]
/linux/arch/sparc/mm/
H A Dtsb.c545 spin_lock_init(&mm->context.lock); in init_new_context()
547 mm->context.sparc64_ctx_val = 0UL; in init_new_context()
549 mm->context.tag_store = NULL; in init_new_context()
559 mm->context.hugetlb_pte_count = 0; in init_new_context()
560 mm->context.thp_pte_count = 0; in init_new_context()
570 mm->context.tsb_block[i].tsb = NULL; in init_new_context()
611 if (CTX_VALID(mm->context)) { in destroy_context()
619 if (mm->context.tag_store) { in destroy_context()
624 tag_desc = mm->context.tag_store; in destroy_context()
632 kfree(mm->context.tag_store); in destroy_context()
[all …]
/linux/arch/powerpc/mm/book3s64/
H A Dmmu_context.c101 if (!mm->context.hash_context) in hash__init_new_context()
118 if (mm->context.id == 0) { in hash__init_new_context()
142 kfree(mm->context.hash_context); in hash__init_new_context()
189 mm->context.hash_context = NULL; in radix__init_new_context()
207 mm->context.id = index; in init_new_context()
209 mm->context.pte_frag = NULL; in init_new_context()
210 mm->context.pmd_frag = NULL; in init_new_context()
265 frag = mm->context.pte_frag; in destroy_pagetable_cache()
269 frag = mm->context.pmd_frag; in destroy_pagetable_cache()
297 destroy_contexts(&mm->context); in destroy_context()
[all …]
/linux/drivers/usb/image/
H A Dmicrotek.c384 context in mts_int_submit_urb()
402 context->final_callback(context->srb); in mts_transfer_cleanup()
410 context->srb->result |= (unsigned)(*context->scsi_status)<<1; in mts_transfer_done()
436 scsi_set_resid(context->srb, context->data_length - in mts_data_done()
477 context->data, in mts_command_done()
500 context->curr_sg = sg_next(context->curr_sg); in mts_do_sg()
502 context->data_pipe, in mts_do_sg()
529 desc->context.srb = srb; in mts_build_transfer_context()
537 desc->context.data = sg_virt(desc->context.curr_sg); in mts_build_transfer_context()
538 desc->context.data_length = desc->context.curr_sg->length; in mts_build_transfer_context()
[all …]

12345678910>>...95