/linux/tools/testing/selftests/bpf/progs/ |
H A D | local_kptr_stash.c | 110 inner_local_kptr = bpf_kptr_xchg(&res->stashed_in_local_kptr, inner_local_kptr); in create_and_stash() 118 res = bpf_kptr_xchg(&mapval->node, res); in create_and_stash() 146 res = bpf_kptr_xchg(&mapval->plain, res); in stash_plain() 179 res = bpf_kptr_xchg(&mapval->local_root, res); in stash_local_with_root() 200 res = bpf_kptr_xchg(&mapval->node, NULL); in unstash_rb_node() 202 inner_local_kptr = bpf_kptr_xchg(&res->stashed_in_local_kptr, inner_local_kptr); in unstash_rb_node() 227 res = bpf_kptr_xchg(&mapval->val, NULL); in stash_test_ref_kfunc() 276 p = bpf_kptr_xchg(&s->stashed, p); in stash_refcounted_node()
|
H A D | cpumask_failure.c | 88 cpumask = bpf_kptr_xchg(&v->cpumask, NULL); in BPF_PROG() 114 prev = bpf_kptr_xchg(&global_mask, local); in BPF_PROG() 148 prev = bpf_kptr_xchg(&global_mask, local); in BPF_PROG() 176 prev = bpf_kptr_xchg(&global_mask, curr); in BPF_PROG() 183 prev = bpf_kptr_xchg(&global_mask, curr); in BPF_PROG()
|
H A D | map_kptr_fail.c | 71 bpf_kptr_xchg((void *)v + id, NULL); in non_const_var_off_kptr_xchg() 215 bpf_kptr_xchg(&v->unref_ptr, NULL); in reject_kptr_xchg_on_unref() 269 bpf_kptr_xchg(&v->ref_ptr, p); in reject_untrusted_xchg() 289 bpf_kptr_xchg(&v->ref_memb_ptr, ref_ptr); in reject_bad_type_xchg() 308 bpf_kptr_xchg(&v->ref_memb_ptr, &ref_ptr->memb); in reject_member_of_ref_xchg() 362 bpf_kptr_xchg(&v->ref_ptr, p); in kptr_xchg_ref_state() 381 p = bpf_kptr_xchg(&v->ref_ptr, p); in kptr_xchg_possibly_null()
|
H A D | percpu_alloc_array.c | 40 p = bpf_kptr_xchg(&e->pc, p); in BPF_PROG() 121 p = bpf_kptr_xchg(&e->pc, NULL); in BPF_PROG() 151 p1 = bpf_kptr_xchg(&e->pc, p); in BPF_PROG() 170 p1 = bpf_kptr_xchg(&e->pc, NULL); in BPF_PROG()
|
H A D | percpu_alloc_fail.c | 54 p = bpf_kptr_xchg(&e->pc, p); in BPF_PROG() 79 p = bpf_kptr_xchg(&e->pc, p2); in BPF_PROG() 103 p1 = bpf_kptr_xchg(&e->pc, p); in BPF_PROG()
|
H A D | test_bpf_ma.c | 74 old = bpf_kptr_xchg(&value->data, new); in batch_alloc() 96 old = bpf_kptr_xchg(&value->data, NULL); in batch_free() 124 old = bpf_kptr_xchg(&value->data, new); in batch_percpu_alloc() 147 old = bpf_kptr_xchg(&value->data, NULL); in batch_percpu_free()
|
H A D | map_kptr.c | 159 p = bpf_kptr_xchg(&v->ref_ptr, NULL); in test_kptr_ref() 179 p = bpf_kptr_xchg(&v->ref_ptr, p); in test_kptr_ref() 323 p = bpf_kptr_xchg(&v->ref_ptr, p); in test_map_kptr_ref_pre() 331 p = bpf_kptr_xchg(&v->ref_ptr, NULL); in test_map_kptr_ref_pre() 343 p = bpf_kptr_xchg(&v->ref_ptr, p); in test_map_kptr_ref_pre() 368 p = bpf_kptr_xchg(&v->ref_ptr, NULL); in test_map_kptr_ref_post() 376 p = bpf_kptr_xchg(&v->ref_ptr, p); in test_map_kptr_ref_post()
|
H A D | task_kfunc_success.c | 166 kptr = bpf_kptr_xchg(&v->task, NULL); in BPF_PROG() 179 kptr = bpf_kptr_xchg(&local->task, kptr); in BPF_PROG() 187 kptr = bpf_kptr_xchg(&local->task, NULL); in BPF_PROG() 204 acquired = bpf_kptr_xchg(&local->task, acquired); in BPF_PROG()
|
H A D | preempted_bpf_ma_op.c | 38 old = bpf_kptr_xchg(&value->data, NULL); in del_array() 61 old = bpf_kptr_xchg(&value->data, new); in add_array()
|
H A D | refcounted_kptr.c | 123 n = bpf_kptr_xchg(&mapval->node, n); in __stash_map_insert_tree() 216 n = bpf_kptr_xchg(&mapval->node, n); in __read_from_unstash() 422 n = bpf_kptr_xchg(&mapval->node, n); in __stash_map_empty_xchg() 462 n = bpf_kptr_xchg(&mapval->node, NULL); in rbtree_wrong_owner_remove_fail_b() 486 m = bpf_kptr_xchg(&mapval->node, NULL); in rbtree_wrong_owner_remove_fail_a2()
|
H A D | cpumask_success.c | 446 cpumask = bpf_kptr_xchg(&v->cpumask, NULL); in BPF_PROG() 467 prev = bpf_kptr_xchg(&global_mask, local); in BPF_PROG() 504 prev = bpf_kptr_xchg(&global_mask_array_one[0], local); in BPF_PROG() 540 local = bpf_kptr_xchg(mask0, local); in _global_mask_array_rcu() 558 local = bpf_kptr_xchg(mask1, local); in _global_mask_array_rcu()
|
H A D | cgrp_kfunc_failure.c | 147 kptr = bpf_kptr_xchg(&v->cgrp, NULL); in BPF_PROG() 231 old = bpf_kptr_xchg(&v->cgrp, acquired); in BPF_PROG()
|
H A D | kptr_xchg_inline.c | 34 __imm(bpf_kptr_xchg), in kptr_xchg_inline()
|
H A D | jit_probe_mem.c | 21 p = bpf_kptr_xchg(&v, p); in test_jit_probe_mem()
|
H A D | task_kfunc_failure.c | 142 kptr = bpf_kptr_xchg(&v->task, NULL); in BPF_PROG() 218 old = bpf_kptr_xchg(&v->task, acquired); in BPF_PROG()
|
H A D | local_kptr_stash_fail.c | 59 res = bpf_kptr_xchg(&mapval->node, res); in stash_rb_nodes()
|
H A D | crypto_common.h | 57 old = bpf_kptr_xchg(&v->ctx, ctx); in crypto_ctx_insert()
|
H A D | task_kfunc_common.h | 67 old = bpf_kptr_xchg(&v->task, acquired); in tasks_kfunc_map_insert()
|
H A D | cgrp_kfunc_common.h | 70 old = bpf_kptr_xchg(&v->cgrp, acquired); in cgrps_kfunc_map_insert()
|
H A D | cb_refs.c | 61 p = bpf_kptr_xchg(&v->ptr, p); in leak_prog()
|
H A D | percpu_alloc_cgrp_local_storage.c | 39 p = bpf_kptr_xchg(&e->pc, p); in BPF_PROG()
|
H A D | cgrp_kfunc_success.c | 96 kptr = bpf_kptr_xchg(&v->cgrp, NULL); in BPF_PROG()
|
H A D | kfunc_call_test.c | 208 ctx = bpf_kptr_xchg(&ctx_val->ctx, ctx); in kfunc_call_ctx()
|
H A D | cpumask_common.h | 114 old = bpf_kptr_xchg(&v->cpumask, mask); in cpumask_map_insert()
|
/linux/tools/sched_ext/ |
H A D | scx_flatcg.bpf.c | 297 cgv_node = bpf_kptr_xchg(&stash->node, NULL); in cgrp_enqueued() 714 cgv_node = bpf_kptr_xchg(&stash->node, cgv_node); in try_pick_next_cgroup() 884 cgv_node = bpf_kptr_xchg(&stash->node, cgv_node); in BPF_STRUCT_OPS_SLEEPABLE()
|