/freebsd/contrib/jemalloc/include/jemalloc/internal/ |
H A D | arena_stats.h | 165 return atomic_load_u64(p, ATOMIC_RELAXED); in arena_stats_read_u64() 176 atomic_fetch_add_u64(p, x, ATOMIC_RELAXED); in arena_stats_add_u64() 205 atomic_store_u64(dst, src + cur_dst, ATOMIC_RELAXED); in arena_stats_accum_u64() 215 return atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_read_zu() 218 return atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_read_zu() 226 atomic_fetch_add_zu(p, x, ATOMIC_RELAXED); in arena_stats_add_zu() 229 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_add_zu() 230 atomic_store_zu(p, cur + x, ATOMIC_RELAXED); in arena_stats_add_zu() 242 size_t cur = atomic_load_zu(p, ATOMIC_RELAXED); in arena_stats_sub_zu() 243 atomic_store_zu(p, cur - x, ATOMIC_RELAXED); in arena_stats_sub_zu() [all …]
|
H A D | prng.h | 73 state0 = atomic_load_u32(state, ATOMIC_RELAXED); in prng_lg_range_u32() 79 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prng_lg_range_u32() 82 atomic_store_u32(state, state1, ATOMIC_RELAXED); in prng_lg_range_u32() 110 state0 = atomic_load_zu(state, ATOMIC_RELAXED); in prng_lg_range_zu() 116 state1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prng_lg_range_zu() 119 atomic_store_zu(state, state1, ATOMIC_RELAXED); in prng_lg_range_zu()
|
H A D | prof_inlines_a.h | 21 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); in prof_accum_add() 30 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prof_accum_add() 58 a0 = atomic_load_u64(&prof_accum->accumbytes, ATOMIC_RELAXED); in prof_accum_cancel() 63 a1, ATOMIC_RELAXED, ATOMIC_RELAXED)); in prof_accum_cancel()
|
H A D | seq.h | 26 size_t old_seq = atomic_load_zu(&dst->seq, ATOMIC_RELAXED); \ 27 atomic_store_zu(&dst->seq, old_seq + 1, ATOMIC_RELAXED); \ 30 atomic_store_zu(&dst->data[i], buf[i], ATOMIC_RELAXED); \ 44 buf[i] = atomic_load_zu(&src->data[i], ATOMIC_RELAXED); \ 47 size_t seq2 = atomic_load_zu(&src->seq, ATOMIC_RELAXED); \
|
H A D | mutex.h | 154 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); in malloc_mutex_lock_final() 180 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); in malloc_mutex_trylock() 205 ATOMIC_RELAXED); in malloc_mutex_prof_merge() 207 &data->n_waiting_thds, ATOMIC_RELAXED); in malloc_mutex_prof_merge() 209 ATOMIC_RELAXED); in malloc_mutex_prof_merge() 220 atomic_store_b(&mutex->locked, true, ATOMIC_RELAXED); in malloc_mutex_lock() 229 atomic_store_b(&mutex->locked, false, ATOMIC_RELAXED); in malloc_mutex_unlock() 261 atomic_store_u32(&data->n_waiting_thds, 0, ATOMIC_RELAXED); in malloc_mutex_prof_read() 281 atomic_store_u32(&data->n_waiting_thds, 0, ATOMIC_RELAXED); in malloc_mutex_prof_accum()
|
H A D | arena_inlines_a.h | 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_add() 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_sub() 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED); in arena_internal_get()
|
H A D | background_thread_inlines.h | 6 return atomic_load_b(&background_thread_enabled_state, ATOMIC_RELAXED); in background_thread_enabled() 12 atomic_store_b(&background_thread_enabled_state, state, ATOMIC_RELAXED); in background_thread_enabled_set()
|
H A D | atomic.h | 48 #define ATOMIC_RELAXED atomic_memory_order_relaxed macro
|
H A D | rtree.h | 176 ? ATOMIC_RELAXED : ATOMIC_ACQUIRE); in rtree_leaf_elm_bits_read() 218 ? ATOMIC_RELAXED : ATOMIC_ACQUIRE); in rtree_leaf_elm_extent_read() 230 return (szind_t)atomic_load_u(&elm->le_szind, dependent ? ATOMIC_RELAXED in rtree_leaf_elm_szind_read() 242 return atomic_load_b(&elm->le_slab, dependent ? ATOMIC_RELAXED : in rtree_leaf_elm_slab_read()
|
H A D | log.h | 68 ATOMIC_RELAXED); \
|
/freebsd/contrib/jemalloc/src/ |
H A D | hook.c | 39 atomic_load_u(&nhooks, ATOMIC_RELAXED) + 1, in hook_install_locked() 40 ATOMIC_RELAXED); in hook_install_locked() 68 atomic_store_u(&nhooks, atomic_load_u(&nhooks, ATOMIC_RELAXED) - 1, in hook_remove_locked() 69 ATOMIC_RELAXED); in hook_remove_locked() 141 if (likely(atomic_load_u(&nhooks, ATOMIC_RELAXED) == 0)) { \
|
H A D | extent_dss.c | 77 true, ATOMIC_ACQ_REL, ATOMIC_RELAXED)) { in extent_dss_extending_start() 86 assert(atomic_load_b(&dss_extending, ATOMIC_RELAXED)); in extent_dss_extending_finish() 266 atomic_store_b(&dss_extending, false, ATOMIC_RELAXED); in extent_dss_boot() 267 atomic_store_b(&dss_exhausted, dss_base == (void *)-1, ATOMIC_RELAXED); in extent_dss_boot() 268 atomic_store_p(&dss_max, dss_base, ATOMIC_RELAXED); in extent_dss_boot()
|
H A D | tsd.c | 116 assert(tsd_atomic_load(&remote_tsd->state, ATOMIC_RELAXED) in tsd_force_recompute() 119 ATOMIC_RELAXED); in tsd_force_recompute() 126 atomic_fetch_add_u32(&tsd_global_slow_count, 1, ATOMIC_RELAXED); in tsd_global_slow_inc() 139 atomic_fetch_sub_u32(&tsd_global_slow_count, 1, ATOMIC_RELAXED); in tsd_global_slow_dec() 152 return atomic_load_u32(&tsd_global_slow_count, ATOMIC_RELAXED) > 0; in tsd_global_slow() 184 uint8_t old_state = tsd_atomic_load(&tsd->state, ATOMIC_RELAXED); in tsd_state_set() 191 tsd_atomic_store(&tsd->state, new_state, ATOMIC_RELAXED); in tsd_state_set() 205 ATOMIC_RELAXED); in tsd_state_set()
|
H A D | arena.c | 106 ATOMIC_RELAXED); in arena_stats_merge() 132 (((atomic_load_zu(&arena->nactive, ATOMIC_RELAXED) + in arena_stats_merge() 136 &arena->stats.abandoned_vm, ATOMIC_RELAXED)); in arena_stats_merge() 188 ATOMIC_RELAXED); in arena_stats_merge() 190 ATOMIC_RELAXED); in arena_stats_merge() 192 ATOMIC_RELAXED); in arena_stats_merge() 520 return atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED); in arena_decay_ms_read() 1147 atomic_store_zu(&arena->nactive, 0, ATOMIC_RELAXED); in arena_reset() 1998 (size_t)(uintptr_t)arena, ATOMIC_RELAXED); in arena_new() 2004 ATOMIC_RELAXED); in arena_new() [all …]
|
H A D | rtree.c | 70 ATOMIC_RELAXED); in rtree_delete_subtree() 80 ATOMIC_RELAXED); in rtree_delete_subtree() 109 rtree_node_elm_t *node = atomic_load_p(elmp, ATOMIC_RELAXED); in rtree_node_init() 135 rtree_leaf_elm_t *leaf = atomic_load_p(elmp, ATOMIC_RELAXED); in rtree_leaf_init() 170 ATOMIC_RELAXED); in rtree_child_node_tryread() 199 ATOMIC_RELAXED); in rtree_child_leaf_tryread()
|
H A D | mutex.c | 69 if (!atomic_load_b(&mutex->locked, ATOMIC_RELAXED) in malloc_mutex_lock_slow() 87 ATOMIC_RELAXED) + 1; in malloc_mutex_lock_slow() 90 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); in malloc_mutex_lock_slow() 98 atomic_fetch_sub_u32(&data->n_waiting_thds, 1, ATOMIC_RELAXED); in malloc_mutex_lock_slow()
|
H A D | log.c | 66 ATOMIC_RELAXED); in log_var_update_state() 72 LOG_INITIALIZED_NOT_ENABLED, ATOMIC_RELAXED); in log_var_update_state()
|
H A D | extent.c | 297 atomic_store_zu(&extents->npages, 0, ATOMIC_RELAXED); in extents_init() 310 return atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_npages_get() 365 atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_insert_locked() 367 ATOMIC_RELAXED); in extents_insert_locked() 395 atomic_load_zu(&extents->npages, ATOMIC_RELAXED); in extents_remove_locked() 587 ATOMIC_RELAXED); in extents_evict() 753 ATOMIC_RELAXED) + nadd; in extent_gdump_add() 754 size_t high = atomic_load_zu(&highpages, ATOMIC_RELAXED); in extent_gdump_add() 756 &highpages, &high, cur, ATOMIC_RELAXED, ATOMIC_RELAXED)) { in extent_gdump_add() 776 atomic_fetch_sub_zu(&curpages, nsub, ATOMIC_RELAXED); in extent_gdump_sub() [all …]
|
H A D | ctl.c | 910 ATOMIC_RELAXED) == 0); in ctl_arena_stats_sdmerge() 1073 ATOMIC_RELAXED); in ctl_refresh() 1078 ATOMIC_RELAXED); in ctl_refresh() 2840 ATOMIC_RELAXED), 2871 ATOMIC_RELAXED), size_t) 3061 ATOMIC_RELAXED), size_t); 3065 ATOMIC_RELAXED), size_t); 3069 ATOMIC_RELAXED), size_t); 3073 ATOMIC_RELAXED), size_t); 3077 ATOMIC_RELAXED), size_t); [all …]
|
H A D | background_thread.c | 124 ssize_t decay_time = atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED); 678 ssize_t decay_time = atomic_load_zd(&decay->time_ms, ATOMIC_RELAXED);
|
H A D | base.c | 365 atomic_store_p(&base->extent_hooks, extent_hooks, ATOMIC_RELAXED); in base_new()
|
H A D | prof.c | 810 unsigned ngctxs = atomic_fetch_add_u(&cum_gctxs, 1, ATOMIC_RELAXED); in prof_gctx_mutex_choose() 2043 atomic_store_u64(&prof_accum->accumbytes, 0, ATOMIC_RELAXED); in prof_accum_init()
|
H A D | jemalloc.c | 390 ATOMIC_RELAXED); in arena_bind()
|