Home
last modified time | relevance | path

Searched refs:arena (Results 1 – 25 of 47) sorted by relevance

12

/freebsd/contrib/jemalloc/src/
H A Darena.c100 + arena_stats_read_zu(tsdn, &arena->stats, &arena->stats.mapped)); in arena_stats_merge()
261 extents_dalloc(tsdn, arena, r_extent_hooks, &arena->extents_dirty, in arena_extents_dirty_dalloc()
812 return arena_decay_ms_set(tsdn, arena, &arena->decay_dirty, in arena_dirty_decay_ms_set()
819 return arena_decay_ms_set(tsdn, arena, &arena->decay_muzzy, in arena_muzzy_decay_ms_set()
900 arena_stats_sub_zu(tsdn, &arena->stats, &arena->stats.mapped, in arena_decay_stashed()
985 return arena_decay_impl(tsdn, arena, &arena->decay_dirty, in arena_decay_dirty()
992 return arena_decay_impl(tsdn, arena, &arena->decay_muzzy, in arena_decay_muzzy()
1519 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, size); in arena_malloc_hard()
1943 arena_t *arena; in arena_new() local
2102 return arena; in arena_new()
[all …]
H A Dlarge.c28 assert(!tsdn_null(tsdn) || arena != NULL); in large_palloc()
45 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, usize); in large_palloc()
53 if (!arena_is_auto(arena)) { in large_palloc()
55 malloc_mutex_lock(tsdn, &arena->large_mtx); in large_palloc()
56 extent_list_append(&arena->large, extent); in large_palloc()
70 arena_decay_tick(tsdn, arena); in large_palloc()
97 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_shrink() local
133 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_expand() local
326 if (!arena_is_auto(arena)) { in large_dalloc_prep_impl()
335 if (!arena_is_auto(arena)) { in large_dalloc_prep_impl()
[all …]
H A Dextent.c1254 arena_t *arena; in extent_alloc_default() local
1262 assert(arena != NULL); in extent_alloc_default()
1341 arena_ind_get(arena)); in extent_grow_retained()
1427 arena->extent_grow_next = arena->retain_grow_limit; in extent_grow_retained()
1863 extent_record(tsdn, arena, r_extent_hooks, &arena->extents_retained, in extent_dalloc_wrapper()
1968 arena_ind_get(arena))); in extent_decommit_wrapper()
2007 arena_ind_get(arena)); in extent_purge_lazy_impl()
2054 arena_ind_get(arena)); in extent_purge_forced_impl()
2148 arena_ind_get(arena)); in extent_split_impl()
2265 arena_ind_get(arena)); in extent_merge_impl()
[all …]
H A Dtcache.c141 arena_t *arena = tcache->arena; in tcache_bin_flush_small() local
142 assert(arena != NULL); in tcache_bin_flush_small()
343 tcache->arena = arena; in tcache_arena_associate()
363 arena_t *arena = tcache->arena; in tcache_arena_dissociate() local
364 assert(arena != NULL); in tcache_arena_dissociate()
385 tcache->arena = NULL; in tcache_arena_dissociate()
414 tcache->arena = NULL; in tcache_init()
466 tcache->arena = NULL; in tsd_tcache_data_init()
467 arena_t *arena; in tsd_tcache_data_init() local
479 assert(arena == tcache->arena); in tsd_tcache_data_init()
[all …]
H A Dextent_dss.c110 extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, in extent_alloc_dss() argument
126 gap = extent_alloc(tsdn, arena); in extent_alloc_dss()
156 extent_init(gap, arena, gap_addr_page, in extent_alloc_dss()
158 arena_extent_sn_next(arena), in extent_alloc_dss()
189 extent_dalloc_gap(tsdn, arena, gap); in extent_alloc_dss()
191 extent_dalloc(tsdn, arena, gap); in extent_alloc_dss()
201 extent_init(&extent, arena, ret, size, in extent_alloc_dss()
206 arena, &extent_hooks, &extent, 0, in extent_alloc_dss()
227 extent_dalloc(tsdn, arena, gap); in extent_alloc_dss()
H A Djemalloc.c323 arena_t *arena; in arena_init_locked() local
340 return arena; in arena_init_locked()
346 return arena; in arena_init_locked()
369 arena_t *arena; in arena_init() local
377 return arena; in arena_init()
413 arena_t *arena; in arena_unbind() local
592 ret = arena; in arena_choose_hard()
628 arena_t *arena; in arena_cleanup() local
1919 arena_t *arena; in imalloc_no_sample() local
1943 arena = NULL; in imalloc_no_sample()
[all …]
H A Dctl.c2145 if (*arena == NULL || arena_is_auto(*arena)) { in arena_i_reset_destroy_helper()
2191 arena_t *arena; in arena_i_reset_ctl() local
2211 arena_t *arena; in arena_i_destroy_ctl() local
2220 if (arena_nthreads_get(arena, false) != 0 || arena_nthreads_get(arena, in arena_i_destroy_ctl()
2313 arena_t *arena; in arena_i_decay_ms_ctl_impl() local
2375 arena_t *arena; in arena_i_extent_hooks_ctl() local
2431 arena_t *arena; in arena_i_retain_grow_limit_ctl() local
2612 arena_t *arena; in arenas_lookup_ctl() local
2623 if (arena == NULL) in arenas_lookup_ctl()
2983 if (!arena) { in stats_mutexes_reset_ctl()
[all …]
H A Dbackground_thread.c57 void background_thread_interval_check(tsdn_t *tsdn, arena_t *arena, in background_thread_create()
205 &arena->extents_dirty);
210 &arena->extents_muzzy);
291 arena_t *arena = arena_get(tsdn, i, false);
292 if (!arena) {
295 arena_decay(tsdn, arena, true, false);
301 arena);
657 background_thread_interval_check(tsdn_t *tsdn, arena_t *arena,
660 arena);
721 (extents_npages_get(&arena->extents_dirty) > 0 ||
[all …]
/freebsd/contrib/jemalloc/include/jemalloc/internal/
H A Darena_externs.h43 ssize_t arena_dirty_decay_ms_get(arena_t *arena);
45 ssize_t arena_muzzy_decay_ms_get(arena_t *arena);
49 void arena_reset(tsd_t *tsd, arena_t *arena);
50 void arena_destroy(tsd_t *tsd, arena_t *arena);
74 dss_prec_t arena_dss_prec_get(arena_t *arena);
85 size_t arena_extent_sn_next(arena_t *arena);
93 void arena_prefork0(tsdn_t *tsdn, arena_t *arena);
94 void arena_prefork1(tsdn_t *tsdn, arena_t *arena);
95 void arena_prefork2(tsdn_t *tsdn, arena_t *arena);
96 void arena_prefork3(tsdn_t *tsdn, arena_t *arena);
[all …]
H A Dextent_externs.h15 extent_t *extent_alloc(tsdn_t *tsdn, arena_t *arena);
18 extent_hooks_t *extent_hooks_get(arena_t *arena);
19 extent_hooks_t *extent_hooks_set(tsd_t *tsd, arena_t *arena,
38 extent_t *extents_alloc(tsdn_t *tsdn, arena_t *arena,
42 void extents_dalloc(tsdn_t *tsdn, arena_t *arena,
44 extent_t *extents_evict(tsdn_t *tsdn, arena_t *arena,
53 void extent_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena,
55 void extent_destroy_wrapper(tsdn_t *tsdn, arena_t *arena,
57 bool extent_commit_wrapper(tsdn_t *tsdn, arena_t *arena,
60 bool extent_decommit_wrapper(tsdn_t *tsdn, arena_t *arena,
[all …]
H A Djemalloc_internal_inlines_b.h11 if (arena != NULL) { in arena_choose_impl()
12 return arena; in arena_choose_impl()
26 if (tcache->arena != NULL) { in arena_choose_impl()
28 assert(tcache->arena == in arena_choose_impl()
30 if (tcache->arena != ret) { in arena_choose_impl()
62 arena_choose(tsd_t *tsd, arena_t *arena) { in arena_choose() argument
63 return arena_choose_impl(tsd, arena, false); in arena_choose()
67 arena_ichoose(tsd_t *tsd, arena_t *arena) { in arena_ichoose() argument
68 return arena_choose_impl(tsd, arena, true); in arena_ichoose()
72 arena_is_auto(arena_t *arena) { in arena_is_auto() argument
[all …]
H A Darena_inlines_a.h5 arena_ind_get(const arena_t *arena) { in arena_ind_get() argument
6 return base_ind_get(arena->base); in arena_ind_get()
10 arena_internal_add(arena_t *arena, size_t size) { in arena_internal_add() argument
11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_add()
15 arena_internal_sub(arena_t *arena, size_t size) { in arena_internal_sub() argument
16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_sub()
20 arena_internal_get(arena_t *arena) { in arena_internal_get() argument
21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED); in arena_internal_get()
25 arena_prof_accum(tsdn_t *tsdn, arena_t *arena, uint64_t accumbytes) { in arena_prof_accum() argument
32 return prof_accum_add(tsdn, &arena->prof_accum, accumbytes); in arena_prof_accum()
H A Darena_inlines_b.h12 arena_has_default_hooks(arena_t *arena) { in arena_has_default_hooks() argument
18 if (arena != NULL) { in arena_choose_maybe_huge()
19 return arena; in arena_choose_maybe_huge()
126 arena_decay(tsdn, arena, false, false); in arena_decay_ticks()
135 arena_decay_ticks(tsdn, arena, 1); in arena_decay_tick()
143 extent_dalloc_wrapper(tsdn, arena, in arena_decay_extent()
147 arena_stats_lock(tsdn, &arena->stats); in arena_decay_extent()
148 arena_stats_add_u64(tsdn, &arena->stats, in arena_decay_extent()
150 arena_stats_add_u64(tsdn, &arena->stats, in arena_decay_extent()
152 arena_stats_sub_zu(tsdn, &arena->stats, &arena->stats.mapped, in arena_decay_extent()
[all …]
H A Djemalloc_internal_inlines_c.h43 bool is_internal, arena_t *arena, bool slow_path) { in iallocztm() argument
47 assert(!is_internal || arena == NULL || arena_is_auto(arena)); in iallocztm()
53 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path); in iallocztm()
68 tcache_t *tcache, bool is_internal, arena_t *arena) { in ipallocztm() argument
74 assert(!is_internal || arena == NULL || arena_is_auto(arena)); in ipallocztm()
78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); in ipallocztm()
88 tcache_t *tcache, arena_t *arena) { in ipalloct() argument
89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); in ipalloct()
136 size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, in iralloct_realign() argument
147 p = ipalloct(tsdn, usize, alignment, zero, tcache, arena); in iralloct_realign()
[all …]
H A Dbackground_thread_inlines.h16 arena_background_thread_info_get(arena_t *arena) { in arena_background_thread_info_get() argument
17 unsigned arena_ind = arena_ind_get(arena); in arena_background_thread_info_get()
49 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, in arena_background_thread_inactivity_check() argument
55 arena_background_thread_info_get(arena); in arena_background_thread_inactivity_check()
57 background_thread_interval_check(tsdn, arena, in arena_background_thread_inactivity_check()
58 &arena->decay_dirty, 0); in arena_background_thread_inactivity_check()
H A Dtcache_inlines.h42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small() argument
55 arena = arena_choose(tsd, arena); in tcache_alloc_small()
56 if (unlikely(arena == NULL)) { in tcache_alloc_small()
60 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache, in tcache_alloc_small()
104 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_large() argument
119 arena = arena_choose(tsd, arena); in tcache_alloc_large()
120 if (unlikely(arena == NULL)) { in tcache_alloc_large()
124 ret = large_malloc(tsd_tsdn(tsd), arena, sz_s2u(size), zero); in tcache_alloc_large()
H A Dtcache_externs.h30 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
37 arena_t *arena);
40 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
45 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
H A Dlarge_externs.h6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero);
7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment,
11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
/freebsd/contrib/llvm-project/clang/lib/Analysis/FlowSensitive/
H A DSimplifyConstraints.cpp20 Arena &arena) { in substitute() argument
30 return arena.makeNot(substitute(*F.operands()[0], Substitutions, arena)); in substitute()
32 return arena.makeAnd(substitute(*F.operands()[0], Substitutions, arena), in substitute()
35 return arena.makeOr(substitute(*F.operands()[0], Substitutions, arena), in substitute()
38 return arena.makeImplies( in substitute()
39 substitute(*F.operands()[0], Substitutions, arena), in substitute()
40 substitute(*F.operands()[1], Substitutions, arena)); in substitute()
42 return arena.makeEquals(substitute(*F.operands()[0], Substitutions, arena), in substitute()
79 Constraints.insert(&arena.makeLiteral(false)); in simplifyConstraints()
122 Substitutions.insert({TheAtom, &arena.makeLiteral(true)}); in simplifyConstraints()
[all …]
H A DDataflowAnalysisContext.cpp81 return arena().create<ScalarStorageLocation>(Type); in createStorageLocation()
152 Atom ForkToken = arena().makeFlowConditionToken(); in forkFlowCondition()
161 Atom Token = arena().makeFlowConditionToken(); in joinFlowConditions()
165 arena().makeOr(arena().makeAtomRef(FirstToken), in joinFlowConditions()
186 Constraints.insert(&arena().makeAtomRef(Token)); in flowConditionImplies()
187 Constraints.insert(&arena().makeNot(F)); in flowConditionImplies()
198 Constraints.insert(&arena().makeAtomRef(Token)); in flowConditionAllows()
207 Constraints.insert(&arena().makeNot(arena().makeEquals(Val1, Val2))); in equivalentFormulas()
231 Constraints.insert(&arena().makeEquals(arena().makeAtomRef(Token), in addTransitiveFlowConditionConstraints()
256 Constraints.insert(&arena().makeAtomRef(Token)); in dumpFlowCondition()
[all …]
/freebsd/contrib/llvm-project/clang/include/clang/Analysis/FlowSensitive/
H A DDataflowEnvironment.h513 return arena().makeIntLiteral(Value); in getIntLiteralValue()
519 return arena().makeBoolValue(arena().makeLiteral(Value)); in getBoolLiteralValue()
524 return arena().makeAtomValue(); in makeAtomicBoolValue()
529 return arena().makeTopValue(); in makeTopBoolValue()
537 return arena().makeBoolValue( in makeAnd()
546 return arena().makeBoolValue( in makeOr()
547 arena().makeOr(LHS.formula(), RHS.formula())); in makeOr()
553 return arena().makeBoolValue(arena().makeNot(Val.formula())); in makeNot()
561 return arena().makeBoolValue( in makeImplication()
570 return arena().makeBoolValue( in makeIff()
[all …]
/freebsd/crypto/openssl/crypto/
H A Dmem_sec.c297 ((char*)(p) >= sh.arena && (char*)(p) < &sh.arena[sh.arena_size])
312 char *arena; member
327 size_t bit = (sh.arena_size + ptr - sh.arena) / sh.minsize; in sh_getlist()
524 sh.arena = (char *)(sh.map_result + pgsize); in sh_init()
525 sh_setbit(sh.arena, 0, sh.bittable); in sh_init()
526 sh_add_to_list(&sh.freelist[0], sh.arena); in sh_init()
553 if (mlock(sh.arena, sh.arena_size) < 0) in sh_init()
560 if (VirtualLock(sh.arena, sh.arena_size) == FALSE) in sh_init()
563 if (mlock(sh.arena, sh.arena_size) < 0) in sh_init()
567 if (madvise(sh.arena, sh.arena_size, MADV_DONTDUMP) < 0) in sh_init()
[all …]
/freebsd/sys/vm/
H A Dvm_glue.c304 vmem_t *arena; in vm_thread_alloc_kstack_kva() local
310 arena = vm_dom[domain].vmd_kernel_arena; in vm_thread_alloc_kstack_kva()
312 arena = vmd_kstack_arena[domain]; in vm_thread_alloc_kstack_kva()
314 rv = vmem_alloc(arena, size, M_BESTFIT | M_NOWAIT, &addr); in vm_thread_alloc_kstack_kva()
335 vmem_t *arena; in vm_thread_free_kstack_kva() local
339 arena = kernel_arena; in vm_thread_free_kstack_kva()
341 arena = vmd_kstack_arena[domain]; in vm_thread_free_kstack_kva()
343 arena = vm_dom[domain].vmd_kernel_arena; in vm_thread_free_kstack_kva()
346 vmem_free(arena, addr, size); in vm_thread_free_kstack_kva()
386 error = vmem_xalloc(arena, vm_thread_kstack_import_quantum(), in vm_thread_kstack_arena_import()
[all …]
H A Dvm_kern.c468 vmem_t *arena; in kmem_malloc_domain() local
474 arena = vm_dom[domain].vmd_kernel_arena; in kmem_malloc_domain()
476 arena = vm_dom[domain].vmd_kernel_rwx_arena; in kmem_malloc_domain()
483 vmem_free(arena, addr, asize); in kmem_malloc_domain()
632 struct vmem *arena; in _kmem_unback() local
649 arena = vm_dom[domain].vmd_kernel_arena; in _kmem_unback()
651 arena = vm_dom[domain].vmd_kernel_rwx_arena; in _kmem_unback()
660 return (arena); in _kmem_unback()
679 struct vmem *arena; in kmem_free() local
684 if (arena != NULL) in kmem_free()
[all …]
/freebsd/contrib/llvm-project/compiler-rt/lib/scudo/standalone/
H A Dwrappers_c.h23 __scudo_mallinfo_data_t arena; member
36 size_t arena; member

12