/freebsd/contrib/jemalloc/src/ |
H A D | arena.c | 100 + arena_stats_read_zu(tsdn, &arena->stats, &arena->stats.mapped)); in arena_stats_merge() 261 extents_dalloc(tsdn, arena, r_extent_hooks, &arena->extents_dirty, in arena_extents_dirty_dalloc() 812 return arena_decay_ms_set(tsdn, arena, &arena->decay_dirty, in arena_dirty_decay_ms_set() 819 return arena_decay_ms_set(tsdn, arena, &arena->decay_muzzy, in arena_muzzy_decay_ms_set() 900 arena_stats_sub_zu(tsdn, &arena->stats, &arena->stats.mapped, in arena_decay_stashed() 985 return arena_decay_impl(tsdn, arena, &arena->decay_dirty, in arena_decay_dirty() 992 return arena_decay_impl(tsdn, arena, &arena->decay_muzzy, in arena_decay_muzzy() 1519 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, size); in arena_malloc_hard() 1943 arena_t *arena; in arena_new() local 2102 return arena; in arena_new() [all …]
|
H A D | large.c | 28 assert(!tsdn_null(tsdn) || arena != NULL); in large_palloc() 45 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, usize); in large_palloc() 53 if (!arena_is_auto(arena)) { in large_palloc() 55 malloc_mutex_lock(tsdn, &arena->large_mtx); in large_palloc() 56 extent_list_append(&arena->large, extent); in large_palloc() 70 arena_decay_tick(tsdn, arena); in large_palloc() 97 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_shrink() local 133 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_expand() local 326 if (!arena_is_auto(arena)) { in large_dalloc_prep_impl() 335 if (!arena_is_auto(arena)) { in large_dalloc_prep_impl() [all …]
|
H A D | extent.c | 1254 arena_t *arena; in extent_alloc_default() local 1262 assert(arena != NULL); in extent_alloc_default() 1341 arena_ind_get(arena)); in extent_grow_retained() 1427 arena->extent_grow_next = arena->retain_grow_limit; in extent_grow_retained() 1863 extent_record(tsdn, arena, r_extent_hooks, &arena->extents_retained, in extent_dalloc_wrapper() 1968 arena_ind_get(arena))); in extent_decommit_wrapper() 2007 arena_ind_get(arena)); in extent_purge_lazy_impl() 2054 arena_ind_get(arena)); in extent_purge_forced_impl() 2148 arena_ind_get(arena)); in extent_split_impl() 2265 arena_ind_get(arena)); in extent_merge_impl() [all …]
|
H A D | tcache.c | 141 arena_t *arena = tcache->arena; in tcache_bin_flush_small() local 142 assert(arena != NULL); in tcache_bin_flush_small() 343 tcache->arena = arena; in tcache_arena_associate() 363 arena_t *arena = tcache->arena; in tcache_arena_dissociate() local 364 assert(arena != NULL); in tcache_arena_dissociate() 385 tcache->arena = NULL; in tcache_arena_dissociate() 414 tcache->arena = NULL; in tcache_init() 466 tcache->arena = NULL; in tsd_tcache_data_init() 467 arena_t *arena; in tsd_tcache_data_init() local 479 assert(arena == tcache->arena); in tsd_tcache_data_init() [all …]
|
H A D | extent_dss.c | 110 extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, in extent_alloc_dss() argument 126 gap = extent_alloc(tsdn, arena); in extent_alloc_dss() 156 extent_init(gap, arena, gap_addr_page, in extent_alloc_dss() 158 arena_extent_sn_next(arena), in extent_alloc_dss() 189 extent_dalloc_gap(tsdn, arena, gap); in extent_alloc_dss() 191 extent_dalloc(tsdn, arena, gap); in extent_alloc_dss() 201 extent_init(&extent, arena, ret, size, in extent_alloc_dss() 206 arena, &extent_hooks, &extent, 0, in extent_alloc_dss() 227 extent_dalloc(tsdn, arena, gap); in extent_alloc_dss()
|
H A D | jemalloc.c | 323 arena_t *arena; in arena_init_locked() local 340 return arena; in arena_init_locked() 346 return arena; in arena_init_locked() 369 arena_t *arena; in arena_init() local 377 return arena; in arena_init() 413 arena_t *arena; in arena_unbind() local 592 ret = arena; in arena_choose_hard() 628 arena_t *arena; in arena_cleanup() local 1919 arena_t *arena; in imalloc_no_sample() local 1943 arena = NULL; in imalloc_no_sample() [all …]
|
H A D | ctl.c | 2145 if (*arena == NULL || arena_is_auto(*arena)) { in arena_i_reset_destroy_helper() 2191 arena_t *arena; in arena_i_reset_ctl() local 2211 arena_t *arena; in arena_i_destroy_ctl() local 2220 if (arena_nthreads_get(arena, false) != 0 || arena_nthreads_get(arena, in arena_i_destroy_ctl() 2313 arena_t *arena; in arena_i_decay_ms_ctl_impl() local 2375 arena_t *arena; in arena_i_extent_hooks_ctl() local 2431 arena_t *arena; in arena_i_retain_grow_limit_ctl() local 2612 arena_t *arena; in arenas_lookup_ctl() local 2623 if (arena == NULL) in arenas_lookup_ctl() 2983 if (!arena) { in stats_mutexes_reset_ctl() [all …]
|
H A D | background_thread.c | 57 void background_thread_interval_check(tsdn_t *tsdn, arena_t *arena, in background_thread_create() 205 &arena->extents_dirty); 210 &arena->extents_muzzy); 291 arena_t *arena = arena_get(tsdn, i, false); 292 if (!arena) { 295 arena_decay(tsdn, arena, true, false); 301 arena); 657 background_thread_interval_check(tsdn_t *tsdn, arena_t *arena, 660 arena); 721 (extents_npages_get(&arena->extents_dirty) > 0 || [all …]
|
/freebsd/contrib/jemalloc/include/jemalloc/internal/ |
H A D | arena_externs.h | 43 ssize_t arena_dirty_decay_ms_get(arena_t *arena); 45 ssize_t arena_muzzy_decay_ms_get(arena_t *arena); 49 void arena_reset(tsd_t *tsd, arena_t *arena); 50 void arena_destroy(tsd_t *tsd, arena_t *arena); 74 dss_prec_t arena_dss_prec_get(arena_t *arena); 85 size_t arena_extent_sn_next(arena_t *arena); 93 void arena_prefork0(tsdn_t *tsdn, arena_t *arena); 94 void arena_prefork1(tsdn_t *tsdn, arena_t *arena); 95 void arena_prefork2(tsdn_t *tsdn, arena_t *arena); 96 void arena_prefork3(tsdn_t *tsdn, arena_t *arena); [all …]
|
H A D | extent_externs.h | 15 extent_t *extent_alloc(tsdn_t *tsdn, arena_t *arena); 18 extent_hooks_t *extent_hooks_get(arena_t *arena); 19 extent_hooks_t *extent_hooks_set(tsd_t *tsd, arena_t *arena, 38 extent_t *extents_alloc(tsdn_t *tsdn, arena_t *arena, 42 void extents_dalloc(tsdn_t *tsdn, arena_t *arena, 44 extent_t *extents_evict(tsdn_t *tsdn, arena_t *arena, 53 void extent_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena, 55 void extent_destroy_wrapper(tsdn_t *tsdn, arena_t *arena, 57 bool extent_commit_wrapper(tsdn_t *tsdn, arena_t *arena, 60 bool extent_decommit_wrapper(tsdn_t *tsdn, arena_t *arena, [all …]
|
H A D | jemalloc_internal_inlines_b.h | 11 if (arena != NULL) { in arena_choose_impl() 12 return arena; in arena_choose_impl() 26 if (tcache->arena != NULL) { in arena_choose_impl() 28 assert(tcache->arena == in arena_choose_impl() 30 if (tcache->arena != ret) { in arena_choose_impl() 62 arena_choose(tsd_t *tsd, arena_t *arena) { in arena_choose() argument 63 return arena_choose_impl(tsd, arena, false); in arena_choose() 67 arena_ichoose(tsd_t *tsd, arena_t *arena) { in arena_ichoose() argument 68 return arena_choose_impl(tsd, arena, true); in arena_ichoose() 72 arena_is_auto(arena_t *arena) { in arena_is_auto() argument [all …]
|
H A D | arena_inlines_a.h | 5 arena_ind_get(const arena_t *arena) { in arena_ind_get() argument 6 return base_ind_get(arena->base); in arena_ind_get() 10 arena_internal_add(arena_t *arena, size_t size) { in arena_internal_add() argument 11 atomic_fetch_add_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_add() 15 arena_internal_sub(arena_t *arena, size_t size) { in arena_internal_sub() argument 16 atomic_fetch_sub_zu(&arena->stats.internal, size, ATOMIC_RELAXED); in arena_internal_sub() 20 arena_internal_get(arena_t *arena) { in arena_internal_get() argument 21 return atomic_load_zu(&arena->stats.internal, ATOMIC_RELAXED); in arena_internal_get() 25 arena_prof_accum(tsdn_t *tsdn, arena_t *arena, uint64_t accumbytes) { in arena_prof_accum() argument 32 return prof_accum_add(tsdn, &arena->prof_accum, accumbytes); in arena_prof_accum()
|
H A D | arena_inlines_b.h | 12 arena_has_default_hooks(arena_t *arena) { in arena_has_default_hooks() argument 18 if (arena != NULL) { in arena_choose_maybe_huge() 19 return arena; in arena_choose_maybe_huge() 126 arena_decay(tsdn, arena, false, false); in arena_decay_ticks() 135 arena_decay_ticks(tsdn, arena, 1); in arena_decay_tick() 143 extent_dalloc_wrapper(tsdn, arena, in arena_decay_extent() 147 arena_stats_lock(tsdn, &arena->stats); in arena_decay_extent() 148 arena_stats_add_u64(tsdn, &arena->stats, in arena_decay_extent() 150 arena_stats_add_u64(tsdn, &arena->stats, in arena_decay_extent() 152 arena_stats_sub_zu(tsdn, &arena->stats, &arena->stats.mapped, in arena_decay_extent() [all …]
|
H A D | jemalloc_internal_inlines_c.h | 43 bool is_internal, arena_t *arena, bool slow_path) { in iallocztm() argument 47 assert(!is_internal || arena == NULL || arena_is_auto(arena)); in iallocztm() 53 ret = arena_malloc(tsdn, arena, size, ind, zero, tcache, slow_path); in iallocztm() 68 tcache_t *tcache, bool is_internal, arena_t *arena) { in ipallocztm() argument 74 assert(!is_internal || arena == NULL || arena_is_auto(arena)); in ipallocztm() 78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); in ipallocztm() 88 tcache_t *tcache, arena_t *arena) { in ipalloct() argument 89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); in ipalloct() 136 size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, in iralloct_realign() argument 147 p = ipalloct(tsdn, usize, alignment, zero, tcache, arena); in iralloct_realign() [all …]
|
H A D | background_thread_inlines.h | 16 arena_background_thread_info_get(arena_t *arena) { in arena_background_thread_info_get() argument 17 unsigned arena_ind = arena_ind_get(arena); in arena_background_thread_info_get() 49 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, in arena_background_thread_inactivity_check() argument 55 arena_background_thread_info_get(arena); in arena_background_thread_inactivity_check() 57 background_thread_interval_check(tsdn, arena, in arena_background_thread_inactivity_check() 58 &arena->decay_dirty, 0); in arena_background_thread_inactivity_check()
|
H A D | tcache_inlines.h | 42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small() argument 55 arena = arena_choose(tsd, arena); in tcache_alloc_small() 56 if (unlikely(arena == NULL)) { in tcache_alloc_small() 60 ret = tcache_alloc_small_hard(tsd_tsdn(tsd), arena, tcache, in tcache_alloc_small() 104 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_large() argument 119 arena = arena_choose(tsd, arena); in tcache_alloc_large() 120 if (unlikely(arena == NULL)) { in tcache_alloc_large() 124 ret = large_malloc(tsd_tsdn(tsd), arena, sz_s2u(size), zero); in tcache_alloc_large()
|
H A D | tcache_externs.h | 30 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 37 arena_t *arena); 40 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 45 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
|
H A D | large_externs.h | 6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, 11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
|
/freebsd/contrib/llvm-project/clang/lib/Analysis/FlowSensitive/ |
H A D | SimplifyConstraints.cpp | 20 Arena &arena) { in substitute() argument 30 return arena.makeNot(substitute(*F.operands()[0], Substitutions, arena)); in substitute() 32 return arena.makeAnd(substitute(*F.operands()[0], Substitutions, arena), in substitute() 35 return arena.makeOr(substitute(*F.operands()[0], Substitutions, arena), in substitute() 38 return arena.makeImplies( in substitute() 39 substitute(*F.operands()[0], Substitutions, arena), in substitute() 40 substitute(*F.operands()[1], Substitutions, arena)); in substitute() 42 return arena.makeEquals(substitute(*F.operands()[0], Substitutions, arena), in substitute() 79 Constraints.insert(&arena.makeLiteral(false)); in simplifyConstraints() 122 Substitutions.insert({TheAtom, &arena.makeLiteral(true)}); in simplifyConstraints() [all …]
|
H A D | DataflowAnalysisContext.cpp | 81 return arena().create<ScalarStorageLocation>(Type); in createStorageLocation() 152 Atom ForkToken = arena().makeFlowConditionToken(); in forkFlowCondition() 161 Atom Token = arena().makeFlowConditionToken(); in joinFlowConditions() 165 arena().makeOr(arena().makeAtomRef(FirstToken), in joinFlowConditions() 186 Constraints.insert(&arena().makeAtomRef(Token)); in flowConditionImplies() 187 Constraints.insert(&arena().makeNot(F)); in flowConditionImplies() 198 Constraints.insert(&arena().makeAtomRef(Token)); in flowConditionAllows() 207 Constraints.insert(&arena().makeNot(arena().makeEquals(Val1, Val2))); in equivalentFormulas() 231 Constraints.insert(&arena().makeEquals(arena().makeAtomRef(Token), in addTransitiveFlowConditionConstraints() 256 Constraints.insert(&arena().makeAtomRef(Token)); in dumpFlowCondition() [all …]
|
/freebsd/contrib/llvm-project/clang/include/clang/Analysis/FlowSensitive/ |
H A D | DataflowEnvironment.h | 513 return arena().makeIntLiteral(Value); in getIntLiteralValue() 519 return arena().makeBoolValue(arena().makeLiteral(Value)); in getBoolLiteralValue() 524 return arena().makeAtomValue(); in makeAtomicBoolValue() 529 return arena().makeTopValue(); in makeTopBoolValue() 537 return arena().makeBoolValue( in makeAnd() 546 return arena().makeBoolValue( in makeOr() 547 arena().makeOr(LHS.formula(), RHS.formula())); in makeOr() 553 return arena().makeBoolValue(arena().makeNot(Val.formula())); in makeNot() 561 return arena().makeBoolValue( in makeImplication() 570 return arena().makeBoolValue( in makeIff() [all …]
|
/freebsd/crypto/openssl/crypto/ |
H A D | mem_sec.c | 297 ((char*)(p) >= sh.arena && (char*)(p) < &sh.arena[sh.arena_size]) 312 char *arena; member 327 size_t bit = (sh.arena_size + ptr - sh.arena) / sh.minsize; in sh_getlist() 524 sh.arena = (char *)(sh.map_result + pgsize); in sh_init() 525 sh_setbit(sh.arena, 0, sh.bittable); in sh_init() 526 sh_add_to_list(&sh.freelist[0], sh.arena); in sh_init() 553 if (mlock(sh.arena, sh.arena_size) < 0) in sh_init() 560 if (VirtualLock(sh.arena, sh.arena_size) == FALSE) in sh_init() 563 if (mlock(sh.arena, sh.arena_size) < 0) in sh_init() 567 if (madvise(sh.arena, sh.arena_size, MADV_DONTDUMP) < 0) in sh_init() [all …]
|
/freebsd/sys/vm/ |
H A D | vm_glue.c | 304 vmem_t *arena; in vm_thread_alloc_kstack_kva() local 310 arena = vm_dom[domain].vmd_kernel_arena; in vm_thread_alloc_kstack_kva() 312 arena = vmd_kstack_arena[domain]; in vm_thread_alloc_kstack_kva() 314 rv = vmem_alloc(arena, size, M_BESTFIT | M_NOWAIT, &addr); in vm_thread_alloc_kstack_kva() 335 vmem_t *arena; in vm_thread_free_kstack_kva() local 339 arena = kernel_arena; in vm_thread_free_kstack_kva() 341 arena = vmd_kstack_arena[domain]; in vm_thread_free_kstack_kva() 343 arena = vm_dom[domain].vmd_kernel_arena; in vm_thread_free_kstack_kva() 346 vmem_free(arena, addr, size); in vm_thread_free_kstack_kva() 386 error = vmem_xalloc(arena, vm_thread_kstack_import_quantum(), in vm_thread_kstack_arena_import() [all …]
|
H A D | vm_kern.c | 468 vmem_t *arena; in kmem_malloc_domain() local 474 arena = vm_dom[domain].vmd_kernel_arena; in kmem_malloc_domain() 476 arena = vm_dom[domain].vmd_kernel_rwx_arena; in kmem_malloc_domain() 483 vmem_free(arena, addr, asize); in kmem_malloc_domain() 632 struct vmem *arena; in _kmem_unback() local 649 arena = vm_dom[domain].vmd_kernel_arena; in _kmem_unback() 651 arena = vm_dom[domain].vmd_kernel_rwx_arena; in _kmem_unback() 660 return (arena); in _kmem_unback() 679 struct vmem *arena; in kmem_free() local 684 if (arena != NULL) in kmem_free() [all …]
|
/freebsd/contrib/llvm-project/compiler-rt/lib/scudo/standalone/ |
H A D | wrappers_c.h | 23 __scudo_mallinfo_data_t arena; member 36 size_t arena; member
|