/freebsd/contrib/jemalloc/include/jemalloc/internal/ |
H A D | prof_inlines_b.h | 84 ssize_t check = update ? 0 : usize; in prof_sample_check() 88 bytes_until_sample -= usize; in prof_sample_check() 108 if (likely(prof_sample_check(tsd, usize, update))) { in prof_sample_accum_update() 151 assert(usize == sz_s2u(usize)); in prof_alloc_prep() 170 assert(usize == isalloc(tsdn, ptr)); in prof_malloc() 173 prof_malloc_sample_object(tsdn, ptr, usize, tctx); in prof_malloc() 175 prof_tctx_set(tsdn, ptr, usize, alloc_ctx, in prof_malloc() 190 assert(usize == isalloc(tsd_tsdn(tsd), ptr)); in prof_realloc() 211 prof_tctx_set(tsd_tsdn(tsd), ptr, usize, NULL, in prof_realloc() 243 assert(usize == isalloc(tsd_tsdn(tsd), ptr)); in prof_free() [all …]
|
H A D | safety_check.h | 9 safety_check_set_redzone(void *ptr, size_t usize, size_t bumped_usize) { in safety_check_set_redzone() argument 10 assert(usize < bumped_usize); in safety_check_set_redzone() 11 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { in safety_check_set_redzone() 17 safety_check_verify_redzone(const void *ptr, size_t usize, size_t bumped_usize) in safety_check_verify_redzone() argument 19 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { in safety_check_verify_redzone()
|
H A D | sz.h | 117 return usize; in sz_psz2u() 193 return usize; in sz_index2size_compute() 234 return usize; in sz_s2u_compute() 264 size_t usize; in sz_sa2u() local 285 if (usize < SC_LARGE_MINCLASS) { in sz_sa2u() 286 return usize; in sz_sa2u() 298 usize = SC_LARGE_MINCLASS; in sz_sa2u() 300 usize = sz_s2u(size); in sz_sa2u() 301 if (usize < size) { in sz_sa2u() 311 if (usize + sz_large_pad + PAGE_CEILING(alignment) - PAGE < usize) { in sz_sa2u() [all …]
|
H A D | tcache_inlines.h | 47 size_t usize JEMALLOC_CC_SILENCE_INIT(0); in tcache_alloc_small() 73 usize = sz_index2size(binind); in tcache_alloc_small() 83 memset(ret, 0, usize); in tcache_alloc_small() 90 memset(ret, 0, usize); in tcache_alloc_small() 97 tcache->prof_accumbytes += usize; in tcache_alloc_small() 134 usize = sz_index2size(binind); in tcache_alloc_large() 135 assert(usize <= tcache_maxclass); in tcache_alloc_large() 142 usize); in tcache_alloc_large() 144 memset(ret, 0, usize); in tcache_alloc_large() 148 memset(ret, 0, usize); in tcache_alloc_large() [all …]
|
H A D | jemalloc_internal_inlines_c.h | 67 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipallocztm() argument 71 assert(usize != 0); in ipallocztm() 72 assert(usize == sz_sa2u(usize, alignment)); in ipallocztm() 78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); in ipallocztm() 87 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipalloct() argument 93 ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero) { in ipalloc() argument 94 return ipallocztm(tsd_tsdn(tsd), usize, alignment, zero, in ipalloc() 141 size_t usize, copysize; in iralloct_realign() local 143 usize = sz_sa2u(size, alignment); in iralloct_realign() 144 if (unlikely(usize == 0 || usize > SC_LARGE_MAXCLASS)) { in iralloct_realign() [all …]
|
H A D | prof_inlines_a.h | 47 size_t usize) { in prof_accum_cancel() argument 60 a1 = (a0 >= SC_LARGE_MINCLASS - usize) in prof_accum_cancel() 61 ? a0 - (SC_LARGE_MINCLASS - usize) : 0; in prof_accum_cancel() 67 a1 = (a0 >= SC_LARGE_MINCLASS - usize) in prof_accum_cancel() 68 ? a0 - (SC_LARGE_MINCLASS - usize) : 0; in prof_accum_cancel()
|
H A D | large_externs.h | 6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, 11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
|
H A D | arena_externs.h | 36 size_t usize, size_t alignment, bool *zero); 61 void *arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, 63 void arena_prof_promote(tsdn_t *tsdn, void *ptr, size_t usize);
|
H A D | prof_externs.h | 47 void prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, 49 void prof_free_sampled_object(tsd_t *tsd, const void *ptr, size_t usize,
|
H A D | arena_inlines_b.h | 57 arena_prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, in arena_prof_tctx_set() argument
|
/freebsd/contrib/jemalloc/src/ |
H A D | large.c | 15 assert(usize == sz_s2u(usize)); in large_malloc() 30 ausize = sz_sa2u(usize, alignment); in large_palloc() 102 assert(oldusize > usize); in large_ralloc_no_move_shrink() 136 size_t trailsize = usize - oldusize; in large_ralloc_no_move_expand() 181 szind_t szind = sz_size2index(usize); in large_ralloc_no_move_expand() 210 JEMALLOC_ALLOC_JUNK, usize - oldusize); in large_ralloc_no_move_expand() 281 assert(usize > 0 && usize <= SC_LARGE_MAXCLASS); in large_ralloc() 284 && usize >= SC_LARGE_MINCLASS); in large_ralloc() 287 if (!large_ralloc_no_move(tsdn, extent, usize, usize, zero)) { in large_ralloc() 290 usize, (uintptr_t)ptr, hook_args->args); in large_ralloc() [all …]
|
H A D | jemalloc.c | 1864 bool usize; member 1893 size_t usize; member 2072 dopts->usize = usize; in imalloc_body() 2073 assert(usize > 0 && usize in imalloc_body() 2083 dopts->usize = usize; in imalloc_body() 2128 sopts, dopts, tsd, usize, usize, ind); in imalloc_body() 2346 size_t usize; in JEMALLOC_ATTR() local 2578 size_t usize; in ifree() local 3204 size_t usize; in je_rallocx() local 3312 size_t usize; in ixallocx_prof_sample() local [all …]
|
H A D | arena.c | 390 usize = SC_LARGE_MINCLASS; in arena_large_malloc_stats_update() 417 size_t usize) { in arena_large_ralloc_stats_update() argument 1117 size_t usize; in arena_reset() local 1463 size_t usize; in arena_malloc_small() local 1498 memset(ret, 0, usize); in arena_malloc_small() 1506 memset(ret, 0, usize); in arena_malloc_small() 1540 ret = arena_malloc(tsdn, arena, usize, sz_size2index(usize), in arena_palloc() 1789 return arena_malloc(tsdn, arena, usize, sz_size2index(usize), in arena_ralloc_move_helper() 1792 usize = sz_sa2u(usize, alignment); in arena_ralloc_move_helper() 1793 if (unlikely(usize == 0 || usize > SC_LARGE_MAXCLASS)) { in arena_ralloc_move_helper() [all …]
|
H A D | ckh.c | 274 size_t usize; in ckh_grow() local 277 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); in ckh_grow() 278 if (unlikely(usize == 0 in ckh_grow() 279 || usize > SC_LARGE_MAXCLASS)) { in ckh_grow() 283 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, in ckh_grow() 314 size_t usize; in ckh_shrink() local 323 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); in ckh_shrink() 324 if (unlikely(usize == 0 || usize > SC_LARGE_MAXCLASS)) { in ckh_shrink() 363 size_t mincells, usize; in ckh_new() local 399 usize = sz_sa2u(sizeof(ckhc_t) << lg_mincells, CACHELINE); in ckh_new() [all …]
|
H A D | base.c | 252 size_t usize = ALIGNMENT_CEILING(size, alignment); in base_block_alloc() local 264 + usize)); in base_block_alloc() 424 size_t usize = ALIGNMENT_CEILING(size, alignment); in base_alloc_impl() local 425 size_t asize = usize + alignment - QUANTUM; in base_alloc_impl() 438 extent = base_extent_alloc(tsdn, base, usize, alignment); in base_alloc_impl() 446 ret = base_extent_bump_alloc(base, extent, usize, alignment); in base_alloc_impl()
|
H A D | prof.c | 146 size_t usize; member 343 prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, in prof_malloc_sample_object() argument 345 prof_tctx_set(tsdn, ptr, usize, NULL, tctx); in prof_malloc_sample_object() 355 tctx->cnts.curbytes += usize; in prof_malloc_sample_object() 358 tctx->cnts.accumbytes += usize; in prof_malloc_sample_object() 506 new_node->usize = usize; in prof_try_log() 521 prof_free_sampled_object(tsd_t *tsd, const void *ptr, size_t usize, in prof_free_sampled_object() argument 526 assert(tctx->cnts.curbytes >= usize); in prof_free_sampled_object() 528 tctx->cnts.curbytes -= usize; in prof_free_sampled_object() 530 prof_try_log(tsd, ptr, usize, tctx); in prof_free_sampled_object() [all …]
|
/freebsd/contrib/libarchive/libarchive/ |
H A D | archive_write_add_filter_lzop.c | 321 lzo_uint usize, csize; in drive_compressor() local 331 usize = (lzo_uint) in drive_compressor() 337 r = lzo1x_1_compress(data->uncompressed, usize, in drive_compressor() 342 r = lzo1x_1_15_compress(data->uncompressed, usize, in drive_compressor() 347 r = lzo1x_999_compress_level(data->uncompressed, usize, in drive_compressor() 359 archive_be32enc(p + header_bytes, (uint32_t)usize); in drive_compressor() 361 checksum = lzo_adler32(1, data->uncompressed, usize); in drive_compressor() 364 if (csize < usize) { in drive_compressor() 374 archive_be32enc(p + header_bytes + 4, (uint32_t)usize); in drive_compressor() 380 usize); in drive_compressor()
|
/freebsd/bin/ed/ |
H A D | undo.c | 34 static long usize = 0; /* stack size variable */ variable 45 (ustack = (undo_t *) malloc((usize = USIZE) * sizeof(undo_t))) == NULL) { in push_undo_stack() 52 if (u_p < usize || in push_undo_stack() 53 (t = (undo_t *) realloc(ustack, (usize += USIZE) * sizeof(undo_t))) != NULL) { in push_undo_stack() 66 usize = 0; in push_undo_stack()
|
/freebsd/sys/contrib/openzfs/tests/zfs-tests/tests/functional/mount/ |
H A D | umount_unlinked_drain.ksh | 52 usize=$(($nunlinks - $nunlinked)) 53 if [[ $iters == $MAX_ITERS && $usize == $1 ]]; then 56 if [[ $usize == $last_usize ]]; then 61 last_usize=$usize
|
/freebsd/usr.bin/gzip/ |
H A D | gzip.c | 1750 off_t usize, gsize; in handle_stdin() local 1856 off_t usize; in handle_stdout() local 1888 usize = in handle_stdout() 1953 off_t usize, gsize; in handle_file() local 1960 print_test(file, usize != -1); in handle_file() 1961 if (usize == -1) in handle_file() 1968 usize = sbp->st_size; in handle_file() 2059 print_ratio(usize, gsize, stderr); in print_verbage() 2110 uint32_t usize; in print_list() local 2119 usize = le32dec(&buf[4]); in print_list() [all …]
|
/freebsd/sys/compat/linux/ |
H A D | linux_fork.c | 437 if (args->usize > PAGE_SIZE) in linux_clone3() 439 if (args->usize < LINUX_CLONE_ARGS_SIZE_VER0) in linux_clone3() 447 size = max(args->usize, sizeof(*uca)); in linux_clone3() 449 error = copyin(args->uargs, uca, args->usize); in linux_clone3()
|
/freebsd/contrib/jemalloc/ |
H A D | FREEBSD-diffs | 446 + size_t usize = je_xallocx(*ptr, size, extra, flags); 447 + ret = (usize >= size) ? ALLOCM_SUCCESS : ALLOCM_ERR_NOT_MOVED; 449 + *rsize = usize; 481 + size_t usize = je_nallocx(size, flags); 482 + if (usize == 0) { 486 + *rsize = usize;
|
/freebsd/sys/contrib/openzfs/cmd/ |
H A D | dbufstat.in | 403 usize = int(line[labels['usize']]) 421 d[pool][objset][key]['cached'] += dbsize + usize
|
/freebsd/sys/dev/iwi/ |
H A D | if_iwireg.h | 165 uint32_t usize; /* size of ucode image */ member
|
/freebsd/sys/dev/qat_c2xxx/ |
H A D | qat_ae.c | 1420 int ii, vali, fixup, usize = 0; in qat_ae_concat_ucode() local 1432 usize = nitems(ae_inst_1b); in qat_ae_concat_ucode() 1436 usize = nitems(ae_inst_2b); in qat_ae_concat_ucode() 1440 usize = nitems(ae_inst_3b); in qat_ae_concat_ucode() 1444 usize = nitems(ae_inst_4b); in qat_ae_concat_ucode() 1449 for (ii = 0; ii < usize; ii++) in qat_ae_concat_ucode()
|