/dragonfly/contrib/zstd/lib/compress/ |
H A D | zstd_cwksp.h | 164 assert(ws->workspace <= ws->objectEnd); in ZSTD_cwksp_assert_internal_consistency() 165 assert(ws->objectEnd <= ws->tableEnd); in ZSTD_cwksp_assert_internal_consistency() 167 assert(ws->tableEnd <= ws->allocStart); in ZSTD_cwksp_assert_internal_consistency() 383 ws->tableValidEnd = ws->objectEnd; in ZSTD_cwksp_mark_tables_dirty() 391 if (ws->tableValidEnd < ws->tableEnd) { in ZSTD_cwksp_mark_tables_clean() 392 ws->tableValidEnd = ws->tableEnd; in ZSTD_cwksp_mark_tables_clean() 428 ws->tableEnd = ws->objectEnd; in ZSTD_cwksp_clear_tables() 461 ws->tableEnd = ws->objectEnd; in ZSTD_cwksp_clear() 462 ws->allocStart = ws->workspaceEnd; in ZSTD_cwksp_clear() 480 ws->objectEnd = ws->workspace; in ZSTD_cwksp_init() [all …]
|
/dragonfly/contrib/gcc-4.7/libgomp/ |
H A D | work.c | 48 return ws; in alloc_work_share() 57 if (ws && ws->next_free) in alloc_work_share() 67 if (ws) in alloc_work_share() 72 return ws; in alloc_work_share() 83 ws[i].next_free = &ws[i + 1]; in alloc_work_share() 85 return ws; in alloc_work_share() 107 ws->ordered_team_ids = ws->inline_ordered_team_ids; in gomp_init_work_share() 127 if (ws->ordered_team_ids != ws->inline_ordered_team_ids) in gomp_fini_work_share() 140 free (ws); in free_work_share() 177 ws = gomp_malloc (sizeof (*ws)); in gomp_work_share_start() [all …]
|
H A D | iter_ull.c | 55 return ws->next_ull == ws->end_ull; in gomp_iter_ull_static_next() 70 n = (ws->end_ull - ws->next_ull + ws->incr_ull - 1) / ws->incr_ull; in gomp_iter_ull_static_next() 72 n = (ws->next_ull - ws->end_ull - ws->incr_ull - 1) / -ws->incr_ull; in gomp_iter_ull_static_next() 95 s = s0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 96 e = e0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 111 n = (ws->end_ull - ws->next_ull + ws->incr_ull - 1) / ws->incr_ull; in gomp_iter_ull_static_next() 113 n = (ws->next_ull - ws->end_ull - ws->incr_ull - 1) / -ws->incr_ull; in gomp_iter_ull_static_next() 129 s = s0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 130 e = e0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 270 if (ws->next_ull == ws->end_ull) in gomp_iter_ull_guided_next_locked() [all …]
|
H A D | iter.c | 54 return ws->next == ws->end; in gomp_iter_static_next() 70 s = ws->incr + (ws->incr > 0 ? -1 : 1); in gomp_iter_static_next() 71 n = (ws->end - ws->next + s) / ws->incr; in gomp_iter_static_next() 94 s = (long)s0 * ws->incr + ws->next; in gomp_iter_static_next() 95 e = (long)e0 * ws->incr + ws->next; in gomp_iter_static_next() 110 s = ws->incr + (ws->incr > 0 ? -1 : 1); in gomp_iter_static_next() 111 n = (ws->end - ws->next + s) / ws->incr; in gomp_iter_static_next() 127 s = (long)s0 * ws->incr + ws->next; in gomp_iter_static_next() 128 e = (long)e0 * ws->incr + ws->next; in gomp_iter_static_next() 269 if (ws->next == ws->end) in gomp_iter_guided_next_locked() [all …]
|
H A D | ordered.c | 46 index = ws->ordered_cur + ws->ordered_num_used; in gomp_ordered_first() 78 ws->ordered_owner = -1; in gomp_ordered_last() 81 if (--ws->ordered_num_used > 0) in gomp_ordered_last() 86 ws->ordered_cur = next; in gomp_ordered_last() 113 ws->ordered_owner = -1; in gomp_ordered_next() 116 if (ws->ordered_num_used == 1) in gomp_ordered_next() 129 index = ws->ordered_cur + ws->ordered_num_used; in gomp_ordered_next() 135 index = ws->ordered_cur + 1; in gomp_ordered_next() 138 ws->ordered_cur = index; in gomp_ordered_next() 177 ws->ordered_owner = -1; in gomp_ordered_static_next() [all …]
|
H A D | sections.c | 33 gomp_sections_init (struct gomp_work_share *ws, unsigned count) in gomp_sections_init() argument 35 ws->sched = GFS_DYNAMIC; in gomp_sections_init() 36 ws->chunk_size = 1; in gomp_sections_init() 37 ws->end = count + 1L; in gomp_sections_init() 38 ws->incr = 1; in gomp_sections_init() 39 ws->next = 1; in gomp_sections_init() 43 ws->mode = 1; in gomp_sections_init() 50 ws->mode = ((nthreads | ws->end) in gomp_sections_init() 54 ws->mode = 0; in gomp_sections_init()
|
H A D | loop.c | 38 ws->sched = sched; in gomp_loop_init() 39 ws->chunk_size = chunk_size; in gomp_loop_init() 43 ws->incr = incr; in gomp_loop_init() 44 ws->next = start; in gomp_loop_init() 47 ws->chunk_size *= incr; in gomp_loop_init() 60 if (__builtin_expect ((nthreads | ws->chunk_size) in gomp_loop_init() 63 ws->mode = 0; in gomp_loop_init() 65 ws->mode = ws->end < (LONG_MAX in gomp_loop_init() 66 - (nthreads + 1) * ws->chunk_size); in gomp_loop_init() 72 ws->mode = 0; in gomp_loop_init() [all …]
|
H A D | loop_ull.c | 40 ws->sched = sched; in gomp_loop_ull_init() 41 ws->chunk_size_ull = chunk_size; in gomp_loop_ull_init() 45 ws->incr_ull = incr; in gomp_loop_ull_init() 46 ws->next_ull = start; in gomp_loop_ull_init() 47 ws->mode = 0; in gomp_loop_ull_init() 50 ws->chunk_size_ull *= incr; in gomp_loop_ull_init() 63 if (__builtin_expect ((nthreads | ws->chunk_size_ull) in gomp_loop_ull_init() 66 ws->mode = ws->end_ull < (__LONG_LONG_MAX__ * 2ULL + 1 in gomp_loop_ull_init() 67 - (nthreads + 1) * ws->chunk_size_ull); in gomp_loop_ull_init() 73 ws->mode = ws->end_ull > ((nthreads + 1) * -ws->chunk_size_ull in gomp_loop_ull_init() [all …]
|
/dragonfly/contrib/gcc-8.0/libgomp/ |
H A D | work.c | 49 return ws; in alloc_work_share() 58 if (ws && ws->next_free) in alloc_work_share() 68 if (ws) in alloc_work_share() 73 return ws; in alloc_work_share() 84 ws[i].next_free = &ws[i + 1]; in alloc_work_share() 86 return ws; in alloc_work_share() 108 ws->ordered_team_ids = ws->inline_ordered_team_ids; in gomp_init_work_share() 128 if (ws->ordered_team_ids != ws->inline_ordered_team_ids) in gomp_fini_work_share() 141 free (ws); in free_work_share() 178 ws = gomp_malloc (sizeof (*ws)); in gomp_work_share_start() [all …]
|
H A D | iter_ull.c | 56 return ws->next_ull == ws->end_ull; in gomp_iter_ull_static_next() 71 n = (ws->end_ull - ws->next_ull + ws->incr_ull - 1) / ws->incr_ull; in gomp_iter_ull_static_next() 73 n = (ws->next_ull - ws->end_ull - ws->incr_ull - 1) / -ws->incr_ull; in gomp_iter_ull_static_next() 96 s = s0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 97 e = e0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 112 n = (ws->end_ull - ws->next_ull + ws->incr_ull - 1) / ws->incr_ull; in gomp_iter_ull_static_next() 114 n = (ws->next_ull - ws->end_ull - ws->incr_ull - 1) / -ws->incr_ull; in gomp_iter_ull_static_next() 130 s = s0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 131 e = e0 * ws->incr_ull + ws->next_ull; in gomp_iter_ull_static_next() 271 if (ws->next_ull == ws->end_ull) in gomp_iter_ull_guided_next_locked() [all …]
|
H A D | iter.c | 55 return ws->next == ws->end; in gomp_iter_static_next() 71 s = ws->incr + (ws->incr > 0 ? -1 : 1); in gomp_iter_static_next() 72 n = (ws->end - ws->next + s) / ws->incr; in gomp_iter_static_next() 95 s = (long)s0 * ws->incr + ws->next; in gomp_iter_static_next() 96 e = (long)e0 * ws->incr + ws->next; in gomp_iter_static_next() 111 s = ws->incr + (ws->incr > 0 ? -1 : 1); in gomp_iter_static_next() 112 n = (ws->end - ws->next + s) / ws->incr; in gomp_iter_static_next() 128 s = (long)s0 * ws->incr + ws->next; in gomp_iter_static_next() 129 e = (long)e0 * ws->incr + ws->next; in gomp_iter_static_next() 270 if (ws->next == ws->end) in gomp_iter_guided_next_locked() [all …]
|
H A D | ordered.c | 50 index = ws->ordered_cur + ws->ordered_num_used; in gomp_ordered_first() 82 ws->ordered_owner = -1; in gomp_ordered_last() 90 ws->ordered_cur = next; in gomp_ordered_last() 117 ws->ordered_owner = -1; in gomp_ordered_next() 133 index = ws->ordered_cur + ws->ordered_num_used; in gomp_ordered_next() 139 index = ws->ordered_cur + 1; in gomp_ordered_next() 142 ws->ordered_cur = index; in gomp_ordered_next() 181 ws->ordered_owner = -1; in gomp_ordered_static_next() 298 if (ws->sched == GFS_STATIC) in gomp_doacross_init() 349 ws->doacross = doacross; in gomp_doacross_init() [all …]
|
H A D | sections.c | 34 gomp_sections_init (struct gomp_work_share *ws, unsigned count) in gomp_sections_init() argument 36 ws->sched = GFS_DYNAMIC; in gomp_sections_init() 37 ws->chunk_size = 1; in gomp_sections_init() 38 ws->end = count + 1L; in gomp_sections_init() 39 ws->incr = 1; in gomp_sections_init() 40 ws->next = 1; in gomp_sections_init() 44 ws->mode = 1; in gomp_sections_init() 51 ws->mode = ((nthreads | ws->end) in gomp_sections_init() 55 ws->mode = 0; in gomp_sections_init()
|
/dragonfly/sbin/routed/ |
H A D | output.c | 59 } ws; variable 317 if (ws.a != NULL && ws.a->type == RIP_AUTH_MD5) in supply_write() 319 if (output(wb->type, &ws.to, ws.ifp, wb->buf, in supply_write() 351 (ws.state & WS_ST_TO_ON_NET) ? ws.ifp : 0); in supply_out() 527 || on_net(dst, ws.to_net, ws.to_mask)) in walk_supply() 561 && !on_net(dst, ws.to_std_net, ws.to_std_mask)) in walk_supply() 672 ws.to_std_mask = std_mask(ws.to.sin_addr.s_addr); in supply() 673 ws.to_std_net = ntohl(ws.to.sin_addr.s_addr) & ws.to_std_mask; in supply() 678 if (on_net(ws.to.sin_addr.s_addr, ws.to_net, ws.to_mask)) in supply() 682 ws.to_mask = ripv1_mask_net(ws.to.sin_addr.s_addr, 0); in supply() [all …]
|
/dragonfly/contrib/gmp/mpn/generic/ |
H A D | toom_eval_pm2rexp.c | 40 __cy = mpn_lshift(ws,src,n,s); in DO_mpn_addlsh_n() 41 return __cy + mpn_add_n(dst,dst,ws,n); in DO_mpn_addlsh_n() 50 unsigned int s, mp_ptr ws) in mpn_toom_eval_pm2rexp() argument 60 ws[n] = mpn_lshift(ws, ap+n, n, s*(q-1)); in mpn_toom_eval_pm2rexp() 62 ASSERT_NOCARRY(mpn_add(ws,ws,n+1,ap+n*q,t)); in mpn_toom_eval_pm2rexp() 71 ws[n] += DO_mpn_addlsh_n(ws, ap+n*i, n, s*(q-i), rm); in mpn_toom_eval_pm2rexp() 74 neg = (mpn_cmp (rp, ws, n + 1) < 0) ? ~0 : 0; in mpn_toom_eval_pm2rexp() 78 mpn_add_n_sub_n (rp, rm, ws, rp, n + 1); in mpn_toom_eval_pm2rexp() 80 mpn_add_n_sub_n (rp, rm, rp, ws, n + 1); in mpn_toom_eval_pm2rexp() 83 mpn_sub_n (rm, ws, rp, n + 1); in mpn_toom_eval_pm2rexp() [all …]
|
H A D | sqr.c | 44 mpn_toom2_sqr (p, a, n, ws); in mpn_sqr() 48 mp_ptr ws; in mpn_sqr() local 51 ws = TMP_SALLOC_LIMBS (mpn_toom3_sqr_itch (n)); in mpn_sqr() 52 mpn_toom3_sqr (p, a, n, ws); in mpn_sqr() 57 mp_ptr ws; in mpn_sqr() local 61 mpn_toom4_sqr (p, a, n, ws); in mpn_sqr() 66 mp_ptr ws; in mpn_sqr() local 70 mpn_toom6_sqr (p, a, n, ws); in mpn_sqr() 75 mp_ptr ws; in mpn_sqr() local 78 ws = TMP_ALLOC_LIMBS (mpn_toom8_sqr_itch (n)); in mpn_sqr() [all …]
|
H A D | mul_n.c | 42 mpn_toom22_mul (p, a, n, b, n, ws); in mpn_mul_n() 46 mp_ptr ws; in mpn_mul_n() local 50 mpn_toom33_mul (p, a, n, b, n, ws); in mpn_mul_n() 55 mp_ptr ws; in mpn_mul_n() local 59 mpn_toom44_mul (p, a, n, b, n, ws); in mpn_mul_n() 64 mp_ptr ws; in mpn_mul_n() local 67 ws = TMP_SALLOC_LIMBS (mpn_toom6_mul_n_itch (n)); in mpn_mul_n() 68 mpn_toom6h_mul (p, a, n, b, n, ws); in mpn_mul_n() 73 mp_ptr ws; in mpn_mul_n() local 76 ws = TMP_ALLOC_LIMBS (mpn_toom8_mul_n_itch (n)); in mpn_mul_n() [all …]
|
H A D | mul.c | 177 mp_ptr ws; in mpn_mul() local 180 ws = TMP_SALLOC_LIMBS (4 * vn); in mpn_mul() 208 MPN_COPY (prodp + vn, ws + vn, un); in mpn_mul() 240 mp_ptr ws; in mpn_mul() local 270 mpn_mul (ws, vp, vn, up, un); in mpn_mul() 272 mpn_mul (ws, up, un, vp, vn); in mpn_mul() 344 mp_ptr ws; in mpn_mul() local 348 ws = TMP_BALLOC_LIMBS (9 * vn >> 1); in mpn_mul() 369 mpn_mul (ws, vp, vn, up, un); in mpn_mul() 371 mpn_mul (ws, up, un, vp, vn); in mpn_mul() [all …]
|
H A D | toom63_mul.c | 77 #define TOOM_63_MUL_N_REC(p, a, b, n, ws) \ argument 147 cy = mpn_add (ws, b0, n, b2, t); in mpn_toom63_mul() 149 if (cy == 0 && mpn_cmp (ws, b1, n) < 0) in mpn_toom63_mul() 164 v3[n] = cy + mpn_add_n (v3, ws, b1, n); in mpn_toom63_mul() 165 if (cy == 0 && mpn_cmp (ws, b1, n) < 0) in mpn_toom63_mul() 167 mpn_sub_n (v1, b1, ws, n); in mpn_toom63_mul() 173 cy -= mpn_sub_n (v1, ws, b1, n); in mpn_toom63_mul() 196 TOOM_63_MUL_N_REC(pp, ap, bp, n, ws); in mpn_toom63_mul() 200 TOOM_63_MUL_REC(r1, a5, s, b2, t, ws); in mpn_toom63_mul() 202 TOOM_63_MUL_REC(r1, b2, t, a5, s, ws); in mpn_toom63_mul() [all …]
|
H A D | toom_interpolate_8pts.c | 58 #define DO_mpn_sublsh_n(dst,src,n,s,ws) mpn_sublsh_n (dst,src,n,s) argument 67 __cy = mpn_lshift (ws,src,n,s); in DO_mpn_sublsh_n() 68 return __cy + mpn_sub_n (dst,dst,ws,n); in DO_mpn_sublsh_n() 75 #define DO_mpn_subrsh(dst,nd,src,ns,s,ws) mpn_subrsh (dst,nd,src,ns,s) argument 78 #define DO_mpn_subrsh(dst,nd,src,ns,s,ws) \ argument 117 mp_size_t spt, mp_ptr ws) in mpn_toom_interpolate_8pts() argument 126 DO_mpn_subrsh(r3+n, 2 * n + 1, pp, 2 * n, 4, ws); in mpn_toom_interpolate_8pts() 127 cy = DO_mpn_sublsh_n (r3, r1, spt, 12, ws); in mpn_toom_interpolate_8pts() 130 DO_mpn_subrsh(r5+n, 2 * n + 1, pp, 2 * n, 2, ws); in mpn_toom_interpolate_8pts() 131 cy = DO_mpn_sublsh_n (r5, r1, spt, 6, ws); in mpn_toom_interpolate_8pts() [all …]
|
/dragonfly/lib/libc/net/ |
H A D | nscache.c | 331 cached_mp_write_session ws; in __nss_mp_cache_write() local 344 ws = cache_info->get_mp_ws_func(); in __nss_mp_cache_write() 354 cache_info->set_mp_ws_func(ws); in __nss_mp_cache_write() 395 cached_mp_write_session ws; in __nss_mp_cache_write_submit() local 403 ws = cache_info->get_mp_ws_func(); in __nss_mp_cache_write_submit() 405 __close_cached_mp_write_session(ws); in __nss_mp_cache_write_submit() 406 ws = INVALID_CACHED_MP_WRITE_SESSION; in __nss_mp_cache_write_submit() 407 cache_info->set_mp_ws_func(ws); in __nss_mp_cache_write_submit() 415 cached_mp_write_session ws; in __nss_mp_cache_end() local 424 ws = cache_info->get_mp_ws_func(); in __nss_mp_cache_end() [all …]
|
/dragonfly/usr.sbin/nscd/ |
H A D | cachelib.c | 189 assert(ws != NULL); in destroy_cache_mp_write_session() 197 free(ws); in destroy_cache_mp_write_session() 938 assert(ws != NULL); in cache_mp_write() 943 (ws->parent_entry->mp_params.max_elemsize == ws->items_size)) { in cache_mp_write() 958 ++ws->items_size; in cache_mp_write() 972 assert(ws != NULL); in abandon_cache_mp_write_session() 976 TAILQ_REMOVE(&ws->parent_entry->ws_head, ws, entries); in abandon_cache_mp_write_session() 991 assert(ws != NULL); in close_cache_mp_write_session() 995 TAILQ_REMOVE(&ws->parent_entry->ws_head, ws, entries); in close_cache_mp_write_session() 1004 ws->parent_entry->completed_write_session = ws; in close_cache_mp_write_session() [all …]
|
/dragonfly/contrib/gdb-7/gdb/ |
H A D | fork-child.c | 436 struct target_waitstatus ws; in startup_inferior() local 437 memset (&ws, 0, sizeof (ws)); in startup_inferior() 438 event_ptid = target_wait (resume_ptid, &ws, 0); in startup_inferior() 440 if (ws.kind == TARGET_WAITKIND_IGNORE) in startup_inferior() 444 switch (ws.kind) in startup_inferior() 460 gdb_signal_to_name (ws.value.sig), in startup_inferior() 461 gdb_signal_to_string (ws.value.sig)); in startup_inferior() 467 if (ws.value.integer) in startup_inferior() 469 ws.value.integer); in startup_inferior() 476 xfree (ws.value.execd_pathname); in startup_inferior() [all …]
|
/dragonfly/libexec/telnetd/ |
H A D | termstat.c | 514 struct winsize ws; in clientstat() local 531 ws.ws_col = parm1; in clientstat() 532 ws.ws_row = parm2; in clientstat() 533 (void) ioctl(pty, TIOCSWINSZ, (char *)&ws); in clientstat() 596 struct winsize ws; in defer_terminit() local 598 memset((char *)&ws, 0, sizeof(ws)); in defer_terminit() 599 ws.ws_col = def_col; in defer_terminit() 600 ws.ws_row = def_row; in defer_terminit() 601 (void) ioctl(pty, TIOCSWINSZ, (char *)&ws); in defer_terminit()
|
/dragonfly/usr.bin/mail/ |
H A D | main.c | 319 struct winsize ws; in setscreensize() local 322 if (ioctl(1, TIOCGWINSZ, (char *)&ws) < 0) in setscreensize() 323 ws.ws_col = ws.ws_row = 0; in setscreensize() 332 else if (ws.ws_row != 0) in setscreensize() 333 screenheight = ws.ws_row; in setscreensize() 336 if ((realscreenheight = ws.ws_row) == 0) in setscreensize() 338 if ((screenwidth = ws.ws_col) == 0) in setscreensize()
|