/dragonfly/contrib/gcc-8.0/libgomp/ |
H A D | team.c | 148 if (last_team != NULL && last_team->nthreads == nthreads) in get_last_team() 178 team->nthreads = nthreads; in gomp_new_team() 332 if (nthreads > 1) in gomp_team_start() 356 if (nthreads == 1) in gomp_team_start() 439 n = nthreads; in gomp_team_start() 525 if (i == nthreads - rest) in gomp_team_start() 577 n = nthreads; in gomp_team_start() 689 if (i == nthreads) in gomp_team_start() 722 * (nthreads-i)); in gomp_team_start() 725 for (; i < nthreads; ++i) in gomp_team_start() [all …]
|
H A D | iter.c | 44 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_static_next() local 50 if (nthreads == 1) in gomp_iter_static_next() 77 q = n / nthreads; in gomp_iter_static_next() 78 t = n % nthreads; in gomp_iter_static_next() 118 s0 = (thr->ts.static_trip * nthreads + i) * c; in gomp_iter_static_next() 266 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_guided_next_locked() local 275 q = (n + nthreads - 1) / nthreads; in gomp_iter_guided_next_locked() 300 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_guided_next() local 318 q = (n + nthreads - 1) / nthreads; in gomp_iter_guided_next()
|
H A D | iter_ull.c | 45 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_ull_static_next() local 51 if (nthreads == 1) in gomp_iter_ull_static_next() 78 q = n / nthreads; in gomp_iter_ull_static_next() 79 t = n % nthreads; in gomp_iter_ull_static_next() 120 s0 = (thr->ts.static_trip * (gomp_ull) nthreads + i) * c; in gomp_iter_ull_static_next() 267 gomp_ull nthreads = team ? team->nthreads : 1; in gomp_iter_ull_guided_next_locked() local 279 q = (n + nthreads - 1) / nthreads; in gomp_iter_ull_guided_next_locked() 304 gomp_ull nthreads = team ? team->nthreads : 1; in gomp_iter_ull_guided_next() local 325 q = (n + nthreads - 1) / nthreads; in gomp_iter_ull_guided_next()
|
H A D | ordered.c | 47 if (team == NULL || team->nthreads == 1) in gomp_ordered_first() 51 if (index >= team->nthreads) in gomp_ordered_first() 52 index -= team->nthreads; in gomp_ordered_first() 78 if (team == NULL || team->nthreads == 1) in gomp_ordered_last() 88 if (next == team->nthreads) in gomp_ordered_last() 134 if (index >= team->nthreads) in gomp_ordered_next() 135 index -= team->nthreads; in gomp_ordered_next() 140 if (index == team->nthreads) in gomp_ordered_next() 184 if (++id == team->nthreads) in gomp_ordered_static_next() 299 num_ents = team->nthreads; in gomp_doacross_init() [all …]
|
H A D | parallel.c | 137 unsigned int nthreads = team ? team->nthreads : 1; in GOMP_parallel_end() local 139 if (nthreads > 1) in GOMP_parallel_end() 149 1UL - nthreads); in GOMP_parallel_end() 152 thr->thread_pool->threads_busy -= nthreads - 1; in GOMP_parallel_end() 240 return team ? team->nthreads : 1; in omp_get_num_threads() 287 return ts->team->nthreads; in omp_get_team_size()
|
H A D | work.c | 94 unsigned nthreads) in gomp_init_work_share() argument 104 if (nthreads > INLINE_ORDERED_TEAM_IDS_CNT) in gomp_init_work_share() 106 = gomp_malloc (nthreads * sizeof (*ws->ordered_team_ids)); in gomp_init_work_share() 110 nthreads * sizeof (*ws->ordered_team_ids)); in gomp_init_work_share() 191 gomp_init_work_share (ws, ordered, team->nthreads); in gomp_work_share_start() 291 if (completed == team->nthreads) in gomp_work_share_end_nowait()
|
H A D | sections.c | 49 long nthreads = team ? team->nthreads : 1; in gomp_sections_init() local 51 ws->mode = ((nthreads | ws->end) in gomp_sections_init()
|
H A D | taskloop.c | 134 num_tasks = team ? team->nthreads : 1; in GOMP_taskloop() 163 || team->task_count + num_tasks > 64 * team->nthreads) in GOMP_taskloop() 325 < team->nthreads) in GOMP_taskloop() 327 do_wake = team->nthreads - team->task_running_count in GOMP_taskloop()
|
/dragonfly/contrib/gcc-4.7/libgomp/ |
H A D | ordered.c | 43 if (team == NULL || team->nthreads == 1) in gomp_ordered_first() 47 if (index >= team->nthreads) in gomp_ordered_first() 48 index -= team->nthreads; in gomp_ordered_first() 74 if (team == NULL || team->nthreads == 1) in gomp_ordered_last() 84 if (next == team->nthreads) in gomp_ordered_last() 109 if (team == NULL || team->nthreads == 1) in gomp_ordered_next() 130 if (index >= team->nthreads) in gomp_ordered_next() 131 index -= team->nthreads; in gomp_ordered_next() 136 if (index == team->nthreads) in gomp_ordered_next() 154 if (team == NULL || team->nthreads == 1) in gomp_ordered_static_init() [all …]
|
H A D | team.c | 137 gomp_new_team (unsigned nthreads) in gomp_new_team() argument 161 team->nthreads = nthreads; in gomp_new_team() 294 if (nthreads > 1) in gomp_team_start() 310 if (nthreads == 1) in gomp_team_start() 324 if (nthreads <= old_threads_used) in gomp_team_start() 325 n = nthreads; in gomp_team_start() 343 pool->threads_used = nthreads; in gomp_team_start() 367 if (i == nthreads) in gomp_team_start() 373 if (nthreads >= pool->threads_size) in gomp_team_start() 375 pool->threads_size = nthreads + 1; in gomp_team_start() [all …]
|
H A D | iter.c | 43 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_static_next() local 49 if (nthreads == 1) in gomp_iter_static_next() 76 q = n / nthreads; in gomp_iter_static_next() 77 t = n % nthreads; in gomp_iter_static_next() 117 s0 = (thr->ts.static_trip * nthreads + i) * c; in gomp_iter_static_next() 265 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_guided_next_locked() local 274 q = (n + nthreads - 1) / nthreads; in gomp_iter_guided_next_locked() 299 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_guided_next() local 317 q = (n + nthreads - 1) / nthreads; in gomp_iter_guided_next()
|
H A D | iter_ull.c | 44 unsigned long nthreads = team ? team->nthreads : 1; in gomp_iter_ull_static_next() local 50 if (nthreads == 1) in gomp_iter_ull_static_next() 77 q = n / nthreads; in gomp_iter_ull_static_next() 78 t = n % nthreads; in gomp_iter_ull_static_next() 119 s0 = (thr->ts.static_trip * (gomp_ull) nthreads + i) * c; in gomp_iter_ull_static_next() 266 gomp_ull nthreads = team ? team->nthreads : 1; in gomp_iter_ull_guided_next_locked() local 278 q = (n + nthreads - 1) / nthreads; in gomp_iter_ull_guided_next_locked() 303 gomp_ull nthreads = team ? team->nthreads : 1; in gomp_iter_ull_guided_next() local 324 q = (n + nthreads - 1) / nthreads; in gomp_iter_ull_guided_next()
|
H A D | parallel.c | 118 if (team && team->nthreads > 1) in GOMP_parallel_end() 122 1UL - team->nthreads); in GOMP_parallel_end() 125 gomp_remaining_threads_count -= team->nthreads - 1; in GOMP_parallel_end() 140 return team ? team->nthreads : 1; in omp_get_num_threads() 187 return ts->team->nthreads; in omp_get_team_size()
|
H A D | work.c | 93 unsigned nthreads) in gomp_init_work_share() argument 103 if (nthreads > INLINE_ORDERED_TEAM_IDS_CNT) in gomp_init_work_share() 105 = gomp_malloc (nthreads * sizeof (*ws->ordered_team_ids)); in gomp_init_work_share() 109 nthreads * sizeof (*ws->ordered_team_ids)); in gomp_init_work_share() 190 gomp_init_work_share (ws, ordered, team->nthreads); in gomp_work_share_start() 261 if (completed == team->nthreads) in gomp_work_share_end_nowait()
|
H A D | sections.c | 48 long nthreads = team ? team->nthreads : 1; in gomp_sections_init() local 50 ws->mode = ((nthreads | ws->end) in gomp_sections_init()
|
H A D | loop.c | 55 long nthreads = team ? team->nthreads : 1; in gomp_loop_init() local 60 if (__builtin_expect ((nthreads | ws->chunk_size) in gomp_loop_init() 66 - (nthreads + 1) * ws->chunk_size); in gomp_loop_init() 69 else if (__builtin_expect ((nthreads | -ws->chunk_size) in gomp_loop_init() 74 ws->mode = ws->end > (nthreads + 1) * -ws->chunk_size - LONG_MAX; in gomp_loop_init()
|
H A D | loop_ull.c | 58 long nthreads = team ? team->nthreads : 1; in gomp_loop_ull_init() local 63 if (__builtin_expect ((nthreads | ws->chunk_size_ull) in gomp_loop_ull_init() 67 - (nthreads + 1) * ws->chunk_size_ull); in gomp_loop_ull_init() 70 else if (__builtin_expect ((nthreads | -ws->chunk_size_ull) in gomp_loop_ull_init() 73 ws->mode = ws->end_ull > ((nthreads + 1) * -ws->chunk_size_ull in gomp_loop_ull_init()
|
/dragonfly/test/testcases/threads/umtx_errno/ |
H A D | umtx_errno.c | 47 int i, ret, nthreads; in main() local 57 nthreads = atoi(argv[1]); in main() 58 threads = malloc(sizeof(pthread_t) * (nthreads + 1)); in main() 64 printf("Trying with %d threads\n", nthreads); in main() 66 if ( (ret = pthread_barrier_init(&bar, NULL, nthreads + 1)) != 0) in main() 79 for (i = 0; i < nthreads; i++) { in main() 88 for (i = 0; i < nthreads; i++) { in main()
|
/dragonfly/test/testcases/mem/mmap_1/ |
H A D | mmap_1.c | 19 int i, ret, nthreads; in main() local 29 nthreads = atoi(argv[1]); in main() 30 threads = malloc(nthreads * sizeof(pthread_t)); in main() 34 printf("Trying with %d threads\n", nthreads); in main() 37 for (i = 0; i < nthreads; i++) in main() 43 for (i = 0; i < nthreads; i++) in main()
|
/dragonfly/usr.bin/sort/ |
H A D | radixsort.c | 143 if (nthreads > 1) in push_ls() 151 if (nthreads > 1) { in push_ls() 157 if (nthreads > 1) in push_ls() 607 if (nthreads < 2) { in run_top_sort_level() 614 for(i = 0; i < nthreads; ++i) { in run_top_sort_level() 636 for (i = 0; i < nthreads; ++i) in run_top_sort_level() 648 size_t nthreads_save = nthreads; in run_sort() 650 nthreads = 1; in run_sort() 652 if (nthreads > 1) { in run_sort() 684 if (nthreads > 1) { in run_sort() [all …]
|
H A D | file.c | 1513 if (nthreads < 2 || list->count < MT_SORT_THRESHOLD) { in mt_sort() 1514 size_t nthreads_save = nthreads; in mt_sort() 1515 nthreads = 1; in mt_sort() 1523 nthreads = nthreads_save; in mt_sort() 1532 avgsize = list->count / nthreads; in mt_sort() 1538 for (i = 0; i < nthreads; ++i) { in mt_sort() 1546 sz = (i == nthreads - 1) ? list->count - cstart : in mt_sort() 1560 for (i = 0; i < nthreads; ++i) { in mt_sort() 1584 for (i = 0; i < nthreads; ++i) { in mt_sort() 1591 merge_list_parts(parts, nthreads, fn); in mt_sort() [all …]
|
H A D | sort.c | 119 size_t nthreads = 1; variable 312 nthreads = ncpu; in set_hw_params() 1186 nthreads = (size_t)(atoi(optarg)); in main() 1187 if (nthreads < 1) in main() 1188 nthreads = 1; in main() 1189 if (nthreads > 1024) in main() 1190 nthreads = 1024; in main() 1297 nthreads = 1; in main() 1341 nthreads = 1; in main()
|
/dragonfly/sys/kern/ |
H A D | kern_checkpoint.c | 77 prstatus_t *status, prfpregset_t *fpregset, int nthreads); 179 int nthreads; in elf_getnotes() local 185 nthreads = (notesz - sizeof(prpsinfo_t))/(sizeof(prstatus_t) + in elf_getnotes() 187 PRINTF(("reading notes header nthreads=%d\n", nthreads)); in elf_getnotes() 188 if (nthreads <= 0 || nthreads > CKPT_MAXTHREADS) in elf_getnotes() 192 status = kmalloc(nthreads*sizeof(prstatus_t), M_TEMP, M_WAITOK); in elf_getnotes() 193 fpregset = kmalloc(nthreads*sizeof(prfpregset_t), M_TEMP, M_WAITOK); in elf_getnotes() 200 error = elf_demarshalnotes(note, psinfo, status, fpregset, nthreads); in elf_getnotes() 356 prfpregset_t *fpregset, int nthreads) in elf_demarshalnotes() argument 380 for (i = 0 ; i < nthreads - 1; i++) { in elf_demarshalnotes()
|
/dragonfly/contrib/gcc-8.0/libgomp/config/posix/ |
H A D | pool.h | 37 gomp_get_thread_pool (struct gomp_thread *thr, unsigned nthreads) in gomp_get_thread_pool() argument 47 pool->threads_busy = nthreads; in gomp_get_thread_pool()
|
/dragonfly/contrib/gdb-7/gdb/python/ |
H A D | py-inferior.c | 51 int nthreads; member 166 inf_obj->nthreads = 0; in inferior_to_inferior_object() 249 inf_obj->nthreads++; in add_thread_object() 287 inf_obj->nthreads--; in delete_thread_object() 311 tuple = PyTuple_New (inf_obj->nthreads); in infpy_threads() 315 for (i = 0, entry = inf_obj->threads; i < inf_obj->nthreads; in infpy_threads() 760 inf_obj->nthreads = 0; in py_free_inferior()
|