/dports/science/hdf5-18/hdf5-1.8.21/fortran/testpar/ |
H A D | ptest.f90 | 41 CALL mpi_comm_rank( MPI_COMM_WORLD, mpi_rank, mpierror ) 43 WRITE(*,*) "MPI_COMM_RANK *FAILED* Process = ", mpi_rank 47 WRITE(*,*) "MPI_COMM_SIZE *FAILED* Process = ", mpi_rank 63 CALL hyper(length, do_collective, do_chunk, mpi_size, mpi_rank, nerrors) 73 CALL hyper(length, do_collective, do_chunk, mpi_size, mpi_rank, nerrors) 83 CALL hyper(length, do_collective, do_chunk, mpi_size, mpi_rank, nerrors) 93 CALL hyper(length, do_collective, do_chunk, mpi_size, mpi_rank, nerrors) 103 CALL multiple_dset_write(length, do_collective, do_chunk, mpi_size, mpi_rank, nerrors) 119 WRITE(*,*) "MPI_FINALIZE *FAILED* Process = ", mpi_rank 122 WRITE(*,*) 'Errors detected in process ', mpi_rank [all …]
|
/dports/science/hdf5-18/hdf5-1.8.21/testpar/ |
H A D | t_span_tree.c | 262 int mpi_size,mpi_rank; in coll_write_test() local 269 MPI_Comm_rank(comm,&mpi_rank); in coll_write_test() 724 int mpi_size,mpi_rank; in coll_read_test() local 731 MPI_Comm_rank(comm,&mpi_rank); in coll_read_test() 1476 if ( mpi_rank == in lower_dim_size_comp_test__verify_data() 1523 if ( mpi_rank == in lower_dim_size_comp_test__verify_data() 1587 int mpi_rank; in lower_dim_size_comp_test__run_test() local 1656 fcnName, mpi_rank, in lower_dim_size_comp_test__run_test() 2104 fcnName, mpi_rank, in lower_dim_size_comp_test__run_test() 2108 fcnName, mpi_rank, in lower_dim_size_comp_test__run_test() [all …]
|
H A D | t_mpi.c | 46 int mpi_size, mpi_rank; in test_mpio_overlap_writes() local 89 mpi_off = mpi_rank*stride; in test_mpio_overlap_writes() 192 int mpi_size, mpi_rank; in test_mpio_gb_file() local 314 mpi_rank, MB, mpi_off); in test_mpio_gb_file() 337 mpi_rank); in test_mpio_gb_file() 530 if (mpi_rank==irank){ in test_mpio_1wMr() 691 int mpi_rank,mpi_size; in test_mpio_derived_dtype() local 828 if(mpi_rank == 0) { in test_mpio_derived_dtype() 886 if(mpi_rank == 0 || mpi_rank == 1) { in test_mpio_special_collective() 1170 if (mpi_rank==0) in main() [all …]
|
H A D | t_mdset.c | 48 int mpi_rank; in get_size() local 350 int mpi_size, mpi_rank; in null_dataset() local 678 req_start[0]=mpi_rank; in dataset_fillvalue() 876 int mpi_rank, m; in independent_group_read() local 895 if(mpi_rank%2==0) { in independent_group_read() 1130 int mpi_rank; in create_group_recursive() local 1596 if(mpi_rank == 0 ) { in io_mode_confusion() 1629 if(mpi_rank == 0 ) { in io_mode_confusion() 1811 int mpi_rank; in rr_obj_hdr_flush_confusion() local 1887 int mpi_rank; in rr_obj_hdr_flush_confusion_writer() local [all …]
|
H A D | t_shapesame.c | 170 mpi_rank = tv_ptr->mpi_rank; in hs_dr_pio_test__setup() 709 mpi_rank = tv_ptr->mpi_rank; in hs_dr_pio_test__takedown() 833 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__d2m_l2s() 1076 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__d2m_s2l() 1361 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__m2d_l2s() 1691 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__m2d_s2l() 2098 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__run_test() 2998 mpi_rank = tv_ptr->mpi_rank; 3276 mpi_rank = tv_ptr->mpi_rank; 3609 mpi_rank = tv_ptr->mpi_rank; [all …]
|
/dports/science/hdf5/hdf5-1.10.6/testpar/ |
H A D | t_span_tree.c | 240 int mpi_size,mpi_rank; in coll_write_test() local 247 MPI_Comm_rank(comm,&mpi_rank); in coll_write_test() 697 int mpi_size,mpi_rank; in coll_read_test() local 704 MPI_Comm_rank(comm,&mpi_rank); in coll_read_test() 1449 if ( mpi_rank == in lower_dim_size_comp_test__verify_data() 1496 if ( mpi_rank == in lower_dim_size_comp_test__verify_data() 1558 int mpi_rank; in lower_dim_size_comp_test__run_test() local 1627 fcnName, mpi_rank, in lower_dim_size_comp_test__run_test() 2075 fcnName, mpi_rank, in lower_dim_size_comp_test__run_test() 2079 fcnName, mpi_rank, in lower_dim_size_comp_test__run_test() [all …]
|
H A D | t_mdset.c | 49 int mpi_rank; in get_size() local 722 req_start[0]=mpi_rank; in dataset_fillvalue() 949 int mpi_rank, m; in independent_group_read() local 970 if(mpi_rank%2==0) { in independent_group_read() 1205 int mpi_rank; in create_group_recursive() local 1520 int mpi_rank, mpi_size; in get_slab() local 1670 if(mpi_rank == 0 ) { in io_mode_confusion() 1703 if(mpi_rank == 0 ) { in io_mode_confusion() 1885 int mpi_rank; in rr_obj_hdr_flush_confusion() local 1961 int mpi_rank; in rr_obj_hdr_flush_confusion_writer() local [all …]
|
H A D | t_mpi.c | 42 int mpi_size, mpi_rank; in test_mpio_overlap_writes() local 184 int mpi_size, mpi_rank; in test_mpio_gb_file() local 438 int mpi_size, mpi_rank; in test_mpio_1wMr() local 533 if (mpi_rank == irank) { in test_mpio_1wMr() 692 int mpi_rank, mpi_size; in test_mpio_derived_dtype() local 826 if (mpi_rank == 0) { in test_mpio_derived_dtype() 862 int mpi_size, mpi_rank; in test_mpio_special_collective() local 883 if (mpi_rank == 0 || mpi_rank == 1) { in test_mpio_special_collective() 969 if (mpi_rank == 0) { in test_mpio_special_collective() 1142 if (mpi_rank==0) in main() [all …]
|
H A D | t_shapesame.c | 164 mpi_rank = tv_ptr->mpi_rank; in hs_dr_pio_test__setup() 699 mpi_rank = tv_ptr->mpi_rank; in hs_dr_pio_test__takedown() 819 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__d2m_l2s() 1058 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__d2m_s2l() 1339 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__m2d_l2s() 1665 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__m2d_s2l() 2072 mpi_rank = tv_ptr->mpi_rank; in contig_hs_dr_pio_test__run_test() 2968 mpi_rank = tv_ptr->mpi_rank; 3242 mpi_rank = tv_ptr->mpi_rank; 3571 mpi_rank = tv_ptr->mpi_rank; [all …]
|
H A D | t_cache_image.c | 91 int mpi_rank, int mpi_size); 3133 if ( mpi_rank == 0 ) { in verify_cache_image_RO() 3333 if ( mpi_rank == 0 ) { in verify_cache_image_RO() 3412 if ( mpi_rank == 0 ) { in verify_cache_image_RW() 3631 if ( mpi_rank == 0 ) { in verify_cache_image_RW() 3693 if ( mpi_rank == 0 ) { in smoke_check_1() 4056 if ( mpi_rank == 0 ) { in smoke_check_1() 4120 int mpi_rank; in main() local 4182 mpi_rank); in main() 4187 if ( mpi_rank == 0 ) { in main() [all …]
|
/dports/science/netcdf/netcdf-c-4.7.4/nc_test4/ |
H A D | tst_parallel3.c | 65 if (mpi_rank == 0) in main() 89 if (mpi_rank == 0) in main() 93 if (mpi_rank == 0) in main() 96 if (mpi_rank == 0) in main() 100 if (mpi_rank == 0) in main() 103 if (mpi_rank == 0) in main() 107 if (mpi_rank == 0) in main() 110 if (mpi_rank == 0) in main() 113 if (mpi_rank == 0) in main() 116 if (mpi_rank == 0) in main() [all …]
|
H A D | tst_parallel5.c | 43 if (!mpi_rank) in main() 46 if (!mpi_rank) in main() 86 if (!mpi_rank) in main() 89 if (!mpi_rank) in main() 147 if (!mpi_rank) in main() 149 if (!mpi_rank) in main() 221 if (!mpi_rank) in main() 223 if (!mpi_rank) in main() 286 if (!mpi_rank) in main() 289 if (!mpi_rank) in main() [all …]
|
H A D | tst_simplerw_coll_r.c | 31 int mpi_size, mpi_rank; in main() local 51 if (!mpi_rank) in main() 57 if (!mpi_rank) in main() 130 if (!mpi_rank) in main() 139 byte_data[i] = mpi_rank; in main() 148 char_data[i] = mpi_rank; in main() 166 int_data[i] = mpi_rank; in main() 280 if (!mpi_rank) in main() 294 if (!mpi_rank) in main() 405 if (!mpi_rank) in main() [all …]
|
H A D | tst_nc4perf.c | 32 MPI_Info info, int mpi_size, int mpi_rank, in test_pio_2d() argument 85 start[1] = mpi_rank * DIMSIZE1/mpi_size; in test_pio_2d() 103 if (!mpi_rank) in test_pio_2d() 111 if (!mpi_rank) in test_pio_2d() 215 if (!mpi_rank) in test_pio_4d() 223 if (!mpi_rank) in test_pio_4d() 258 int mpi_size, mpi_rank; in main() local 273 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in main() 283 if (!mpi_rank) in main() 302 if (!mpi_rank) in main() [all …]
|
H A D | tst_parallel4.c | 37 int mpi_size, mpi_rank; in main() local 59 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in main() 66 if (!mpi_rank) printf("NUM_SLABS (%d) is not evenly divisible by mpi_size(%d)\n", in main() 101 data[i] = mpi_rank; in main() 129 if (mpi_rank) in main() 130 sleep(mpi_rank); in main() 136 if (!mpi_rank) in main() 142 start[0] = NUM_SLABS / mpi_size * mpi_rank + i; in main() 156 if (!mpi_rank) in main() 187 start[0] = NUM_SLABS / mpi_size * mpi_rank + i; in main() [all …]
|
/dports/math/libmesh/libmesh-1.6.2/contrib/netcdf/netcdf-c-4.6.2/nc_test4/ |
H A D | tst_parallel3.c | 65 if (mpi_rank == 0) in main() 89 if (mpi_rank == 0) in main() 93 if (mpi_rank == 0) in main() 96 if (mpi_rank == 0) in main() 100 if (mpi_rank == 0) in main() 103 if (mpi_rank == 0) in main() 107 if (mpi_rank == 0) in main() 110 if (mpi_rank == 0) in main() 113 if (mpi_rank == 0) in main() 116 if (mpi_rank == 0) in main() [all …]
|
H A D | tst_parallel5.c | 44 if (!mpi_rank) in main() 47 if (!mpi_rank) in main() 66 if (!mpi_rank) in main() 87 if (!mpi_rank) in main() 90 if (!mpi_rank) in main() 148 if (!mpi_rank) in main() 150 if (!mpi_rank) in main() 222 if (!mpi_rank) in main() 224 if (!mpi_rank) in main() 287 if (!mpi_rank) in main() [all …]
|
H A D | tst_simplerw_coll_r.c | 31 int mpi_size, mpi_rank; in main() local 51 if (!mpi_rank) in main() 57 if (!mpi_rank) in main() 130 if (!mpi_rank) in main() 139 byte_data[i] = mpi_rank; in main() 148 char_data[i] = mpi_rank; in main() 166 int_data[i] = mpi_rank; in main() 280 if (!mpi_rank) in main() 294 if (!mpi_rank) in main() 405 if (!mpi_rank) in main() [all …]
|
H A D | tst_nc4perf.c | 32 MPI_Info info, int mpi_size, int mpi_rank, in test_pio_2d() argument 85 start[1] = mpi_rank * DIMSIZE1/mpi_size; in test_pio_2d() 103 if (!mpi_rank) in test_pio_2d() 111 if (!mpi_rank) in test_pio_2d() 215 if (!mpi_rank) in test_pio_4d() 223 if (!mpi_rank) in test_pio_4d() 258 int mpi_size, mpi_rank; in main() local 273 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in main() 283 if (!mpi_rank) in main() 302 if (!mpi_rank) in main() [all …]
|
H A D | tst_parallel4.c | 37 int mpi_size, mpi_rank; in main() local 59 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in main() 66 if (!mpi_rank) printf("NUM_SLABS (%d) is not evenly divisible by mpi_size(%d)\n", in main() 101 data[i] = mpi_rank; in main() 129 if (mpi_rank) in main() 130 sleep(mpi_rank); in main() 136 if (!mpi_rank) in main() 142 start[0] = NUM_SLABS / mpi_size * mpi_rank + i; in main() 156 if (!mpi_rank) in main() 187 start[0] = NUM_SLABS / mpi_size * mpi_rank + i; in main() [all …]
|
/dports/science/hdf5/hdf5-1.10.6/fortran/testpar/ |
H A D | ptest.f90 | 30 INTEGER :: mpi_rank ! rank of the calling process in the communicator variable 47 CALL mpi_comm_rank( MPI_COMM_WORLD, mpi_rank, mpierror ) 49 WRITE(*,*) "MPI_COMM_RANK *FAILED* Process = ", mpi_rank 53 WRITE(*,*) "MPI_COMM_SIZE *FAILED* Process = ", mpi_rank 65 CALL hyper(length, do_collective(j), do_chunk(i), mpi_size, mpi_rank, ret_total_error) 66 IF(mpi_rank==0) CALL write_test_status(ret_total_error, & 76 …CALL multiple_dset_write(length, do_collective(1), do_chunk(1), mpi_size, mpi_rank, ret_total_erro… 77 IF(mpi_rank==0) CALL write_test_status(ret_total_error, & 91 WRITE(*,*) "MPI_FINALIZE *FAILED* Process = ", mpi_rank 94 WRITE(*,*) 'Errors detected in process ', mpi_rank [all …]
|
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/examples/proxies/kripke/Kripke/ |
H A D | ParallelComm.cpp | 19 int ParallelComm::computeTag(int mpi_rank, int sdom_id){ in computeTag() argument 23 int tag = mpi_rank + mpi_size*sdom_id; in computeTag() 32 mpi_rank = tag % mpi_size; in computeRankSdom() 77 int mpi_rank, mpi_size; in postRecvs() local 78 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in postRecvs() 85 if(sdom.upwind[dim].mpi_rank < 0){ in postRecvs() 90 if(sdom.upwind[dim].mpi_rank == mpi_rank){ in postRecvs() 119 int mpi_rank, mpi_size; in postSends() local 120 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in postSends() 124 if(sdom->downwind[dim].mpi_rank < 0){ in postSends() [all …]
|
/dports/math/libmesh/libmesh-1.6.2/contrib/netcdf/netcdf-c-4.6.2/nc_test/ |
H A D | tst_parallel2.c | 35 int mpi_size, mpi_rank; in main() local 57 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in main() 93 if (!mpi_rank) in main() 103 data[i] = mpi_rank; in main() 134 if (mpi_rank) in main() 135 sleep(mpi_rank); in main() 146 if (!mpi_rank) in main() 166 if (!mpi_rank) in main() 207 if (data_in[j] != mpi_rank) in main() 230 if (mpi_rank == 0) in main() [all …]
|
/dports/science/netcdf/netcdf-c-4.7.4/nc_test/ |
H A D | tst_parallel2.c | 46 int mpi_size, mpi_rank; in main() local 68 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in main() 104 if (!mpi_rank) in main() 114 data[i] = mpi_rank; in main() 145 if (mpi_rank) in main() 146 sleep(mpi_rank); in main() 157 if (!mpi_rank) in main() 177 if (!mpi_rank) in main() 218 if (data_in[j] != mpi_rank) in main() 241 if (mpi_rank == 0) in main() [all …]
|
/dports/science/pynn/PyNN-0.10.0/test/system/scenarios/ |
H A D | test_parameter_handling.py | 58 if mpi_rank == 0: 66 if mpi_rank == 0: 73 if mpi_rank == 0: 80 if mpi_rank == 0: 95 if mpi_rank == 0: 118 if mpi_rank == 0: 126 if mpi_rank == 0: 133 if mpi_rank == 0: 140 if mpi_rank == 0: 152 if mpi_rank == 0: [all …]
|