/dports/science/hdf5-18/hdf5-1.8.21/src/ |
H A D | H5FDmpi.c | 381 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor() 383 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor() 431 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor() 432 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
|
H A D | H5AC.c | 427 int mpi_rank; in H5AC_create() local 444 aux_ptr->mpi_rank = mpi_rank; in H5AC_create() 470 if(mpi_rank == 0) { in H5AC_create() 3128 if(aux_ptr->mpi_rank == 0) { in H5AC_log_deleted_entry() 3219 if ( aux_ptr->mpi_rank == 0 ) { in H5AC_log_dirtied_entry() 3461 if(aux_ptr->mpi_rank == 0) { in H5AC_log_inserted_entry() 3588 if ( aux_ptr->mpi_rank == 0 ) { in H5AC_log_moved_entry() 3828 if(aux_ptr->mpi_rank == 0) { in H5AC_propagate_and_apply_candidate_list() 3872 if(aux_ptr->mpi_rank == 0) { in H5AC_propagate_and_apply_candidate_list() 3986 if(aux_ptr->mpi_rank == 0) { in H5AC_propagate_flushed_and_still_clean_entries_list() [all …]
|
/dports/science/hdf5/hdf5-1.10.6/src/ |
H A D | H5FDmpi.c | 425 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor() 427 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor() 475 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor() 476 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
|
/dports/math/vtk6/VTK-6.2.0/ThirdParty/hdf5/vtkhdf5/src/ |
H A D | H5FDmpi.c | 383 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor() 385 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor() 433 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor() 434 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
|
/dports/math/vtk8/VTK-8.2.0/ThirdParty/hdf5/vtkhdf5/src/ |
H A D | H5FDmpi.c | 420 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor() 422 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor() 470 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor() 471 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
|
/dports/graphics/opencv/opencv-4.5.3/modules/flann/include/opencv2/flann/ |
H A D | hdf5.h | 176 int mpi_size, mpi_rank; in load_from_file() local 178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file() 203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file() 207 offset[0] = mpi_rank*item_cnt; in load_from_file()
|
/dports/science/hdf5-18/hdf5-1.8.21/test/ |
H A D | h5test.c | 544 int mpi_rank; in h5_fixname_real() local 546 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in h5_fixname_real() 548 if (mpi_rank == 0) in h5_fixname_real() 883 int mpi_rank, mpi_initialized, mpi_finalized; in h5_show_hostname() local 889 MPI_Comm_rank(MPI_COMM_WORLD,&mpi_rank); in h5_show_hostname() 890 printf("MPI-process %d.", mpi_rank); in h5_show_hostname() 1278 int mpi_size, mpi_rank, mpi_initialized, mpi_finalized; in getenv_all() local 1288 MPI_Comm_rank(comm, &mpi_rank); in getenv_all() 1294 if(mpi_rank == root) { in getenv_all()
|
H A D | testhdf5.h | 140 int mpi_rank; \ 142 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); \ 143 if(mpi_rank == 0 && HDGetTestVerbosity() > (V)) \
|
/dports/math/flann/flann-1.9.1-29-g1d04523/src/cpp/flann/io/ |
H A D | hdf5.h | 176 int mpi_size, mpi_rank; in load_from_file() local 178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file() 203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file() 207 offset[0] = mpi_rank*item_cnt; in load_from_file()
|
/dports/graphics/colmap/colmap-3.6/lib/FLANN/io/ |
H A D | hdf5.h | 176 int mpi_size, mpi_rank; in load_from_file() local 178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file() 203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file() 207 offset[0] = mpi_rank*item_cnt; in load_from_file()
|
/dports/graphics/hugin/hugin-2020.0.0/src/foreign/flann/io/ |
H A D | hdf5.h | 176 int mpi_size, mpi_rank; in load_from_file() local 178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file() 203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file() 207 offset[0] = mpi_rank*item_cnt; in load_from_file()
|
/dports/science/hdf5-18/hdf5-1.8.21/fortran/testpar/ |
H A D | mdset.f90 | 29 …INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator local 75 istart = mpi_rank*icount ! start position 177 wbuf(i) = n + mpi_rank 279 wbuf(i) = n + mpi_rank
|
/dports/science/hdf5/hdf5-1.10.6/fortran/testpar/ |
H A D | mdset.f90 | 29 …INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator local 75 istart = mpi_rank*icount ! start position 177 wbuf(i) = n + mpi_rank 279 wbuf(i) = n + mpi_rank
|
/dports/misc/openmvg/openMVG-2.0/src/third_party/flann/src/cpp/flann/io/ |
H A D | hdf5.h | 176 int mpi_size, mpi_rank; in load_from_file() local 178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file() 203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file() 207 offset[0] = mpi_rank*item_cnt; in load_from_file()
|
/dports/science/py-pymatgen/pymatgen-2022.0.15/pymatgen/io/abinit/ |
H A D | abitimer.py | 172 mpi_rank = info["mpi_rank"] 173 data[mpi_rank] = timer 204 def timers(self, filename=None, mpi_rank="0"): argument 209 return [self._timers[filename][mpi_rank]] 210 return [self._timers[filename][mpi_rank] for filename in self._filenames] 452 mpi_rank = "0" 453 timers = self.timers(mpi_rank=mpi_rank) 671 self.mpi_rank = info["mpi_rank"].strip() 749 frame.mpi_rank = self.mpi_rank
|
/dports/science/hdf5/hdf5-1.10.6/testpar/ |
H A D | t_filters_parallel.c | 114 static int mpi_rank; variable 1195 start[0] = (hsize_t) mpi_rank; in test_write_filtered_dataset_interleaved_write() 1368 start[2] = (hsize_t) mpi_rank; in test_write_3d_filtered_dataset_no_overlap_separate_pages() 1695 start[0] = (hsize_t) mpi_rank; in test_write_3d_filtered_dataset_overlap() 2057 start[0] = (hsize_t) mpi_rank; in test_write_cmpd_filtered_dataset_no_conversion_shared() 2445 start[0] = (hsize_t) mpi_rank; in test_write_cmpd_filtered_dataset_type_conversion_shared() 3843 start[0] = (hsize_t) mpi_rank; in test_read_filtered_dataset_interleaved_read() 4063 start[2] = (hsize_t) mpi_rank; in test_read_3d_filtered_dataset_no_overlap_separate_pages() 4490 start[0] = (hsize_t) mpi_rank; in test_read_3d_filtered_dataset_overlap() 4948 start[0] = (hsize_t) mpi_rank; in test_read_cmpd_filtered_dataset_no_conversion_shared() [all …]
|
H A D | t_file.c | 39 int mpi_size, mpi_rank; variable 75 MPI_Comm_rank(MPI_COMM_WORLD,&mpi_rank); in test_split_comm_access() 76 is_old = mpi_rank%2; in test_split_comm_access() 77 mrc = MPI_Comm_split(MPI_COMM_WORLD, is_old, mpi_rank, &comm); in test_split_comm_access() 133 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in test_page_buffer_access() 491 start[0] = mpi_rank*block[0]; in create_file() 500 data_array[i] = mpi_rank + 1; in create_file() 553 VRFY((data_array[i] == mpi_rank+1), "Dataset Verify failed"); in create_file() 650 start[0] = mpi_rank*block[0]; in open_file() 687 VRFY((data_array[i] == mpi_rank+1), "Dataset Verify failed"); in open_file() [all …]
|
/dports/devel/py-ipyparallel/ipyparallel-6.3.0/examples/ |
H A D | Using MPI with IPython Parallel.ipynb | 62 "def mpi_rank():\n", 87 "mpi_rank()" 117 "mpi_rank.block = False\n", 118 "ar = mpi_rank()\n",
|
/dports/science/hdf5/hdf5-1.10.6/test/ |
H A D | testhdf5.h | 150 int mpi_rank; \ 152 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); \ 153 if(mpi_rank == 0 && HDGetTestVerbosity() > (V)) \
|
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/examples/proxies/kripke/Kripke/ |
H A D | Grid.cpp | 155 int mpi_rank; in Grid_Data() local 156 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in Grid_Data() 157 if(mpi_rank == 0){ in Grid_Data() 418 int mpi_rank; in writeSilo() local 419 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in writeSilo() 421 if(mpi_rank == 0){ in writeSilo() 444 ss_proc << fname_base << "/rank_" << mpi_rank << ".silo"; in writeSilo()
|
/dports/science/pynn/PyNN-0.10.0/pyNN/common/ |
H A D | projections.py | 361 all_values = {self._simulator.state.mpi_rank: values} 363 if gather == 'all' or self._simulator.state.mpi_rank == 0: 378 all_values = {self._simulator.state.mpi_rank: values} 380 if gather == 'all' or self._simulator.state.mpi_rank == 0: 392 if gather == 'all' or self._simulator.state.mpi_rank == 0: 449 if self._simulator.state.mpi_rank == 0:
|
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/examples/paraview-vis/ |
H A D | paraview_ascent_source.py | 259 def extents(originDims, comm, mpi_size, mpi_rank, topology): argument 297 rankIndex = np.flatnonzero(rank == mpi_rank)[0] 299 extent[2*i+1] = startIndex[rankIndex] + size[mpi_rank] - 1 303 def extents_rectilinear(node, comm, mpi_size, mpi_rank, topology): argument 314 return extents(originDims, comm, mpi_size, mpi_rank, topology) 317 def extents_uniform(node, comm, mpi_size, mpi_rank, topology): argument 331 return extents(originDims, comm, mpi_size, mpi_rank, topology)
|
/dports/science/pynn/PyNN-0.10.0/pyNN/neuron/ |
H A D | simulator.py | 162 self.mpi_rank = int(self.parallel_context.id()) 194 self.parallel_context.set_gid2node(gid, self.mpi_rank) # assign the gid to this node 233 logger.info("Setting up transfer on MPI process {}".format(state.mpi_rank)) 237 logger.debug("default_maxstep on host #%d = %g" % (self.mpi_rank, self.default_maxstep)) 238 … logger.debug("local_minimum_delay on host #%d = %g" % (self.mpi_rank, local_minimum_delay)) 476 state.mpi_rank)) 486 state.mpi_rank))
|
/dports/math/optpp/optpp-2.4/include/ |
H A D | OptGSS.h | 94 int mpi_rank; // also used in serial code variable 102 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in setpid() 338 int getMPIRank() { return mpi_rank; } in getMPIRank()
|
/dports/science/dakota/dakota-6.13.0-release-public.src-UI/packages/external/OPTPP/include/ |
H A D | OptGSS.h | 92 int mpi_rank; // also used in serial code variable 100 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in setpid() 336 int getMPIRank() { return mpi_rank; } in getMPIRank()
|