Home
last modified time | relevance | path

Searched refs:mpi_rank (Results 101 – 125 of 373) sorted by relevance

12345678910>>...15

/dports/science/hdf5-18/hdf5-1.8.21/src/
H A DH5FDmpi.c381 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor()
383 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor()
431 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor()
432 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
H A DH5AC.c427 int mpi_rank; in H5AC_create() local
444 aux_ptr->mpi_rank = mpi_rank; in H5AC_create()
470 if(mpi_rank == 0) { in H5AC_create()
3128 if(aux_ptr->mpi_rank == 0) { in H5AC_log_deleted_entry()
3219 if ( aux_ptr->mpi_rank == 0 ) { in H5AC_log_dirtied_entry()
3461 if(aux_ptr->mpi_rank == 0) { in H5AC_log_inserted_entry()
3588 if ( aux_ptr->mpi_rank == 0 ) { in H5AC_log_moved_entry()
3828 if(aux_ptr->mpi_rank == 0) { in H5AC_propagate_and_apply_candidate_list()
3872 if(aux_ptr->mpi_rank == 0) { in H5AC_propagate_and_apply_candidate_list()
3986 if(aux_ptr->mpi_rank == 0) { in H5AC_propagate_flushed_and_still_clean_entries_list()
[all …]
/dports/science/hdf5/hdf5-1.10.6/src/
H A DH5FDmpi.c425 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor()
427 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor()
475 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor()
476 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
/dports/math/vtk6/VTK-6.2.0/ThirdParty/hdf5/vtkhdf5/src/
H A DH5FDmpi.c383 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor()
385 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor()
433 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor()
434 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
/dports/math/vtk8/VTK-8.2.0/ThirdParty/hdf5/vtkhdf5/src/
H A DH5FDmpi.c420 if (file->mpi_rank != 0) { in H5FD_mpio_wait_for_left_neighbor()
422 file->mpi_rank-1, MPI_ANY_TAG, file->comm, &rcvstat ))) in H5FD_mpio_wait_for_left_neighbor()
470 if(file->mpi_rank != (file->mpi_size - 1)) in H5FD_mpio_signal_right_neighbor()
471 …if(MPI_SUCCESS != (mpi_code=MPI_Send(&msgbuf, 0/*empty msg*/, MPI_CHAR, file->mpi_rank + 1, 0, fil… in H5FD_mpio_signal_right_neighbor()
/dports/graphics/opencv/opencv-4.5.3/modules/flann/include/opencv2/flann/
H A Dhdf5.h176 int mpi_size, mpi_rank; in load_from_file() local
178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file()
203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file()
207 offset[0] = mpi_rank*item_cnt; in load_from_file()
/dports/science/hdf5-18/hdf5-1.8.21/test/
H A Dh5test.c544 int mpi_rank; in h5_fixname_real() local
546 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in h5_fixname_real()
548 if (mpi_rank == 0) in h5_fixname_real()
883 int mpi_rank, mpi_initialized, mpi_finalized; in h5_show_hostname() local
889 MPI_Comm_rank(MPI_COMM_WORLD,&mpi_rank); in h5_show_hostname()
890 printf("MPI-process %d.", mpi_rank); in h5_show_hostname()
1278 int mpi_size, mpi_rank, mpi_initialized, mpi_finalized; in getenv_all() local
1288 MPI_Comm_rank(comm, &mpi_rank); in getenv_all()
1294 if(mpi_rank == root) { in getenv_all()
H A Dtesthdf5.h140 int mpi_rank; \
142 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); \
143 if(mpi_rank == 0 && HDGetTestVerbosity() > (V)) \
/dports/math/flann/flann-1.9.1-29-g1d04523/src/cpp/flann/io/
H A Dhdf5.h176 int mpi_size, mpi_rank; in load_from_file() local
178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file()
203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file()
207 offset[0] = mpi_rank*item_cnt; in load_from_file()
/dports/graphics/colmap/colmap-3.6/lib/FLANN/io/
H A Dhdf5.h176 int mpi_size, mpi_rank; in load_from_file() local
178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file()
203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file()
207 offset[0] = mpi_rank*item_cnt; in load_from_file()
/dports/graphics/hugin/hugin-2020.0.0/src/foreign/flann/io/
H A Dhdf5.h176 int mpi_size, mpi_rank; in load_from_file() local
178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file()
203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file()
207 offset[0] = mpi_rank*item_cnt; in load_from_file()
/dports/science/hdf5-18/hdf5-1.8.21/fortran/testpar/
H A Dmdset.f9029 …INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator local
75 istart = mpi_rank*icount ! start position
177 wbuf(i) = n + mpi_rank
279 wbuf(i) = n + mpi_rank
/dports/science/hdf5/hdf5-1.10.6/fortran/testpar/
H A Dmdset.f9029 …INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator local
75 istart = mpi_rank*icount ! start position
177 wbuf(i) = n + mpi_rank
279 wbuf(i) = n + mpi_rank
/dports/misc/openmvg/openMVG-2.0/src/third_party/flann/src/cpp/flann/io/
H A Dhdf5.h176 int mpi_size, mpi_rank; in load_from_file() local
178 MPI_Comm_rank(comm, &mpi_rank); in load_from_file()
203 hsize_t cnt = (mpi_rank<mpi_size-1 ? item_cnt : dims[0]-item_cnt*(mpi_size-1)); in load_from_file()
207 offset[0] = mpi_rank*item_cnt; in load_from_file()
/dports/science/py-pymatgen/pymatgen-2022.0.15/pymatgen/io/abinit/
H A Dabitimer.py172 mpi_rank = info["mpi_rank"]
173 data[mpi_rank] = timer
204 def timers(self, filename=None, mpi_rank="0"): argument
209 return [self._timers[filename][mpi_rank]]
210 return [self._timers[filename][mpi_rank] for filename in self._filenames]
452 mpi_rank = "0"
453 timers = self.timers(mpi_rank=mpi_rank)
671 self.mpi_rank = info["mpi_rank"].strip()
749 frame.mpi_rank = self.mpi_rank
/dports/science/hdf5/hdf5-1.10.6/testpar/
H A Dt_filters_parallel.c114 static int mpi_rank; variable
1195 start[0] = (hsize_t) mpi_rank; in test_write_filtered_dataset_interleaved_write()
1368 start[2] = (hsize_t) mpi_rank; in test_write_3d_filtered_dataset_no_overlap_separate_pages()
1695 start[0] = (hsize_t) mpi_rank; in test_write_3d_filtered_dataset_overlap()
2057 start[0] = (hsize_t) mpi_rank; in test_write_cmpd_filtered_dataset_no_conversion_shared()
2445 start[0] = (hsize_t) mpi_rank; in test_write_cmpd_filtered_dataset_type_conversion_shared()
3843 start[0] = (hsize_t) mpi_rank; in test_read_filtered_dataset_interleaved_read()
4063 start[2] = (hsize_t) mpi_rank; in test_read_3d_filtered_dataset_no_overlap_separate_pages()
4490 start[0] = (hsize_t) mpi_rank; in test_read_3d_filtered_dataset_overlap()
4948 start[0] = (hsize_t) mpi_rank; in test_read_cmpd_filtered_dataset_no_conversion_shared()
[all …]
H A Dt_file.c39 int mpi_size, mpi_rank; variable
75 MPI_Comm_rank(MPI_COMM_WORLD,&mpi_rank); in test_split_comm_access()
76 is_old = mpi_rank%2; in test_split_comm_access()
77 mrc = MPI_Comm_split(MPI_COMM_WORLD, is_old, mpi_rank, &comm); in test_split_comm_access()
133 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in test_page_buffer_access()
491 start[0] = mpi_rank*block[0]; in create_file()
500 data_array[i] = mpi_rank + 1; in create_file()
553 VRFY((data_array[i] == mpi_rank+1), "Dataset Verify failed"); in create_file()
650 start[0] = mpi_rank*block[0]; in open_file()
687 VRFY((data_array[i] == mpi_rank+1), "Dataset Verify failed"); in open_file()
[all …]
/dports/devel/py-ipyparallel/ipyparallel-6.3.0/examples/
H A DUsing MPI with IPython Parallel.ipynb62 "def mpi_rank():\n",
87 "mpi_rank()"
117 "mpi_rank.block = False\n",
118 "ar = mpi_rank()\n",
/dports/science/hdf5/hdf5-1.10.6/test/
H A Dtesthdf5.h150 int mpi_rank; \
152 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); \
153 if(mpi_rank == 0 && HDGetTestVerbosity() > (V)) \
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/examples/proxies/kripke/Kripke/
H A DGrid.cpp155 int mpi_rank; in Grid_Data() local
156 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in Grid_Data()
157 if(mpi_rank == 0){ in Grid_Data()
418 int mpi_rank; in writeSilo() local
419 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in writeSilo()
421 if(mpi_rank == 0){ in writeSilo()
444 ss_proc << fname_base << "/rank_" << mpi_rank << ".silo"; in writeSilo()
/dports/science/pynn/PyNN-0.10.0/pyNN/common/
H A Dprojections.py361 all_values = {self._simulator.state.mpi_rank: values}
363 if gather == 'all' or self._simulator.state.mpi_rank == 0:
378 all_values = {self._simulator.state.mpi_rank: values}
380 if gather == 'all' or self._simulator.state.mpi_rank == 0:
392 if gather == 'all' or self._simulator.state.mpi_rank == 0:
449 if self._simulator.state.mpi_rank == 0:
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/examples/paraview-vis/
H A Dparaview_ascent_source.py259 def extents(originDims, comm, mpi_size, mpi_rank, topology): argument
297 rankIndex = np.flatnonzero(rank == mpi_rank)[0]
299 extent[2*i+1] = startIndex[rankIndex] + size[mpi_rank] - 1
303 def extents_rectilinear(node, comm, mpi_size, mpi_rank, topology): argument
314 return extents(originDims, comm, mpi_size, mpi_rank, topology)
317 def extents_uniform(node, comm, mpi_size, mpi_rank, topology): argument
331 return extents(originDims, comm, mpi_size, mpi_rank, topology)
/dports/science/pynn/PyNN-0.10.0/pyNN/neuron/
H A Dsimulator.py162 self.mpi_rank = int(self.parallel_context.id())
194 self.parallel_context.set_gid2node(gid, self.mpi_rank) # assign the gid to this node
233 logger.info("Setting up transfer on MPI process {}".format(state.mpi_rank))
237 logger.debug("default_maxstep on host #%d = %g" % (self.mpi_rank, self.default_maxstep))
238 … logger.debug("local_minimum_delay on host #%d = %g" % (self.mpi_rank, local_minimum_delay))
476 state.mpi_rank))
486 state.mpi_rank))
/dports/math/optpp/optpp-2.4/include/
H A DOptGSS.h94 int mpi_rank; // also used in serial code variable
102 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in setpid()
338 int getMPIRank() { return mpi_rank; } in getMPIRank()
/dports/science/dakota/dakota-6.13.0-release-public.src-UI/packages/external/OPTPP/include/
H A DOptGSS.h92 int mpi_rank; // also used in serial code variable
100 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); in setpid()
336 int getMPIRank() { return mpi_rank; } in getMPIRank()

12345678910>>...15