/dports/science/silo/silo-4.10.2-bsd/tests/ |
H A D | ioperf.h | 77 int mpi_rank; member
|
/dports/science/pynn/PyNN-0.10.0/pyNN/mock/ |
H A D | simulator.py | 18 self.mpi_rank = 0
|
H A D | __init__.py | 40 simulator.state.mpi_rank = extra_params.get('rank', 0)
|
/dports/science/chemps2/CheMPS2-1.8.10/tests/ |
H A D | test4.cpp.in | 68 if ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ) 84 if ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER )
|
/dports/science/hdf5-18/hdf5-1.8.21/fortran/testpar/ |
H A D | hyper.f90 | 29 …INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator local 75 istart = mpi_rank*icount ! start position
|
/dports/science/apbs/apbs-pdb2pqr-apbs-1.5-102-g500c1473/apbs/externals/fetk/maloc/src/psh/maloc/ |
H A D | vcom.h | 63 int mpi_rank; member
|
/dports/science/tinker/tinker/apbs/include/linux/maloc/ |
H A D | vcom.h | 52 int mpi_rank; /**< Local PE rank from MPI */ member
|
/dports/science/tinker/tinker/apbs/include/macos/maloc/ |
H A D | vcom.h | 52 int mpi_rank; /**< Local PE rank from MPI */ member
|
/dports/science/libvdwxc/libvdwxc-b1e0dd854310410012d05daf4c6661b49f31b2ab/src/ |
H A D | vdw_core.h | 58 int mpi_rank; member
|
/dports/science/hdf5/hdf5-1.10.6/testpar/ |
H A D | t_filters_parallel.h | 57 #define INCREMENTAL_DATA(i) ((size_t) mpi_rank + i) /* Generates incremental test data */ 58 #define RANK_DATA(i) (mpi_rank) /* Generates test data to visibly show which …
|
/dports/science/cp2k/cp2k-2e995eec7fd208c8a72d9544807bd8b8ba8cd1cc/src/fm/ |
H A D | cp_fm_types.F | 1452 mpi_rank, ncol_block_dest, ncol_block_src, ncol_local_recv, ncol_local_send, ncols, & local 1744 mpi_rank = dest_blacs2mpi(dest_p_i, dest_q_j) 1745 send_count(mpi_rank) = send_count(mpi_rank) + 1 1746 send_buf(send_disp(mpi_rank) + send_count(mpi_rank)) = source%local_data(i, j) 1783 INTEGER :: handle, i, j, k, mpi_rank, send_size, & local 1833 mpi_rank = src_blacs2mpi(src_p_i, src_q_j) 1834 recv_count(mpi_rank) = recv_count(mpi_rank) + 1 1835 destination%local_data(i, j) = recv_buf(recv_disp(mpi_rank) + recv_count(mpi_rank))
|
/dports/science/cp2k-data/cp2k-7.1.0/src/fm/ |
H A D | cp_fm_types.F | 1434 mpi_rank, ncol_block_dest, ncol_block_src, ncol_local_recv, ncol_local_send, ncols, & local 1726 mpi_rank = dest_blacs2mpi(dest_p_i, dest_q_j) 1727 send_count(mpi_rank) = send_count(mpi_rank) + 1 1728 send_buf(send_disp(mpi_rank) + send_count(mpi_rank)) = source%local_data(i, j) 1765 INTEGER :: handle, i, j, k, mpi_rank, send_size, & local 1815 mpi_rank = src_blacs2mpi(src_p_i, src_q_j) 1816 recv_count(mpi_rank) = recv_count(mpi_rank) + 1 1817 destination%local_data(i, j) = recv_buf(recv_disp(mpi_rank) + recv_count(mpi_rank))
|
/dports/science/hdf5/hdf5-1.10.6/src/ |
H A D | H5AC.c | 313 int mpi_rank; in H5AC_create() local 319 if((mpi_rank = H5F_mpi_get_rank(f)) < 0) in H5AC_create() 330 aux_ptr->mpi_rank = mpi_rank; in H5AC_create() 352 HDsprintf(prefix, "%d:", mpi_rank); in H5AC_create() 354 if(mpi_rank == 0) { in H5AC_create() 370 if(aux_ptr->mpi_rank == 0) in H5AC_create() 1596 if(deleted && aux_ptr->mpi_rank == 0) in H5AC_unprotect() 2066 …if((aux_ptr->mpi_rank == 0) || (aux_ptr->metadata_write_strategy == H5AC_METADATA_WRITE_STRATEGY__… in H5AC__check_if_write_permitted()
|
H A D | H5ACpkg.h | 370 int mpi_rank; member
|
/dports/science/hdf5/hdf5-1.10.6/release_docs/ |
H A D | INSTALL_parallel | 247 #define PRINTID printf("Proc %d: ", mpi_rank) 252 int mpi_size, mpi_rank; 267 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); 273 if (mpi_rank==0){ 300 writedata[i] = mpi_rank*DIMSIZE + i; 301 mpi_off = mpi_rank*DIMSIZE;
|
/dports/science/chemps2/CheMPS2-1.8.10/CheMPS2/ |
H A D | executable.cpp | 67 const bool am_i_master = ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ); in file_exists() 85 const bool am_i_master = ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ); in find_integer() 114 const bool am_i_master = ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ); in find_double() 142 const bool am_i_master = ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ); in find_character() 175 const bool am_i_master = ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ); in find_boolean() 410 const bool am_i_master = ( CheMPS2::MPIchemps2::mpi_rank() == MPI_CHEMPS2_MASTER ); in main()
|
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/examples/proxies/kripke/Kripke/ |
H A D | Layout.h | 13 int mpi_rank; // Neighbors MPI rank, or -1 for boundary condition member
|
/dports/math/vtk9/VTK-9.1.0/ThirdParty/hdf5/vtkhdf5/src/ |
H A D | H5ACpkg.h | 364 int mpi_rank; member
|
H A D | H5AC.c | 297 int mpi_rank; in H5AC_create() local 303 if ((mpi_rank = H5F_mpi_get_rank(f)) < 0) in H5AC_create() 314 aux_ptr->mpi_rank = mpi_rank; in H5AC_create() 336 HDsprintf(prefix, "%d:", mpi_rank); in H5AC_create() 338 if (mpi_rank == 0) { in H5AC_create() 353 if (aux_ptr->mpi_rank == 0) in H5AC_create() 1669 if (deleted && aux_ptr->mpi_rank == 0) in H5AC_unprotect() 2156 if ((aux_ptr->mpi_rank == 0) || in H5AC__check_if_write_permitted()
|
/dports/science/pynn/PyNN-0.10.0/test/unittests/ |
H A D | test_population.py | 266 sim.simulator.state.mpi_rank = 1 277 sim.simulator.state.mpi_rank = 0 325 sim.simulator.state.mpi_rank = 1 333 sim.simulator.state.mpi_rank = 0 339 sim.simulator.state.mpi_rank = 1 347 sim.simulator.state.mpi_rank = 0
|
/dports/science/hdf5-18/hdf5-1.8.21/release_docs/ |
H A D | INSTALL_parallel | 246 #define PRINTID printf("Proc %d: ", mpi_rank) 251 int mpi_size, mpi_rank; 266 MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); 272 if (mpi_rank==0){ 299 writedata[i] = mpi_rank*DIMSIZE + i; 300 mpi_off = mpi_rank*DIMSIZE;
|
/dports/science/hdf5/hdf5-1.10.6/fortran/testpar/ |
H A D | hyper.f90 | 29 …INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator local 76 istart = mpi_rank*icount ! start position
|
/dports/science/ascent/ascent-0.7.1-66-gbcf2742a/src/ascent/utils/ |
H A D | ascent_mpi_utils.cpp | 87 int mpi_rank() in mpi_rank() function
|
/dports/science/pynn/PyNN-0.10.0/pyNN/recording/ |
H A D | __init__.py | 299 mpi_node = self._simulator.state.mpi_rank # for debugging 359 io.filename += '.%d' % self._simulator.state.mpi_rank 363 if self._simulator.state.mpi_rank == 0 or gather is False:
|
/dports/science/pynn/PyNN-0.10.0/pyNN/neuroml/ |
H A D | simulator.py | 69 self.mpi_rank = 0
|