/dports/net/mpich/mpich-3.4.3/test/mpi/pt2pt/ |
H A D | many_isend.c | 22 MPI_Request *sendreqs = NULL; in main() local 30 sendreqs = (MPI_Request *) malloc(nprocs * sizeof(MPI_Request)); in main() 43 MPI_Isend(sendbuf, BUF_COUNT, MPI_INT, i, 0, MPI_COMM_WORLD, &sendreqs[i]); in main() 62 MPI_Waitall(nprocs, sendreqs, sendstats); in main() 67 free(sendreqs); in main()
|
/dports/science/PETSc/petsc-3.14.1/src/sys/utils/ |
H A D | mpits.c | 89 MPI_Request *sendreqs,barrier; in PetscCommBuildTwoSided_Ibarrier() local 98 ierr = PetscMalloc1(nto,&sendreqs);CHKERRQ(ierr); in PetscCommBuildTwoSided_Ibarrier() 126 ierr = MPI_Testall(nsends,sendreqs,&sent,MPI_STATUSES_IGNORE);CHKERRQ(ierr); in PetscCommBuildTwoSided_Ibarrier() 134 ierr = PetscFree(sendreqs);CHKERRQ(ierr); in PetscCommBuildTwoSided_Ibarrier() 156 MPI_Request *reqs,*sendreqs; in PetscCommBuildTwoSided_Allreduce() local 181 sendreqs = reqs + nrecvs; in PetscCommBuildTwoSided_Allreduce() 207 MPI_Request *reqs,*sendreqs; in PetscCommBuildTwoSided_RedScatter() local 230 sendreqs = reqs + nrecvs; in PetscCommBuildTwoSided_RedScatter() 375 MPI_Request *sendreqs,*usendreqs,*req,barrier; in PetscCommBuildTwoSidedFReq_Ibarrier() local 388 ierr = PetscMalloc1(nto,&sendreqs);CHKERRQ(ierr); in PetscCommBuildTwoSidedFReq_Ibarrier() [all …]
|
/dports/math/combblas/CombBLAS_beta_16_2/graph500-1.2/mpi/ |
H A D | convert_to_csr.c | 105 MPI_Request* sendreqs = (MPI_Request*)xmalloc(size * sizeof(MPI_Request)); in convert_graph_to_csr() local 163 MPI_Test(&sendreqs[c], &flag, MPI_STATUS_IGNORE); \ in convert_graph_to_csr() 193 …, coalescing_counts[(dest)] * 2, INT64_T_MPI_TYPE, (dest), 0, MPI_COMM_WORLD, &sendreqs[(dest)]); \ in convert_graph_to_csr() 228 free(sendreqs); in convert_graph_to_csr()
|
/dports/net/openmpi3/openmpi-3.1.6/ompi/mca/common/ompio/ |
H A D | common_ompio_aggregators.c | 662 MPI_Request *sendreqs = NULL; in mca_common_ompio_merge_initial_groups() local 771 sendreqs = (MPI_Request *)malloc ( 2 *fh->f_procs_per_group * sizeof(MPI_Request)); in mca_common_ompio_merge_initial_groups() 772 if (NULL == sendreqs) { in mca_common_ompio_merge_initial_groups() 788 sendreqs + r++)); in mca_common_ompio_merge_initial_groups() 801 sendreqs + r++)); in mca_common_ompio_merge_initial_groups() 845 ret = ompi_request_wait_all (r, sendreqs, MPI_STATUSES_IGNORE); in mca_common_ompio_merge_initial_groups() 849 if (NULL != sendreqs) { in mca_common_ompio_merge_initial_groups() 850 free(sendreqs); in mca_common_ompio_merge_initial_groups()
|
/dports/net/openmpi/openmpi-4.1.1/ompi/mca/common/ompio/ |
H A D | common_ompio_aggregators.c | 694 MPI_Request *sendreqs = NULL; in mca_common_ompio_merge_initial_groups() local 803 sendreqs = (MPI_Request *)malloc ( 2 *fh->f_procs_per_group * sizeof(MPI_Request)); in mca_common_ompio_merge_initial_groups() 804 if (NULL == sendreqs) { in mca_common_ompio_merge_initial_groups() 820 sendreqs + r++)); in mca_common_ompio_merge_initial_groups() 833 sendreqs + r++)); in mca_common_ompio_merge_initial_groups() 877 ret = ompi_request_wait_all (r, sendreqs, MPI_STATUSES_IGNORE); in mca_common_ompio_merge_initial_groups() 881 if (NULL != sendreqs) { in mca_common_ompio_merge_initial_groups() 882 free(sendreqs); in mca_common_ompio_merge_initial_groups()
|
/dports/science/PETSc/petsc-3.14.1/src/vec/vec/impls/mpi/ |
H A D | pvecimpl.h | 37 MPI_Request *sendreqs; member
|
H A D | pbvec.c | 251 …ierr = VecAssemblySend_MPI_Private(comm,tag,i,x->sendranks[i],x->sendhdr+i,x->sendreqs+4*i,X);CHKE… in VecAssemblyBegin_MPI_BTS() 258 …ks,(PetscInt*)x->sendhdr,&x->nrecvranks,&x->recvranks,&x->recvhdr,4,&x->sendreqs,&x->recvreqs,VecA… in VecAssemblyBegin_MPI_BTS() 356 ierr = MPI_Waitall(4*x->nsendranks,x->sendreqs,MPI_STATUSES_IGNORE);CHKERRQ(ierr); in VecAssemblyEnd_MPI_BTS() 379 ierr = PetscFree(x->sendreqs);CHKERRQ(ierr); in VecAssemblyReset_MPI()
|
/dports/science/PETSc/petsc-3.14.1/src/mat/utils/ |
H A D | matstash.c | 935 …ivate(stash->comm,&tag,i,stash->sendranks[i],&stash->sendhdr[i],&stash->sendreqs[i],stash);CHKERRQ… in MatStashScatterBegin_BTS() 940 …stash->nrecvranks,&stash->recvranks,(PetscInt*)&stash->recvhdr,1,&stash->sendreqs,&stash->recvreqs, in MatStashScatterBegin_BTS() 1001 ierr = MPI_Waitall(stash->nsendranks,stash->sendreqs,MPI_STATUSES_IGNORE);CHKERRQ(ierr); in MatStashScatterEnd_BTS() 1045 ierr = PetscFree(stash->sendreqs);CHKERRQ(ierr); in MatStashScatterDestroy_BTS()
|
/dports/net/py-mpi4py/mpi4py-3.1.3/src/mpi4py/futures/ |
H A D | _lib.py | 434 sendreqs, recvreqs = [], [] 437 sendreqs.append(comm.Issend(buf, pid, tag)) 441 MPI.Request.Waitall(sendreqs)
|
/dports/science/PETSc/petsc-3.14.1/include/petsc/private/ |
H A D | matimpl.h | 351 MPI_Request *sendreqs; member
|
/dports/science/PETSc/petsc-3.14.1/ |
H A D | TAGS | 10271 MPI_Request *sendreqs;sendreqs351,17037 86266 MPI_Request *sendreqs;sendreqs37,1171
|
H A D | CTAGS | 77319 sendreqs include/petsc/private/matimpl.h /^ MPI_Request *sendreqs;$/;" m language:C++ struct:_M… 77320 sendreqs src/vec/vec/impls/mpi/pvecimpl.h /^ MPI_Request *sendreqs;$/;" m language:C++ struct:_…
|