Home
last modified time | relevance | path

Searched refs:sendreqs (Results 1 – 12 of 12) sorted by relevance

/dports/net/mpich/mpich-3.4.3/test/mpi/pt2pt/
H A Dmany_isend.c22 MPI_Request *sendreqs = NULL; in main() local
30 sendreqs = (MPI_Request *) malloc(nprocs * sizeof(MPI_Request)); in main()
43 MPI_Isend(sendbuf, BUF_COUNT, MPI_INT, i, 0, MPI_COMM_WORLD, &sendreqs[i]); in main()
62 MPI_Waitall(nprocs, sendreqs, sendstats); in main()
67 free(sendreqs); in main()
/dports/science/PETSc/petsc-3.14.1/src/sys/utils/
H A Dmpits.c89 MPI_Request *sendreqs,barrier; in PetscCommBuildTwoSided_Ibarrier() local
98 ierr = PetscMalloc1(nto,&sendreqs);CHKERRQ(ierr); in PetscCommBuildTwoSided_Ibarrier()
126 ierr = MPI_Testall(nsends,sendreqs,&sent,MPI_STATUSES_IGNORE);CHKERRQ(ierr); in PetscCommBuildTwoSided_Ibarrier()
134 ierr = PetscFree(sendreqs);CHKERRQ(ierr); in PetscCommBuildTwoSided_Ibarrier()
156 MPI_Request *reqs,*sendreqs; in PetscCommBuildTwoSided_Allreduce() local
181 sendreqs = reqs + nrecvs; in PetscCommBuildTwoSided_Allreduce()
207 MPI_Request *reqs,*sendreqs; in PetscCommBuildTwoSided_RedScatter() local
230 sendreqs = reqs + nrecvs; in PetscCommBuildTwoSided_RedScatter()
375 MPI_Request *sendreqs,*usendreqs,*req,barrier; in PetscCommBuildTwoSidedFReq_Ibarrier() local
388 ierr = PetscMalloc1(nto,&sendreqs);CHKERRQ(ierr); in PetscCommBuildTwoSidedFReq_Ibarrier()
[all …]
/dports/math/combblas/CombBLAS_beta_16_2/graph500-1.2/mpi/
H A Dconvert_to_csr.c105 MPI_Request* sendreqs = (MPI_Request*)xmalloc(size * sizeof(MPI_Request)); in convert_graph_to_csr() local
163 MPI_Test(&sendreqs[c], &flag, MPI_STATUS_IGNORE); \ in convert_graph_to_csr()
193 …, coalescing_counts[(dest)] * 2, INT64_T_MPI_TYPE, (dest), 0, MPI_COMM_WORLD, &sendreqs[(dest)]); \ in convert_graph_to_csr()
228 free(sendreqs); in convert_graph_to_csr()
/dports/net/openmpi3/openmpi-3.1.6/ompi/mca/common/ompio/
H A Dcommon_ompio_aggregators.c662 MPI_Request *sendreqs = NULL; in mca_common_ompio_merge_initial_groups() local
771 sendreqs = (MPI_Request *)malloc ( 2 *fh->f_procs_per_group * sizeof(MPI_Request)); in mca_common_ompio_merge_initial_groups()
772 if (NULL == sendreqs) { in mca_common_ompio_merge_initial_groups()
788 sendreqs + r++)); in mca_common_ompio_merge_initial_groups()
801 sendreqs + r++)); in mca_common_ompio_merge_initial_groups()
845 ret = ompi_request_wait_all (r, sendreqs, MPI_STATUSES_IGNORE); in mca_common_ompio_merge_initial_groups()
849 if (NULL != sendreqs) { in mca_common_ompio_merge_initial_groups()
850 free(sendreqs); in mca_common_ompio_merge_initial_groups()
/dports/net/openmpi/openmpi-4.1.1/ompi/mca/common/ompio/
H A Dcommon_ompio_aggregators.c694 MPI_Request *sendreqs = NULL; in mca_common_ompio_merge_initial_groups() local
803 sendreqs = (MPI_Request *)malloc ( 2 *fh->f_procs_per_group * sizeof(MPI_Request)); in mca_common_ompio_merge_initial_groups()
804 if (NULL == sendreqs) { in mca_common_ompio_merge_initial_groups()
820 sendreqs + r++)); in mca_common_ompio_merge_initial_groups()
833 sendreqs + r++)); in mca_common_ompio_merge_initial_groups()
877 ret = ompi_request_wait_all (r, sendreqs, MPI_STATUSES_IGNORE); in mca_common_ompio_merge_initial_groups()
881 if (NULL != sendreqs) { in mca_common_ompio_merge_initial_groups()
882 free(sendreqs); in mca_common_ompio_merge_initial_groups()
/dports/science/PETSc/petsc-3.14.1/src/vec/vec/impls/mpi/
H A Dpvecimpl.h37 MPI_Request *sendreqs; member
H A Dpbvec.c251 …ierr = VecAssemblySend_MPI_Private(comm,tag,i,x->sendranks[i],x->sendhdr+i,x->sendreqs+4*i,X);CHKE… in VecAssemblyBegin_MPI_BTS()
258 …ks,(PetscInt*)x->sendhdr,&x->nrecvranks,&x->recvranks,&x->recvhdr,4,&x->sendreqs,&x->recvreqs,VecA… in VecAssemblyBegin_MPI_BTS()
356 ierr = MPI_Waitall(4*x->nsendranks,x->sendreqs,MPI_STATUSES_IGNORE);CHKERRQ(ierr); in VecAssemblyEnd_MPI_BTS()
379 ierr = PetscFree(x->sendreqs);CHKERRQ(ierr); in VecAssemblyReset_MPI()
/dports/science/PETSc/petsc-3.14.1/src/mat/utils/
H A Dmatstash.c935 …ivate(stash->comm,&tag,i,stash->sendranks[i],&stash->sendhdr[i],&stash->sendreqs[i],stash);CHKERRQ… in MatStashScatterBegin_BTS()
940 …stash->nrecvranks,&stash->recvranks,(PetscInt*)&stash->recvhdr,1,&stash->sendreqs,&stash->recvreqs, in MatStashScatterBegin_BTS()
1001 ierr = MPI_Waitall(stash->nsendranks,stash->sendreqs,MPI_STATUSES_IGNORE);CHKERRQ(ierr); in MatStashScatterEnd_BTS()
1045 ierr = PetscFree(stash->sendreqs);CHKERRQ(ierr); in MatStashScatterDestroy_BTS()
/dports/net/py-mpi4py/mpi4py-3.1.3/src/mpi4py/futures/
H A D_lib.py434 sendreqs, recvreqs = [], []
437 sendreqs.append(comm.Issend(buf, pid, tag))
441 MPI.Request.Waitall(sendreqs)
/dports/science/PETSc/petsc-3.14.1/include/petsc/private/
H A Dmatimpl.h351 MPI_Request *sendreqs; member
/dports/science/PETSc/petsc-3.14.1/
H A DTAGS10271 MPI_Request *sendreqs;sendreqs351,17037
86266 MPI_Request *sendreqs;sendreqs37,1171
H A DCTAGS77319 sendreqs include/petsc/private/matimpl.h /^ MPI_Request *sendreqs;$/;" m language:C++ struct:_M…
77320 sendreqs src/vec/vec/impls/mpi/pvecimpl.h /^ MPI_Request *sendreqs;$/;" m language:C++ struct:_…