/dports/science/py-gpaw/gpaw-21.6.0/gpaw/test/ |
H A D | test_mpicomm.py | 38 subcomm = comm.new_communicator(np.array([comm.rank])) 39 assert subcomm.parent is comm 40 assert subcomm.rank == 0 and subcomm.size == 1 41 assert subcomm.get_members().item() == comm.rank 46 assert isinstance(subcomm, _Communicator) 50 assert isinstance(subcomm, SerialCommunicator) 54 assert isinstance(subcomm, _gpaw.Communicator)
|
/dports/math/py-yt/yt-4.0.1/yt/utilities/parallel_tools/ |
H A D | task_queue.py | 21 def __init__(self, tasks, comm, subcomm): argument 25 self.subcomm = subcomm 30 if self.subcomm.rank == 0: 32 self.subcomm.barrier() 36 if self.subcomm.rank == 0: 39 msg = self.subcomm.bcast(msg, root=0) 133 subcomm = communication_system.push_with_ids(all_new_comms[my_new_id].tolist()) 138 my_q = TaskQueueNonRoot(None, comm, subcomm) 165 subcomm = communication_system.push_with_ids(all_new_comms[my_new_id].tolist()) 171 my_q = TaskQueueNonRoot(None, comm, subcomm)
|
/dports/science/PETSc/petsc-3.14.1/src/sys/tests/ |
H A D | ex50.c | 19 MPI_Comm comm,subcomm,subsubcomm; in main() local 45 subcomm = PetscSubcommChild(psubcomm); in main() 47 ierr = PetscViewerGetSubViewer(viewer,subcomm,&subviewer);CHKERRQ(ierr); in main() 51 ierr = PetscSubcommCreate(subcomm,&psubsubcomm);CHKERRQ(ierr); in main() 63 ierr = PetscViewerRestoreSubViewer(viewer,subcomm,&subviewer);CHKERRQ(ierr); in main()
|
/dports/science/PETSc/petsc-3.14.1/src/ksp/ksp/tests/ |
H A D | ex37.c | 24 MPI_Comm comm,subcomm; in main() local 76 subcomm = PetscSubcommChild(psubcomm); in main() 93 ierr = MPI_Comm_split(dcomm,color,subrank,&subcomm);CHKERRQ(ierr); in main() 95 ierr = MatCreate(subcomm,&subA);CHKERRQ(ierr); in main() 115 ierr = MatGetMultiProcBlock(A,subcomm,MAT_INITIAL_MATRIX,&subA);CHKERRQ(ierr); in main() 116 ierr = MatGetMultiProcBlock(A,subcomm,MAT_REUSE_MATRIX,&subA);CHKERRQ(ierr); in main() 120 ierr = VecCreateMPIWithArray(subcomm,1,m,PETSC_DECIDE,NULL,&subb);CHKERRQ(ierr); in main() 121 ierr = VecCreateMPIWithArray(subcomm,1,n,PETSC_DECIDE,NULL,&subx);CHKERRQ(ierr); in main() 122 ierr = VecCreateMPIWithArray(subcomm,1,n,PETSC_DECIDE,NULL,&subu);CHKERRQ(ierr); in main() 132 ierr = KSPCreate(subcomm,&subksp);CHKERRQ(ierr); in main() [all …]
|
/dports/science/PETSc/petsc-3.14.1/src/sys/objects/ |
H A D | subcomm.c | 205 MPI_Comm subcomm=0,dupcomm=0,comm=psubcomm->parent; in PetscSubcommSetTypeGeneral() local 213 ierr = MPI_Comm_split(comm,color,subrank,&subcomm);CHKERRQ(ierr); in PetscSubcommSetTypeGeneral() 221 ierr = MPI_Comm_size(subcomm,&mysubsize);CHKERRQ(ierr); in PetscSubcommSetTypeGeneral() 245 ierr = PetscCommDuplicate(subcomm,&psubcomm->child,NULL);CHKERRQ(ierr); in PetscSubcommSetTypeGeneral() 247 ierr = MPI_Comm_free(&subcomm);CHKERRQ(ierr); in PetscSubcommSetTypeGeneral() 391 MPI_Comm subcomm=0,dupcomm=0,comm=psubcomm->parent; in PetscSubcommCreate_contiguous() local 418 ierr = MPI_Comm_split(comm,color,subrank,&subcomm);CHKERRQ(ierr); in PetscSubcommCreate_contiguous() 425 ierr = MPI_Comm_free(&subcomm);CHKERRQ(ierr); in PetscSubcommCreate_contiguous() 463 MPI_Comm subcomm=0,dupcomm=0,comm=psubcomm->parent; in PetscSubcommCreate_interlaced() local 483 ierr = MPI_Comm_split(comm,color,subrank,&subcomm);CHKERRQ(ierr); in PetscSubcommCreate_interlaced() [all …]
|
/dports/science/PETSc/petsc-3.14.1/src/vec/vscat/tests/ |
H A D | ex9.c | 11 MPI_Comm subcomm; in main() local 31 ierr = MPI_Comm_split(PETSC_COMM_WORLD,mycolor,grank,&subcomm);CHKERRQ(ierr); in main() 55 ierr = VecCreateMPI(subcomm,PETSC_DECIDE,N,&y);CHKERRQ(ierr); in main() 80 ierr = VecView(y,PETSC_VIEWER_STDOUT_(subcomm));CHKERRQ(ierr); in main() 138 ierr = MPI_Comm_rank(subcomm,&lrank);CHKERRQ(ierr); in main() 197 ierr = MPI_Comm_rank(subcomm,&lrank);CHKERRQ(ierr); in main() 203 ierr = MPI_Bcast(&N,1,MPIU_INT,0/*local root*/,subcomm);CHKERRQ(ierr); in main() 211 ierr = VecCreateMPI(subcomm,PETSC_DECIDE,N,&y);CHKERRQ(ierr); in main() 236 ierr = VecView(y,PETSC_VIEWER_STDOUT_(subcomm));CHKERRQ(ierr); in main() 270 ierr = VecCreateMPI(subcomm,PETSC_DECIDE,N,&y);CHKERRQ(ierr); in main() [all …]
|
/dports/math/vtk6/VTK-6.2.0/Parallel/MPI/ |
H A D | vtkMPIController.cxx | 316 VTK_CREATE(vtkMPICommunicator, subcomm); in CreateSubController() 318 if (!subcomm->Initialize(group)) return NULL; in CreateSubController() 326 if (*(subcomm->GetMPIComm()->Handle) == MPI_COMM_NULL) return NULL; in CreateSubController() 329 controller->SetCommunicator(subcomm); in CreateSubController() 337 VTK_CREATE(vtkMPICommunicator, subcomm); in PartitionController() 339 if (!subcomm->SplitInitialize(this->Communicator, localColor, localKey)) in PartitionController() 345 controller->SetCommunicator(subcomm); in PartitionController()
|
/dports/math/vtk8/VTK-8.2.0/Parallel/MPI/ |
H A D | vtkMPIController.cxx | 317 VTK_CREATE(vtkMPICommunicator, subcomm); in CreateSubController() 319 if (!subcomm->Initialize(group)) in CreateSubController() 330 if (*(subcomm->GetMPIComm()->Handle) == MPI_COMM_NULL) in CreateSubController() 336 controller->SetCommunicator(subcomm); in CreateSubController() 344 VTK_CREATE(vtkMPICommunicator, subcomm); in PartitionController() 346 if (!subcomm->SplitInitialize(this->Communicator, localColor, localKey)) in PartitionController() 352 controller->SetCommunicator(subcomm); in PartitionController()
|
/dports/math/vtk9/VTK-9.1.0/Parallel/MPI/ |
H A D | vtkMPIController.cxx | 304 VTK_CREATE(vtkMPICommunicator, subcomm); in CreateSubController() 306 if (!subcomm->Initialize(group)) in CreateSubController() 317 if (*(subcomm->GetMPIComm()->Handle) == MPI_COMM_NULL) in CreateSubController() 323 controller->SetCommunicator(subcomm); in CreateSubController() 330 VTK_CREATE(vtkMPICommunicator, subcomm); in PartitionController() 332 if (!subcomm->SplitInitialize(this->Communicator, localColor, localKey)) in PartitionController() 338 controller->SetCommunicator(subcomm); in PartitionController()
|
/dports/science/PETSc/petsc-3.14.1/src/dm/impls/da/ftn-custom/ |
H A D | zdasubf.c | 20 MPI_Comm subcomm; in dmdagetprocessorsubsets_() local 21 *__ierr = DMDAGetProcessorSubsets(*da,*dir,&subcomm); in dmdagetprocessorsubsets_() 22 *subfcomm = MPI_Comm_c2f(subcomm); in dmdagetprocessorsubsets_()
|
/dports/math/tblis/tblis-1.2.0/src/nodes/ |
H A D | gemm.hpp | 132 node<partition_gemm_nc>(child).subcomm = &comm_nc; in operator ()() 133 node<partition_gemm_kc>(child).subcomm = &comm_kc; in operator ()() 134 node<partition_gemm_mc>(child).subcomm = &comm_mc; in operator ()() 135 node<partition_gemm_nr>(child).subcomm = &comm_nr; in operator ()() 136 node<partition_gemm_mr>(child).subcomm = &comm_mr; in operator ()()
|
H A D | partm.hpp | 16 communicator* subcomm = nullptr; member 40 subcomm->distribute_over_gangs({std::min(m_u, m_v), M_iota}, in operator ()() 82 child(*subcomm, cfg, alpha, local_A, local_B, local_beta, local_C); in operator ()()
|
/dports/science/PETSc/petsc-3.14.1/src/ksp/pc/impls/telescope/ |
H A D | telescope.c | 142 MPI_Comm comm,subcomm; in PCTelescopeSetUp_default() local 149 subcomm = PetscSubcommChild(sred->psubcomm); in PCTelescopeSetUp_default() 199 MPI_Comm comm,subcomm; in PCTelescopeMatCreate_default() local 240 MPI_Comm subcomm; in PCTelescopeSubNullSpaceCreate_Telescope() local 334 MPI_Comm comm,subcomm; in PCView_Telescope() local 363 subcomm = sred->subcomm; in PCView_Telescope() 365 subcomm = PETSC_COMM_SELF; in PCView_Telescope() 437 MPI_Comm comm,subcomm=0; in PCSetUp_Telescope() local 547 sred->subcomm = comm_coarse_partition; in PCSetUp_Telescope() 550 subcomm = sred->subcomm; in PCSetUp_Telescope() [all …]
|
H A D | telescope.h | 12 MPI_Comm subcomm; member 48 if (sred->subcomm != MPI_COMM_NULL) return(PETSC_TRUE); in PCTelescope_isActiveRank()
|
/dports/math/tblis/tblis-1.2.0/src/external/tci/tci/ |
H A D | task_set.c | 28 tci_comm_gang(comm, &set->subcomm, TCI_EVENLY, nt_outer, 0); in tci_task_set_init() 34 tci_comm_destroy(&set->subcomm); in tci_task_set_destroy() 43 if (!tci_slot_try_fill(set->slots+task+1, 0, set->subcomm.gid+1)) in tci_task_set_visit() 46 func(&set->subcomm, task, payload); in tci_task_set_visit() 164 *(dispatch_queue_t*)&set->subcomm = 201 dispatch_queue_t queue = *(dispatch_queue_t*)&set->subcomm;
|
/dports/math/tblis/tblis-1.2.0/src/internal/1t/indexed/ |
H A D | add.cxx | 77 [&,idx_A,idx_B,next_A](const communicator& subcomm) in trace_block() 93 add(subcomm, cfg, group_A.dense_len, {}, group_AB.dense_len, in trace_block() 141 [&,idx_A,local_idx_B,factor](const communicator& subcomm) in replicate_block() 149 add(subcomm, cfg, {}, group_B.dense_len, group_AB.dense_len, in replicate_block() 188 [&,idx_A,idx_B,factor](const communicator& subcomm) in transpose_block() 197 add(subcomm, cfg, {}, {}, group_AB.dense_len, in transpose_block()
|
H A D | dot.cxx | 71 [&,idx_A,idx_B,factor](const communicator& subcomm) in dot_block() 81 dot(subcomm, cfg, group_AB.dense_len, in dot_block() 86 if (subcomm.master()) local_result += factor*block_result; in dot_block()
|
/dports/science/PETSc/petsc-3.14.1/src/mat/tests/ |
H A D | ex9.c | 17 MPI_Comm subcomm; in main() local 87 ierr = PetscObjectGetComm((PetscObject)Credundant,&subcomm);CHKERRQ(ierr); in main() 88 ierr = MPI_Comm_size(subcomm,&subsize);CHKERRQ(ierr); in main() 91 …ierr = PetscViewerASCIIPrintf(PETSC_VIEWER_STDOUT_(subcomm),"\n[%d] Credundant:\n",rank);CHKERRQ(i… in main() 92 ierr = MatView(Credundant,PETSC_VIEWER_STDOUT_(subcomm));CHKERRQ(ierr); in main()
|
H A D | ex183.c | 16 MPI_Comm subcomm; in main() local 88 ierr = MPI_Comm_split(PETSC_COMM_WORLD,color,rank,&subcomm);CHKERRQ(ierr); in main() 89 ierr = MPI_Comm_size(subcomm,&subsize);CHKERRQ(ierr); in main() 90 ierr = MPI_Comm_rank(subcomm,&subrank);CHKERRQ(ierr); in main() 104 ierr = ISCreateGeneral(subcomm,rend-rstart,rowindices,PETSC_COPY_VALUES,&rowis[0]);CHKERRQ(ierr); in main() 105 ierr = ISCreateGeneral(subcomm,rend-rstart,colindices,PETSC_COPY_VALUES,&colis[0]);CHKERRQ(ierr); in main() 209 ierr = MPI_Comm_free(&subcomm);CHKERRQ(ierr); in main()
|
/dports/science/PETSc/petsc-3.14.1/src/sys/classes/viewer/tests/ |
H A D | ex6.c | 78 MPI_Comm subcomm = PetscObjectComm((PetscObject)subviewer); in TestRead() local 81 if (idata != 42) SETERRQ1(subcomm,PETSC_ERR_FILE_UNEXPECTED,"Unexpected idata=%D",idata); in TestRead() 82 … if (rdata != 42) SETERRQ1(subcomm,PETSC_ERR_FILE_UNEXPECTED,"Unexpected rdata=%g",(double)rdata); in TestRead() 93 MPI_Comm subcomm = PetscObjectComm((PetscObject)subviewer); in TestRead() local 96 if (idata != 42) SETERRQ1(subcomm,PETSC_ERR_FILE_UNEXPECTED,"Unexpected idata=%D",idata); in TestRead() 97 … if (rdata != 42) SETERRQ1(subcomm,PETSC_ERR_FILE_UNEXPECTED,"Unexpected rdata=%g",(double)rdata); in TestRead()
|
/dports/math/libmesh/libmesh-1.6.2/tests/parallel/ |
H A D | parallel_test.C | 677 Parallel::Communicator subcomm; in testSplit() local 680 TestCommWorld->split(color, rank, subcomm); in testSplit() 682 CPPUNIT_ASSERT(subcomm.size() >= 1); in testSplit() 683 CPPUNIT_ASSERT(subcomm.size() >= TestCommWorld->size() / 2); in testSplit() 684 CPPUNIT_ASSERT(subcomm.size() <= TestCommWorld->size() / 2 + 1); in testSplit() 690 Parallel::Communicator subcomm; in testSplitByType() local 698 TestCommWorld->split_by_type(type, rank, i, subcomm); in testSplitByType() 700 CPPUNIT_ASSERT(subcomm.size() >= 1); in testSplitByType() 701 CPPUNIT_ASSERT(subcomm.size() <= TestCommWorld->size()); in testSplitByType()
|
/dports/math/tblis/tblis-1.2.0/src/internal/3t/dense/ |
H A D | mult.cxx | 75 auto subcomm = comm.gang(TCI_EVENLY, nt_l); in mult_blis() local 77 subcomm.distribute_over_gangs(m2, in mult_blis() 103 mult(subcomm, cfg, m, n, in mult_blis() 164 auto subcomm = comm.gang(TCI_EVENLY, nt_l); in mult_blis() local 166 subcomm.distribute_over_gangs(m2*n2, in mult_blis() 188 mult(subcomm, cfg, m, n, in mult_blis() 325 subcomm.distribute_over_gangs(l*m2, in mult_blis() 356 mult(subcomm, cfg, m, n, in mult_blis() 425 subcomm.distribute_over_gangs(l*m2*n2, in mult_blis() 451 mult(subcomm, cfg, m, n, in mult_blis() [all …]
|
/dports/science/octopus/octopus-10.5/src/poisson/ |
H A D | poisson_fmm.F90 | 99 integer :: subcomm, cdim local 199 subcomm = this%all_nodes_grp%comm 210 call MPI_Cart_sub(this%all_nodes_grp%comm, remains(1), subcomm, mpi_err) 211 call mpi_grp_init(this%perp_grp, subcomm) 239 ret = fcs_init(this%handle, trim(adjustl(method)) // c_null_char, subcomm)
|
/dports/math/libmesh/libmesh-1.6.2/contrib/timpi/test/ |
H A D | parallel_unit.C | 825 TIMPI::Communicator subcomm; in testSplit() local 828 TestCommWorld->split(color, rank, subcomm); in testSplit() 830 TIMPI_UNIT_ASSERT(subcomm.size() >= 1); in testSplit() 831 TIMPI_UNIT_ASSERT(subcomm.size() >= TestCommWorld->size() / 2); in testSplit() 832 TIMPI_UNIT_ASSERT(subcomm.size() <= TestCommWorld->size() / 2 + 1); in testSplit() 838 TIMPI::Communicator subcomm; in testSplitByType() local 846 TestCommWorld->split_by_type(type, rank, i, subcomm); in testSplitByType() 848 TIMPI_UNIT_ASSERT(subcomm.size() >= 1); in testSplitByType() 849 TIMPI_UNIT_ASSERT(subcomm.size() <= TestCommWorld->size()); in testSplitByType()
|
/dports/math/tblis/tblis-1.2.0/src/internal/1t/indexed_dpd/ |
H A D | add.cxx | 100 [&,idx_A,idx_B,block_AB,next_A](const communicator& subcomm) in trace_block() 152 add(subcomm, cfg, len_A, {}, len_AB, in trace_block() 222 (const communicator& subcomm) in replicate_block() 261 add(subcomm, cfg, {}, len_B, len_AB, in replicate_block() 318 [&,factor,idx_A,idx_B,block_AB](const communicator& subcomm) in transpose_block() 345 add(subcomm, cfg, {}, {}, len_AB, in transpose_block()
|