| /petsc/src/sys/utils/ |
| H A D | mpimesg.c | 123 (*onodes)[i] = w_status[i].MPI_SOURCE; in PetscGatherMessageLengths() 131 if (w_status[i].MPI_SOURCE == rank) (*olengths)[i] = ilengths[rank]; in PetscGatherMessageLengths() 201 (*onodes)[i] = w_status[i].MPI_SOURCE; in PetscGatherMessageLengths_Private() 202 …if (w_status[i].MPI_SOURCE == rank) (*olengths)[i] = ilengths[rank]; /* See comments in PetscGathe… in PetscGatherMessageLengths_Private() 279 (*onodes)[i] = w_status[i].MPI_SOURCE; in PetscGatherMessageLengths2()
|
| H A D | mpits.c | 106 *recvrank = status.MPI_SOURCE; in PetscCommBuildTwoSided_Ibarrier() 107 PetscCallMPI(MPI_Recv(buf, count, dtype, status.MPI_SOURCE, tag, comm, MPI_STATUS_IGNORE)); in PetscCommBuildTwoSided_Ibarrier() 168 for (i = 0; i < nrecvs; i++) franks[i] = statuses[i].MPI_SOURCE; in PetscCommBuildTwoSided_Allreduce() 212 for (i = 0; i < nrecvs; i++) franks[i] = statuses[i].MPI_SOURCE; in PetscCommBuildTwoSided_RedScatter() 374 *recvrank = status.MPI_SOURCE; in PetscCommBuildTwoSidedFReq_Ibarrier() 375 PetscCallMPI(MPI_Recv(buf, count, dtype, status.MPI_SOURCE, tag, comm, MPI_STATUS_IGNORE)); in PetscCommBuildTwoSidedFReq_Ibarrier() 378 PetscCall((*recv)(comm, tags, status.MPI_SOURCE, buf, req, ctx)); in PetscCommBuildTwoSidedFReq_Ibarrier()
|
| /petsc/include/petsc/mpiuni/ |
| H A D | mpif.h | 46 MPIUNI_FInt MPI_SOURCE,MPI_TAG,MPI_ERROR 47 PARAMETER(MPI_SOURCE=1,MPI_TAG=2,MPI_ERROR=3)
|
| H A D | mpi.h | 199 int MPI_SOURCE, MPI_TAG, MPI_ERROR; member 434 …count), MPIUNI_ARG(array_of_requests), MPIUNI_ARG(status), (*(status)).MPI_SOURCE = 0, *(index) = …
|
| /petsc/src/vec/vec/utils/ |
| H A D | vecstash.c | 352 flg_v[2 * recv_status.MPI_SOURCE + 1] = i / 2; in VecStashScatterGetMesg_Private() 355 flg_v[2 * recv_status.MPI_SOURCE] = i / 2; in VecStashScatterGetMesg_Private() 360 i1 = flg_v[2 * recv_status.MPI_SOURCE]; in VecStashScatterGetMesg_Private() 361 i2 = flg_v[2 * recv_status.MPI_SOURCE + 1]; in VecStashScatterGetMesg_Private()
|
| /petsc/src/tao/leastsquares/tutorials/ |
| H A D | chwirut2.c | 134 …PetscCallMPI(MPI_Send(x, NPARAMETERS, MPIU_REAL, status.MPI_SOURCE, (PetscMPIInt)next_task, PETSC_… in EvaluateFunction() 139 … PetscCallMPI(MPI_Send(x, NPARAMETERS, MPIU_REAL, status.MPI_SOURCE, IDLE_TAG, PETSC_COMM_WORLD)); in EvaluateFunction() 652 PetscCallMPI(MPI_Send(x, NPARAMETERS, MPIU_REAL, status.MPI_SOURCE, DIE_TAG, PETSC_COMM_WORLD)); in StopWorkers()
|
| H A D | chwirut2f.F90 | 308 source = status%MPI_SOURCE 310 source = status(MPI_SOURCE) 373 source = status%MPI_SOURCE 376 source = status(MPI_SOURCE)
|
| /petsc/src/mat/utils/ |
| H A D | matstash.c | 649 … PetscCheck(recv_status.MPI_SOURCE >= 0, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Negative MPI source!"); in MatStashScatterGetMesg_Ref() 654 flg_v[2 * recv_status.MPI_SOURCE] = i / 2; in MatStashScatterGetMesg_Ref() 658 flg_v[2 * recv_status.MPI_SOURCE + 1] = i / 2; in MatStashScatterGetMesg_Ref() 663 i1 = flg_v[2 * recv_status.MPI_SOURCE]; in MatStashScatterGetMesg_Ref() 664 i2 = flg_v[2 * recv_status.MPI_SOURCE + 1]; in MatStashScatterGetMesg_Ref()
|
| /petsc/src/vec/is/ao/impls/memscalable/ |
| H A D | aomemscalable.c | 199 source = recv_status.MPI_SOURCE; in AOMap_MemoryScalable_private() 216 source = recv_status.MPI_SOURCE; in AOMap_MemoryScalable_private()
|
| /petsc/src/mat/impls/sbaij/mpi/ |
| H A D | sbaijov.c | 351 proc_id = r_status.MPI_SOURCE; in MatIncreaseOverlap_MPISBAIJ_Once() 404 proc_id = r_status.MPI_SOURCE; in MatIncreaseOverlap_MPISBAIJ_Once()
|
| /petsc/src/vec/is/utils/ |
| H A D | psort.c | 217 sender = status.MPI_SOURCE; in PetscParallelRedistribute()
|
| /petsc/src/mat/impls/dense/mpi/ |
| H A D | mmdense.c | 241 s_proc = r_status1[i].MPI_SOURCE; /* send processor */ in MatCreateSubMatrices_MPIDense_Local()
|
| /petsc/src/mat/graphops/coarsen/impls/hem/ |
| H A D | hem.c | 868 const PetscMPIInt proc = status.MPI_SOURCE; in MatCoarsenApply_HEM_private() 939 proc = status.MPI_SOURCE; in MatCoarsenApply_HEM_private()
|
| /petsc/src/mat/impls/aij/mpi/ |
| H A D | mpiov.c | 720 proc = recv_status[i].MPI_SOURCE; in MatIncreaseOverlap_MPIAIJ_Once() 736 …i) PetscCallMPI(MPIU_Isend(xdata[i], isz1[i], MPIU_INT, recv_status[i].MPI_SOURCE, tag2, comm, s_w… in MatIncreaseOverlap_MPIAIJ_Once() 1398 req_source1[i] = r_status1[i].MPI_SOURCE; in MatCreateSubMatrices_MPIAIJ_SingleIS_Local() 1416 req_source2[i] = r_status2[i].MPI_SOURCE; in MatCreateSubMatrices_MPIAIJ_SingleIS_Local()
|