Home
last modified time | relevance | path

Searched refs:MPI_INT (Results 1 – 25 of 84) sorted by relevance

1234

/petsc/src/sys/utils/
H A Dmpiu.c17 if (rank) PetscCallMPI(MPI_Recv(NULL, 0, MPI_INT, rank - 1, tag, comm, &status)); in PetscSequentialPhaseBegin_Private()
19 …if ((rank % ng) < ng - 1 && rank != size - 1) PetscCallMPI(MPI_Send(NULL, 0, MPI_INT, rank + 1, ta… in PetscSequentialPhaseBegin_Private()
34 …if ((rank % ng) == ng - 1 || rank == size - 1) PetscCallMPI(MPI_Send(NULL, 0, MPI_INT, (rank + 1) … in PetscSequentialPhaseEnd_Private()
35 if (rank == 0) PetscCallMPI(MPI_Recv(NULL, 0, MPI_INT, size - 1, tag, comm, &status)); in PetscSequentialPhaseEnd_Private()
H A Dmpimesg.c54 PetscCallMPI(MPIU_Allreduce(iflags_local, recv_buf, size, MPI_INT, MPI_SUM, comm)); in PetscGatherNumberOfMessages()
107 …for (i = 0; i < nrecvs; i++) PetscCallMPI(MPIU_Irecv((*olengths) + i, 1, MPI_INT, MPI_ANY_SOURCE, … in PetscGatherMessageLengths()
112 PetscCallMPI(MPIU_Isend((void *)(ilengths + i), 1, MPI_INT, i, tag, comm, s_waits + j)); in PetscGatherMessageLengths()
160 PetscCallMPI(MPIU_Allreduce(iflags_local, recv_buf, size, MPI_INT, MPI_SUM, comm)); in PetscGatherNumberOfMessages_Private()
258 PetscCallMPI(MPIU_Irecv(buf_j, 2, MPI_INT, MPI_ANY_SOURCE, tag, comm, r_waits + i)); in PetscGatherMessageLengths2()
267 PetscCallMPI(MPIU_Isend(buf_j, 2, MPI_INT, i, tag, comm, s_waits + j)); in PetscGatherMessageLengths2()
H A Dpsplit.c78 PetscCallMPI(MPIU_Allreduce(l, g, 2, MPI_INT, MPI_SUM, comm)); in PetscSplitOwnership()
134 PetscCallMPI(MPIU_Allreduce(l, g, 2, MPI_INT, MPI_SUM, comm)); in PetscSplitOwnershipEqual()
H A Dpdisplay.c93 PetscCallMPI(MPIU_Allreduce(&localmatch, &allmatch, 1, MPI_INT, MPI_LAND, PETSC_COMM_WORLD)); in PetscWorldIsSingleHost()
/petsc/src/vec/is/sf/tests/
H A Dex1.c162 PetscCall(PetscSFBcastBegin(sf, MPI_INT, NULL, NULL, MPI_REPLACE)); in main()
163 PetscCall(PetscSFBcastEnd(sf, MPI_INT, NULL, NULL, MPI_REPLACE)); in main()
173 PetscCall(PetscSFReduceBegin(sf, MPI_INT, NULL, NULL, MPI_REPLACE)); in main()
174 PetscCall(PetscSFReduceEnd(sf, MPI_INT, NULL, NULL, MPI_REPLACE)); in main()
175 PetscCall(PetscSFReduceBegin(sf, MPI_INT, NULL, NULL, MPI_SUM)); in main()
176 PetscCall(PetscSFReduceEnd(sf, MPI_INT, NULL, NULL, MPI_SUM)); in main()
184 PetscCall(PetscSFFetchAndOpBegin(sf, MPI_INT, NULL, NULL, NULL, MPI_SUM)); in main()
185 PetscCall(PetscSFFetchAndOpEnd(sf, MPI_INT, NULL, NULL, NULL, MPI_SUM)); in main()
H A Dex25.c44 PetscCallMPI(MPI_Type_contiguous_c(m, MPI_INT, &newtype)); in main()
46 PetscCallMPI(MPI_Type_contiguous(m, MPI_INT, &newtype)); in main()
/petsc/src/sys/error/
H A Derr.c500 PetscCallMPI(MPI_Gather(&NN, 1, MPI_INT, NULL, 0, MPI_INT, 0, comm)); in PetscIntViewNumColumns()
504 PetscCallMPI(MPI_Gather(&NN, 1, MPI_INT, sizes, 1, MPI_INT, 0, comm)); in PetscIntViewNumColumns()
608 PetscCallMPI(MPI_Gather(&NN, 1, MPI_INT, NULL, 0, MPI_INT, 0, comm)); in PetscRealViewNumColumns()
612 PetscCallMPI(MPI_Gather(&NN, 1, MPI_INT, sizes, 1, MPI_INT, 0, comm)); in PetscRealViewNumColumns()
720 PetscCallMPI(MPI_Gather(&NN, 1, MPI_INT, NULL, 0, MPI_INT, 0, comm)); in PetscScalarViewNumColumns()
724 PetscCallMPI(MPI_Gather(&NN, 1, MPI_INT, sizes, 1, MPI_INT, 0, comm)); in PetscScalarViewNumColumns()
/petsc/src/sys/objects/
H A Dptype.c40 else if (ptype == PETSC_ENUM) *mtype = MPI_INT; in PetscDataTypeToMPIDataType()
41 else if (ptype == PETSC_BOOL) *mtype = MPI_INT; in PetscDataTypeToMPIDataType()
77 else if (mtype == MPI_INT) *ptype = PETSC_ENUM; in PetscMPIDataTypeToPetscDataType()
H A Dsubcomm.c217 PetscCallMPI(MPI_Allgather(sendbuf, 2, MPI_INT, recvbuf, 2, MPI_INT, comm)); in PetscSubcommSetTypeGeneral()
/petsc/src/sys/fileio/
H A Dmprint.c594 PetscCallMPI(MPI_Send(&dummy, 1, MPI_INT, i, tag, comm)); in PetscSynchronizedFlush()
595 PetscCallMPI(MPI_Recv(&n, 1, MPI_INT, i, tag, comm, &status)); in PetscSynchronizedFlush()
599 PetscCallMPI(MPI_Recv(&size, 1, MPI_INT, i, tag, comm, &status)); in PetscSynchronizedFlush()
609 PetscCallMPI(MPI_Recv(&dummy, 1, MPI_INT, 0, tag, comm, &status)); in PetscSynchronizedFlush()
610 PetscCallMPI(MPI_Send(&petsc_printfqueuelength, 1, MPI_INT, 0, tag, comm)); in PetscSynchronizedFlush()
612 PetscCallMPI(MPI_Send(&next->size, 1, MPI_INT, 0, tag, comm)); in PetscSynchronizedFlush()
H A Dfretrieve.c187 PetscCallMPI(MPIU_Allreduce(&cnt, &sum, 1, MPI_INT, MPI_SUM, comm)); in PetscSharedTmp()
300 PetscCallMPI(MPIU_Allreduce(&cnt, &sum, 1, MPI_INT, MPI_SUM, comm)); in PetscSharedWorkingDirectory()
/petsc/src/dm/impls/da/ftn-custom/
H A Dzdaf.c27 *ierr = F90Array1dCreate((PetscInt *)r, MPI_INT, 1, n, ptr PETSC_F90_2PTR_PARAM(ptrd)); in dmdagetneighbors_()
32 *ierr = F90Array1dDestroy(ptr, MPI_INT PETSC_F90_2PTR_PARAM(ptrd)); in dmdarestoreneighbors_()
/petsc/src/mat/impls/aij/mpi/
H A Dmpb_aij.c30 PetscCallMPI(MPI_Allgather(&commRank, 1, MPI_INT, commRankMap, 1, MPI_INT, subComm)); in MatGetMultiProcBlock_MPIAIJ()
/petsc/src/dm/partitioner/impls/multistage/
H A Dmspart.c94 …PetscCallMPI(MPIU_Allreduce(MPI_IN_PLACE, &gem, 1, MPI_INT, MPI_SUM, PetscObjectComm((PetscObject)… in PetscPartitionerView_Multistage()
221 PetscCallMPI(MPI_Allgather(granks, 1, MPI_INT, ranks, 1, MPI_INT, comm)); in PetscPartitionerMultistage_CreateStages()
393 PetscCallMPI(MPIU_Allreduce(b, b2, 2, MPI_INT, MPI_MAX, comm)); in PetscPartitionerMultistageSetStages()
401 PetscCallMPI(MPI_Allgather(ranks, gsize, MPI_INT, granks, gsize, MPI_INT, comm)); in PetscPartitionerMultistageSetStages()
438 PetscCallMPI(MPI_Allgather(&gsize, 1, MPI_INT, lparts[l], 1, MPI_INT, comm)); in PetscPartitionerMultistageSetStages()
501 PetscCallMPI(MPIU_Allreduce(&isize, &isizer, 1, MPI_INT, MPI_MAX, pcomm)); in PetscPartitionerMultistageSetStages()
/petsc/src/mat/impls/baij/mpi/
H A Dmpb_baij.c35 PetscCallMPI(MPI_Allgather(&commRank, 1, MPI_INT, commRankMap, 1, MPI_INT, subComm)); in MatGetMultiProcBlock_MPIBAIJ()
/petsc/src/sys/ftn-custom/
H A Df90_cwrap.c31 if (ftype == MPI_INTEGER || ftype == MPI_INT) *dtype = MPI_INT; in PetscMPIFortranDatatypeToC()
133 } else if (type == MPI_INT) { in F90Array1dCreate()
179 } else if (type == MPI_INT) { in F90Array1dAccess()
218 } else if (type == MPI_INT) { in F90Array1dDestroy()
/petsc/src/mat/impls/adj/mpi/
H A Dmpiadj.c811 PetscCallMPI(MPI_Allgather(&mnz, 1, MPI_INT, allnz, 1, MPI_INT, PetscObjectComm((PetscObject)A))); in MatMPIAdjToSeq_MPIAdj()
828 PetscCallMPI(MPI_Allgather(&mm, 1, MPI_INT, allm, 1, MPI_INT, PetscObjectComm((PetscObject)A))); in MatMPIAdjToSeq_MPIAdj()
858 PetscCallMPI(MPI_Gather(&mnz, 1, MPI_INT, allnz, 1, MPI_INT, 0, PetscObjectComm((PetscObject)A))); in MatMPIAdjToSeqRankZero_MPIAdj()
881 PetscCallMPI(MPI_Gather(&mm, 1, MPI_INT, allm, 1, MPI_INT, 0, PetscObjectComm((PetscObject)A))); in MatMPIAdjToSeqRankZero_MPIAdj()
888 PetscCallMPI(MPI_Gather(&mm, 1, MPI_INT, NULL, 1, MPI_INT, 0, PetscObjectComm((PetscObject)A))); in MatMPIAdjToSeqRankZero_MPIAdj()
/petsc/include/petsc/private/
H A Dviewercgnsimpl.h83 #define MPIU_CGSIZE MPI_INT
/petsc/src/vec/is/is/utils/
H A Discoloring.c540 PetscCallMPI(MPIU_Allreduce(&np, &npt, 1, MPI_INT, MPI_MAX, comm)); in ISPartitioningToNumbering()
693 PetscCallMPI(MPI_Allgather(&nn, 1, MPI_INT, sizes, 1, MPI_INT, comm)); in ISAllGather()
745 PetscCallMPI(MPI_Allgather(&nn, 1, MPI_INT, sizes, 1, MPI_INT, comm)); in ISAllGatherColors()
/petsc/src/vec/is/sf/interface/ftn-custom/
H A Dzsf.c63 …*ierr = F90Array1dCreate((void *)iranks, MPI_INT, 1, *niranks, airanks PETSC_F90_2PTR_PARAM(pirank… in petscsfgetleafranks_()
78 *ierr = F90Array1dCreate((void *)ranks, MPI_INT, 1, *nranks, aranks PETSC_F90_2PTR_PARAM(pranks)); in petscsfgetrootranks_()
/petsc/src/sys/objects/ftn-custom/
H A Dzstart.c81 PetscCallMPI(MPI_Bcast(argc, 1, MPI_INT, 0, PETSC_COMM_WORLD)); in PETScParseFortranArgs_Private()
/petsc/src/sys/classes/viewer/impls/ascii/
H A Dfilev.c648 PetscCallMPI(MPI_Bcast(&PETSC_VIEWER_ASCII_STDOUT_fileunit, 1, MPI_INT, 0, comm)); in PetscViewerASCIIGetStdout()
981 PetscCallMPI(MPI_Send(&dummy, 1, MPI_INT, i, tag, comm)); in PetscViewerFlush_ASCII()
982 PetscCallMPI(MPI_Recv(&n, 1, MPI_INT, i, tag, comm, &status)); in PetscViewerFlush_ASCII()
1001 PetscCallMPI(MPI_Recv(&dummy, 1, MPI_INT, 0, tag, comm, &status)); in PetscViewerFlush_ASCII()
1002 PetscCallMPI(MPI_Send(&vascii->petsc_printfqueuelength, 1, MPI_INT, 0, tag, comm)); in PetscViewerFlush_ASCII()
/petsc/src/vec/is/ao/impls/mapping/
H A Daomapping.c264 PetscCallMPI(MPI_Allgather(&nnapp, 1, MPI_INT, lens, 1, MPI_INT, comm)); in AOCreateMapping()
/petsc/src/vec/vec/impls/shared/
H A Dshvec.c114 PetscCallMPI(MPI_Scan(&llen, &shift, 1, MPI_INT, MPI_SUM, comm)); in PetscSharedMalloc()
/petsc/src/dm/impls/plex/
H A Dplexvtu.c519 … PetscCall(TransferWrite(comm, viewer, fp, r, 0, connectivity, buffer, piece.nconn, MPI_INT, tag)); in DMPlexVTKWriteAll_VTU()
520 … PetscCall(TransferWrite(comm, viewer, fp, r, 0, offsets, buffer, piece.ncells, MPI_INT, tag)); in DMPlexVTKWriteAll_VTU()
531 … PetscCall(TransferWrite(comm, viewer, fp, r, 0, owners, buffer, piece.ncells, MPI_INT, tag)); in DMPlexVTKWriteAll_VTU()
759 …PetscCall(TransferWrite(comm, viewer, fp, r, 0, NULL, buffer, gpiece[r].nconn, MPI_INT, tag)); … in DMPlexVTKWriteAll_VTU()
760 …PetscCall(TransferWrite(comm, viewer, fp, r, 0, NULL, buffer, gpiece[r].ncells, MPI_INT, tag)); … in DMPlexVTKWriteAll_VTU()
762 …PetscCall(TransferWrite(comm, viewer, fp, r, 0, NULL, buffer, gpiece[r].ncells, MPI_INT, tag)); … in DMPlexVTKWriteAll_VTU()

1234