Home
last modified time | relevance | path

Searched refs:req (Results 1 – 16 of 16) sorted by relevance

/petsc/include/petsc/private/
H A Dsfimpl.h154 #define MPIU_Ibcast(a, b, c, d, e, req) MPI_Ibcast(a, b, c, d, e, req) argument
155 #define MPIU_Ireduce(a, b, c, d, e, f, g, req) MPI_Ireduce(a, b, c, d, e, f, g, req) argument
156 #define MPIU_Iscatter(a, b, c, d, e, f, g, h, req) MPI_Iscatter(a, b, c, d, e, f, g, h, req) argument
157 …#define MPIU_Iscatterv(a, b, c, d, e, f, g, h, i, req) MPI_Iscatterv(a, b, c, d, e, f, g, h, i, re… argument
158 #define MPIU_Igather(a, b, c, d, e, f, g, h, req) MPI_Igather(a, b, c, d, e, f, g, h, req) argument
159 …#define MPIU_Igatherv(a, b, c, d, e, f, g, h, i, req) MPI_Igatherv(a, b, c, d, e, f, g, h, i, req) argument
160 #define MPIU_Iallgather(a, b, c, d, e, f, g, req) MPI_Iallgather(a, b, c, d, e, f, g, req) argument
161 …#define MPIU_Iallgatherv(a, b, c, d, e, f, g, h, req) MPI_Iallgatherv(a, b, c, d, e, f, g, h, req) argument
162 #define MPIU_Ialltoall(a, b, c, d, e, f, g, req) MPI_Ialltoall(a, b, c, d, e, f, g, req) argument
167 #define MPIU_Ibcast(a, b, c, d, e, req) MPI_Bcast(a, b, c, d, e) argument
[all …]
/petsc/src/vec/is/sf/impls/basic/allgather/
H A Dsfallgather.c37 MPI_Request *req = NULL; in PetscSFBcastBegin_Allgather() local
45 …(PetscSFLinkGetMPIBuffersAndRequests(sf, link, PETSCSF_ROOT2LEAF, &rootbuf, &leafbuf, &req, NULL)); in PetscSFBcastBegin_Allgather()
47 PetscCallMPI(MPIU_Iallgather(rootbuf, sendcount, unit, leafbuf, sendcount, unit, comm, req)); in PetscSFBcastBegin_Allgather()
59 MPI_Request *req = NULL; in PetscSFReduceBegin_Allgather() local
73 …(PetscSFLinkGetMPIBuffersAndRequests(sf, link, PETSCSF_LEAF2ROOT, &rootbuf, &leafbuf, &req, NULL)); in PetscSFReduceBegin_Allgather()
80 …REMOTE][link->leafmtype_mpi], recvcount, unit, rootbuf, recvcount, unit, 0 /*rank 0*/, comm, req)); in PetscSFReduceBegin_Allgather()
92 MPI_Request *req = NULL; in PetscSFBcastToZero_Allgather() local
100 …(PetscSFLinkGetMPIBuffersAndRequests(sf, link, PETSCSF_ROOT2LEAF, &rootbuf, &leafbuf, &req, NULL)); in PetscSFBcastToZero_Allgather()
102 …buf ? MPI_IN_PLACE : rootbuf, sendcount, unit, leafbuf, sendcount, unit, 0 /*rank 0*/, comm, req)); in PetscSFBcastToZero_Allgather()
/petsc/src/vec/is/sf/impls/basic/gather/
H A Dsfgather.c11 MPI_Request *req = NULL; in PetscSFLinkStartCommunication_Gather() local
23 …PetscCall(PetscSFLinkGetMPIBuffersAndRequests(sf, link, direction, &rootbuf, &leafbuf, &req, NULL)… in PetscSFLinkStartCommunication_Gather()
27 … == leafbuf ? MPI_IN_PLACE : rootbuf, count, unit, leafbuf, count, unit, 0 /*rank 0*/, comm, req)); in PetscSFLinkStartCommunication_Gather()
29 …, count, unit, rootbuf == leafbuf ? MPI_IN_PLACE : rootbuf, count, unit, 0 /*rank 0*/, comm, req)); in PetscSFLinkStartCommunication_Gather()
/petsc/src/vec/is/sf/impls/basic/gatherv/
H A Dsfgatherv.c14 MPI_Request *req = NULL; in PetscSFLinkStartCommunication_Gatherv() local
25 …PetscCall(PetscSFLinkGetMPIBuffersAndRequests(sf, link, direction, &rootbuf, &leafbuf, &req, NULL)… in PetscSFLinkStartCommunication_Gatherv()
29 …therv(rootbuf, count, unit, leafbuf, dat->recvcounts, dat->displs, unit, 0 /*rank 0*/, comm, req)); in PetscSFLinkStartCommunication_Gatherv()
31 …I(MPIU_Iscatterv(leafbuf, dat->recvcounts, dat->displs, unit, rootbuf, count, unit, 0, comm, req)); in PetscSFLinkStartCommunication_Gatherv()
/petsc/src/sys/tests/
H A Dex8.c50 …scMPIInt tag[], PetscMPIInt tonum, PetscMPIInt rank, void *todata, MPI_Request req[], PetscCtx ctx) in FSend() argument
57 PetscCallMPI(MPI_Isend(&fctx->todata[tonum].rank, 1, MPIU_INT, rank, tag[0], comm, &req[0])); in FSend()
58 PetscCallMPI(MPI_Isend(&fctx->todata[tonum].value, 1, MPIU_SCALAR, rank, tag[1], comm, &req[1])); in FSend()
62 …m comm, const PetscMPIInt tag[], PetscMPIInt rank, void *fromdata, MPI_Request req[], PetscCtx ctx) in FRecv() argument
70 PetscCallMPI(MPI_Irecv(&buf->rank, 1, MPIU_INT, rank, tag[0], comm, &req[0])); in FRecv()
71 PetscCallMPI(MPI_Irecv(&buf->value, 1, MPIU_SCALAR, rank, tag[1], comm, &req[1])); in FRecv()
/petsc/src/vec/is/sf/impls/basic/neighbor/
H A Dsfneighbor.c92 MPI_Request *req = NULL; in PetscSFLinkStartCommunication_Neighbor() local
102 …PetscCall(PetscSFLinkGetMPIBuffersAndRequests(sf, link, direction, &rootbuf, &leafbuf, &req, NULL)… in PetscSFLinkStartCommunication_Neighbor()
107 …at->rootdispls, link->unit, leafbuf, dat->leafcounts, dat->leafdispls, link->unit, distcomm, req)); in PetscSFLinkStartCommunication_Neighbor()
110 …at->leafdispls, link->unit, rootbuf, dat->rootcounts, dat->rootdispls, link->unit, distcomm, req)); in PetscSFLinkStartCommunication_Neighbor()
124 MPI_Request *req = link->rootreqs[direction][rootmtype_mpi][rootdirect_mpi]; in PetscSFLinkInitMPIRequests_Persistent_Neighbor() local
134 …otdispls, link->unit, leafbuf, dat->leafcounts, dat->leafdispls, link->unit, distcomm, info, req)); in PetscSFLinkInitMPIRequests_Persistent_Neighbor()
136 …afdispls, link->unit, rootbuf, dat->rootcounts, dat->rootdispls, link->unit, distcomm, info, req)); in PetscSFLinkInitMPIRequests_Persistent_Neighbor()
149 MPI_Request *req = NULL; in PetscSFLinkStartCommunication_Persistent_Neighbor() local
158 PetscCall(PetscSFLinkGetMPIBuffersAndRequests(sf, link, direction, NULL, NULL, &req, NULL)); in PetscSFLinkStartCommunication_Persistent_Neighbor()
161 PetscCallMPI(MPI_Start(req)); in PetscSFLinkStartCommunication_Persistent_Neighbor()
/petsc/src/vec/is/sf/impls/basic/allgatherv/
H A Dsfallgatherv.c108 MPI_Request *req = NULL; in PetscSFBcastBegin_Allgatherv() local
118 …(PetscSFLinkGetMPIBuffersAndRequests(sf, link, PETSCSF_ROOT2LEAF, &rootbuf, &leafbuf, &req, NULL)); in PetscSFBcastBegin_Allgatherv()
124 …(dat->bcast_pattern) PetscCallMPI(MPIU_Ibcast(leafbuf, nleaves, unit, dat->bcast_root, comm, req)); in PetscSFBcastBegin_Allgatherv()
125 …PIU_Iallgatherv(rootbuf, sendcount, unit, leafbuf, dat->recvcounts, dat->displs, unit, comm, req)); in PetscSFBcastBegin_Allgatherv()
137 MPI_Request *req = NULL; in PetscSFReduceBegin_Allgatherv() local
150 …(PetscSFLinkGetMPIBuffersAndRequests(sf, link, PETSCSF_LEAF2ROOT, &rootbuf, &leafbuf, &req, NULL)); in PetscSFReduceBegin_Allgatherv()
175 *req = MPI_REQUEST_NULL; /* Set NULL so that we can safely MPI_Wait(req) */ in PetscSFReduceBegin_Allgatherv()
178 … PetscCallMPI(MPIU_Ireduce(leafbuf, rootbuf, nleavesi, baseunit, op, dat->bcast_root, comm, req)); in PetscSFReduceBegin_Allgatherv()
189 …link->leafmtype_mpi], dat->recvcounts, dat->displs, unit, rootbuf, recvcount, unit, 0, comm, req)); in PetscSFReduceBegin_Allgatherv()
223 MPI_Request *req = NULL; in PetscSFBcastToZero_Allgatherv() local
[all …]
/petsc/src/ksp/ksp/impls/cg/pipelcg/
H A Dpipelcg.c12 #define req(i) (plcg->req[i]) macro
26 MPI_Request *req; /* request array for asynchronous global collective */ member
180 PetscCallMPI(MPI_Wait(&req(0), MPI_STATUS_IGNORE)); in KSPSolve_InnerLoop_PIPELCG()
190 PetscCallMPI(MPI_Wait(&req(it - l + 1), MPI_STATUS_IGNORE)); in KSPSolve_InnerLoop_PIPELCG()
217 for (i = start; i < end; ++i) PetscCallMPI(MPI_Wait(&req(i), MPI_STATUS_IGNORE)); in KSPSolve_InnerLoop_PIPELCG()
295 …MPI(MPIU_Iallreduce(MPI_IN_PLACE, &G(0, it + 1), mpin, MPIU_SCALAR, MPIU_SUM, comm, &req(it + 1))); in KSPSolve_InnerLoop_PIPELCG()
302 …Iallreduce(MPI_IN_PLACE, &G(it - l + 1, it + 1), mpin, MPIU_SCALAR, MPIU_SUM, comm, &req(it + 1))); in KSPSolve_InnerLoop_PIPELCG()
336 for (i = start; i < end; ++i) PetscCallMPI(MPI_Wait(&req(i), MPI_STATUS_IGNORE)); in KSPSolve_InnerLoop_PIPELCG()
388 PetscCall(PetscCalloc1(max_it + 1, &plcg->req)); in KSPSolve_PIPELCG()
414 PetscCallMPI(MPIU_Iallreduce(MPI_IN_PLACE, &G(0, 0), 1, MPIU_SCALAR, MPIU_SUM, comm, &req(0))); in KSPSolve_PIPELCG()
[all …]
/petsc/src/sys/tutorials/
H A Dex3f90.F9041 MPIU_Request req
90 …Len, MPIU_REAL, mod(rank + 1_PETSC_MPIINT_KIND, size), tagMsg + rank, PETSC_COMM_WORLD, req, ierr))
92 PetscCallMPIA(MPI_Wait(req, MPI_STATUS_IGNORE, ierr))
200 …Len, MPIU_REAL, mod(rank + 1_PETSC_MPIINT_KIND, size), tagMsg + rank, PETSC_COMM_WORLD, req, ierr))
202 PetscCallMPIA(MPI_Wait(req, MPI_STATUS_IGNORE, ierr))
/petsc/src/vec/is/sf/impls/basic/alltoall/
H A Dsfalltoall.c12 MPI_Request *req = NULL; in PetscSFLinkStartCommunication_Alltoall() local
22 …PetscCall(PetscSFLinkGetMPIBuffersAndRequests(sf, link, direction, &rootbuf, &leafbuf, &req, NULL)… in PetscSFLinkStartCommunication_Alltoall()
26 PetscCallMPI(MPIU_Ialltoall(rootbuf, 1, unit, leafbuf, 1, unit, comm, req)); in PetscSFLinkStartCommunication_Alltoall()
28 PetscCallMPI(MPIU_Ialltoall(leafbuf, 1, unit, rootbuf, 1, unit, comm, req)); in PetscSFLinkStartCommunication_Alltoall()
/petsc/src/vec/vec/impls/mpi/
H A Dpbvec.c142 …PetscMPIInt tag[], PetscMPIInt rankid, PetscMPIInt rank, void *sdata, MPI_Request req[], void *ctx) in VecAssemblySend_MPI_Private() argument
155 …scCallMPI(MPIU_Isend(x->sendptrs[rankid].ints, hdr->count, MPIU_INT, rank, tag[0], comm, &req[0])); in VecAssemblySend_MPI_Private()
156 …MPI(MPIU_Isend(x->sendptrs[rankid].scalars, hdr->count, MPIU_SCALAR, rank, tag[1], comm, &req[1])); in VecAssemblySend_MPI_Private()
159 …cCallMPI(MPIU_Isend(x->sendptrs[rankid].intb, hdr->bcount, MPIU_INT, rank, tag[2], comm, &req[2])); in VecAssemblySend_MPI_Private()
160 …IU_Isend(x->sendptrs[rankid].scalarb, hdr->bcount * bs, MPIU_SCALAR, rank, tag[3], comm, &req[3])); in VecAssemblySend_MPI_Private()
165 …PI_Comm comm, const PetscMPIInt tag[], PetscMPIInt rank, void *rdata, MPI_Request req[], void *ctx) in VecAssemblyRecv_MPI_Private() argument
178 PetscCallMPI(MPIU_Irecv(frame->ints, hdr->count, MPIU_INT, rank, tag[0], comm, &req[0])); in VecAssemblyRecv_MPI_Private()
180 PetscCallMPI(MPIU_Irecv(frame->scalars, hdr->count, MPIU_SCALAR, rank, tag[1], comm, &req[1])); in VecAssemblyRecv_MPI_Private()
190 PetscCallMPI(MPIU_Irecv(frame->intb, hdr->bcount, MPIU_INT, rank, tag[2], comm, &req[2])); in VecAssemblyRecv_MPI_Private()
192 …PetscCallMPI(MPIU_Irecv(frame->scalarb, hdr->bcount * bs, MPIU_SCALAR, rank, tag[3], comm, &req[3]… in VecAssemblyRecv_MPI_Private()
/petsc/config/BuildSystem/
H A Dretrieval.py198 req = Request(url)
199 req.headers['User-Agent'] = 'PetscConfigure/'+self.ver
201 f.write(urlopen(req).read())
/petsc/src/sys/utils/
H A Dmpits.c333 MPI_Request *sendreqs, *usendreqs, *req, barrier; in PetscCommBuildTwoSidedFReq_Ibarrier() local
376 PetscCall(PetscSegBufferGet(segreq, ntags, &req)); in PetscCommBuildTwoSidedFReq_Ibarrier()
377 for (k = 0; k < ntags; k++) req[k] = MPI_REQUEST_NULL; in PetscCommBuildTwoSidedFReq_Ibarrier()
378 PetscCall((*recv)(comm, tags, status.MPI_SOURCE, buf, req, ctx)); in PetscCommBuildTwoSidedFReq_Ibarrier()
/petsc/src/ksp/ksp/impls/cg/pipecg2/
H A Dpipecg2.c559 MPI_Request req; in KSPSolve_PIPECG2() local
615 PetscCallMPI(MPI_Iallreduce(MPI_IN_PLACE, &lambda[10], 3, MPIU_SCALAR, MPIU_SUM, pcomm, &req)); in KSPSolve_PIPECG2()
618 req = MPI_REQUEST_NULL; in KSPSolve_PIPECG2()
630 PetscCallMPI(MPI_Wait(&req, &stat)); in KSPSolve_PIPECG2()
683 PetscCallMPI(MPI_Iallreduce(MPI_IN_PLACE, lambda, 15, MPIU_SCALAR, MPIU_SUM, pcomm, &req)); in KSPSolve_PIPECG2()
686 req = MPI_REQUEST_NULL; in KSPSolve_PIPECG2()
695 PetscCallMPI(MPI_Wait(&req, &stat)); in KSPSolve_PIPECG2()
/petsc/src/mat/utils/
H A Dmatstash.c786 …scMPIInt tag[], PetscMPIInt rankid, PetscMPIInt rank, void *sdata, MPI_Request req[], PetscCtx ctx) in MatStashBTSSend_Private() argument
793 …send(stash->sendframes[rankid].buffer, hdr->count, stash->blocktype, rank, tag[0], comm, &req[0])); in MatStashBTSSend_Private()
803 …Comm comm, const PetscMPIInt tag[], PetscMPIInt rank, void *rdata, MPI_Request req[], PetscCtx ctx) in MatStashBTSRecv_Private() argument
812 …PetscCallMPI(MPIU_Irecv(frame->buffer, hdr->count, stash->blocktype, rank, tag[0], comm, &req[0])); in MatStashBTSRecv_Private()
/petsc/config/BuildSystem/config/packages/
H A DMPI.py498 MPI_Request req; \n\
501 … if (MPI_Ineighbor_alltoallv(0,0,0,MPI_INT,0,0,0,MPI_INT,distcomm,&req)) { }\n'):