| /petsc/lib/petsc/bin/ |
| H A D | petscnagfor | 43 status = sub.returncode 52 return (output,error,status) 88 (output,error,status) = runnagfor(['nagfor']+args) 95 (output,error,status) = runnagfor(['nagfor','-dryrun']+args) 96 if status: 99 exit(status) 135 (suboutput,error,status) = runnagfor(js) 148 if status: 151 exit(status) 163 if status and (suboutput+error).find('multiple') > -1: [all …]
|
| H A D | taucc.py | 21 (status, output) = commands.getstatusoutput(cmd) 22 if status: 47 (status, output) = commands.getstatusoutput(cmd) 48 if status:
|
| /petsc/src/dm/tutorials/output/ |
| H A D | swarm_ex1_4.out | 7 topology status: finalized 8 message lengths status: finalized 9 packer status status: finalized 10 communication status: finalized
|
| H A D | swarm_ex1_2.out | 7 topology status: finalized 8 message lengths status: finalized 9 packer status status: finalized 10 communication status: finalized
|
| H A D | swarm_ex1_8.out | 10 topology status: finalized 11 message lengths status: finalized 12 packer status status: finalized 13 communication status: finalized
|
| H A D | swarm_ex1_3.out | 26 topology status: finalized 27 message lengths status: finalized 28 packer status status: finalized 29 communication status: finalized
|
| H A D | swarm_ex1_6.out | 10 topology status: finalized 11 message lengths status: finalized 12 packer status status: finalized 13 communication status: finalized
|
| /petsc/src/ksp/pc/impls/tfs/ |
| H A D | comm.c | 67 MPI_Status status; in PCTFS_giop() local 97 …PI(MPI_Recv(work, n, MPIU_INT, MPI_ANY_SOURCE, MSGTAG0 + edge_not_pow_2, MPI_COMM_WORLD, &status)); in PCTFS_giop() 109 …PetscCallMPI(MPI_Recv(work, n, MPIU_INT, MPI_ANY_SOURCE, MSGTAG2 + dest, MPI_COMM_WORLD, &status)); in PCTFS_giop() 122 …PetscCallMPI(MPI_Recv(vals, n, MPIU_INT, MPI_ANY_SOURCE, MSGTAG4 + dest, MPI_COMM_WORLD, &status)); in PCTFS_giop() 130 …PI(MPI_Recv(vals, n, MPIU_INT, MPI_ANY_SOURCE, MSGTAG5 + edge_not_pow_2, MPI_COMM_WORLD, &status)); in PCTFS_giop() 144 MPI_Status status; in PCTFS_grop() local 173 …MPI_Recv(work, n, MPIU_SCALAR, MPI_ANY_SOURCE, MSGTAG0 + edge_not_pow_2, MPI_COMM_WORLD, &status)); in PCTFS_grop() 185 …scCallMPI(MPI_Recv(work, n, MPIU_SCALAR, MPI_ANY_SOURCE, MSGTAG2 + dest, MPI_COMM_WORLD, &status)); in PCTFS_grop() 198 …scCallMPI(MPI_Recv(vals, n, MPIU_SCALAR, MPI_ANY_SOURCE, MSGTAG4 + dest, MPI_COMM_WORLD, &status)); in PCTFS_grop() 206 …MPI_Recv(vals, n, MPIU_SCALAR, MPI_ANY_SOURCE, MSGTAG5 + edge_not_pow_2, MPI_COMM_WORLD, &status)); in PCTFS_grop() [all …]
|
| /petsc/src/sys/dll/ |
| H A D | demangle.c | 16 int status; in PetscDemangleSymbol() local 66 newname = cxa_demangle(mangledName, PETSC_NULLPTR, PETSC_NULLPTR, &status); in PetscDemangleSymbol() 67 if (status) { in PetscDemangleSymbol() 68 …PetscCheck(status != -1, PETSC_COMM_SELF, PETSC_ERR_MEM, "Failed to allocate memory for symbol %s"… in PetscDemangleSymbol() 69 …PetscCheck(status == -2, PETSC_COMM_SELF, PETSC_ERR_LIB, "Demangling failed for symbol %s", mangle… in PetscDemangleSymbol()
|
| /petsc/config/BuildSystem/config/utilities/ |
| H A D | getResidentSetSize.py | 57 …(output,status) = self.outputRun('''#include <stdio.h>\n#include <ctype.h>\n#include <sys/times.h>… 102 if status > 0: 103 if status == 2: 105 if status == 3: 107 elif status == 0:
|
| H A D | debuggers.py | 69 …(output, error, status) = config.base.Configure.executeShellCommand(self.dbx+' -c conftest -p '+st… 70 if not status: 79 …(output, error, status) = config.base.Configure.executeShellCommand(self.dbx+' -c conftest -a '+st… 80 if not status: 89 …(output, error, status) = config.base.Configure.executeShellCommand(self.dbx+' -c conftest -pid '+… 90 if not status:
|
| /petsc/config/BuildSystem/config/ |
| H A D | programs.py | 32 …(output, error, status) = config.base.Configure.executeShellCommand(self.mkdir+' -p '+conftmpDir, … 33 if not status and os.path.isdir(conftmpDir): 88 …(out,err,status) = Configure.executeShellCommand('%s s/sed/sd/g "%s"'%(sedcmd,sed1), log = self.lo… 91 status = 1 96 if not status: 119 …(out,err,status) = Configure.executeShellCommand('"%s" -w "%s" "%s"' % (self.diff,diff1,diff2), lo… 121 status = 1 124 if status: 125 (buf,err,status) = Configure.executeShellCommand('/bin/rpm -q diffutils', log = self.log)
|
| H A D | setCompilers.py | 11 def noCheck(command, status, output, error): argument 213 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' -V',checkCommand = … 226 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' -v',checkCommand = … 238 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --help | head -n 20… 256 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --version', log = l… 268 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --help | head -n 50… 282 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --version', log = l… 294 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --version', log = l… 307 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --version', log = l… 321 …(output, error, status) = config.base.Configure.executeShellCommand(compiler+' --version', log = l… [all …]
|
| H A D | base.py | 118 def logError(self, component, status, output, error): argument 119 if status: 120 exitstr = ' exit code ' + str(status) 475 def report(command, status, output, error): argument 476 if error or status: 477 self.logError('preprocessor', status, output, error) 531 def report(command, status, output, error): argument 532 if error or status: 533 self.logError('compiler', status, output, error) 613 def report(command, status, output, error): argument [all …]
|
| /petsc/include/petsc/mpiuni/ |
| H A D | mpi.h | 416 …status) (MPIUNI_ARG(buf), MPIUNI_ARG(count), MPIUNI_ARG(datatype), M… argument 418 …#define MPI_Get_count(status, datatype, count) (M… argument 431 …ine MPI_Wait(request, status) (MPIUN… argument 432 … MPI_Test(request, flag, status) (MPIUNI_A… argument 434 …quests, index, status) (MPIUNI_ARG(count), MPIUNI_ARG(array_of_… argument 443 … flag, status) (MPIUNI_ARG(source), MPIUNI_ARG(tag), M… argument 444 …ag, comm, status) (MPIUNI_ARG(source), MPIUNI_A… argument 446 …#define MPI_Test_cancelled(status, flag) (M… argument 457 … sendcount, sendtype, dest, sendtag, recvbuf, recvcount, recvtype, source, recvtag, comm, status) \ argument 458 …), MPIUNI_ARG(source), MPIUNI_ARG(recvtag), MPIUNI_ARG(comm), MPIUNI_ARG(status), MPIUNI_Memcpy(re… [all …]
|
| /petsc/ |
| H A D | setup.py | 169 status = os.system(" ".join(command)) 170 if status != 0: 171 raise RuntimeError(status) 212 status = os.system(" ".join(command)) 213 if status != 0: 214 raise RuntimeError(status) 229 status = os.system(" ".join(command)) 230 if status != 0: 231 raise RuntimeError(status)
|
| /petsc/src/mat/impls/aij/seq/umfpack/ |
| H A D | umfpack.c | 116 int status; in MatSolve_UMFPACK_Private() local 134 …status = umfpack_UMF_wsolve(uflag, ai, aj, (PetscReal *)av, NULL, (PetscReal *)xa, NULL, (PetscRea… in MatSolve_UMFPACK_Private() 136 …status = umfpack_UMF_wsolve(uflag, ai, aj, av, xa, ba, lu->Numeric, lu->Control, lu->Info, lu->Wi,… in MatSolve_UMFPACK_Private() 139 if (status < 0) { in MatSolve_UMFPACK_Private() 140 umfpack_UMF_report_status(lu->Control, status); in MatSolve_UMFPACK_Private() 170 int status; in MatLUFactorNumeric_UMFPACK() local 179 …status = umfpack_UMF_numeric(ai, aj, (double *)av, NULL, lu->Symbolic, &lu->Numeric, lu->Control, … in MatLUFactorNumeric_UMFPACK() 181 status = umfpack_UMF_numeric(ai, aj, av, lu->Symbolic, &lu->Numeric, lu->Control, lu->Info); in MatLUFactorNumeric_UMFPACK() 183 if (status < 0) { in MatLUFactorNumeric_UMFPACK() 184 umfpack_UMF_report_status(lu->Control, status); in MatLUFactorNumeric_UMFPACK() [all …]
|
| /petsc/src/snes/tutorials/network/power/ |
| H A D | PFReadData.c | 92 Load[loadi].status = 1; in PFReadMatPowerData() 107 int bus_i, status; in PFReadMatPowerData() local 108 …d %lf %lf %lf %lf %lf %lf %d %lf %lf", &bus_i, &pg, &qg, &qt, &qb, &vs, &mbase, &status, &pt, &pb); in PFReadMatPowerData() 110 Gen[geni].status = status; in PFReadMatPowerData() 133 int fbus, tbus, status; in PFReadMatPowerData() local 134 …lf %lf %lf %d", &fbus, &tbus, &r, &x, &b, &rateA, &rateB, &rateC, &tapratio, &phaseshift, &status); in PFReadMatPowerData() 137 Branch[bri].status = status; in PFReadMatPowerData()
|
| /petsc/lib/petsc/bin/maint/ |
| H A D | generateetags.py | 87 …status = subprocess.call('ctags --fields=+l --tag-relative=yes --langmap=c:+.cu -I PeNS,PeOP -a -f… 88 if status: 89 …status = subprocess.call('/usr/local/bin/ctags -a -f '+ctagfile+' '+' '.join(frlist), shell=True, … 90 if status: 91 …status = subprocess.call('ctags -a -f '+ctagfile+' '+' '.join(frlist), shell=True, stdout=DEVNULL,… 92 if status:
|
| /petsc/src/dm/interface/ |
| H A D | dmget.c | 318 …PetscCheck(nlink->status == DMVEC_STATUS_IN, ((PetscObject)dm)->comm, PETSC_ERR_ARG_WRONGSTATE, "D… in DMClearNamedGlobalVectors() 348 …PetscCheck(nlink->status == DMVEC_STATUS_IN, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "DM still … in DMClearNamedLocalVectors() 426 …PetscCheck(link->status == DMVEC_STATUS_IN, PetscObjectComm((PetscObject)dm), PETSC_ERR_ARG_WRONGS… in DMGetNamedGlobalVector() 443 link->status = DMVEC_STATUS_OUT; in DMGetNamedGlobalVector() 478 …PetscCheck(link->status == DMVEC_STATUS_OUT, PetscObjectComm((PetscObject)dm), PETSC_ERR_ARG_WRONG… in DMRestoreNamedGlobalVector() 482 link->status = DMVEC_STATUS_IN; in DMRestoreNamedGlobalVector() 564 …PetscCheck(link->status == DMVEC_STATUS_IN, PetscObjectComm((PetscObject)dm), PETSC_ERR_ARG_WRONGS… in DMGetNamedLocalVector() 581 link->status = DMVEC_STATUS_OUT; in DMGetNamedLocalVector() 616 …PetscCheck(link->status == DMVEC_STATUS_OUT, PetscObjectComm((PetscObject)dm), PETSC_ERR_ARG_WRONG… in DMRestoreNamedLocalVector() 620 link->status = DMVEC_STATUS_IN; in DMRestoreNamedLocalVector()
|
| /petsc/src/sys/utils/ |
| H A D | mpiu.c | 11 MPI_Status status; in PetscSequentialPhaseBegin_Private() local 17 if (rank) PetscCallMPI(MPI_Recv(NULL, 0, MPI_INT, rank - 1, tag, comm, &status)); in PetscSequentialPhaseBegin_Private() 26 MPI_Status status; in PetscSequentialPhaseEnd_Private() local 35 if (rank == 0) PetscCallMPI(MPI_Recv(NULL, 0, MPI_INT, size - 1, tag, comm, &status)); in PetscSequentialPhaseEnd_Private()
|
| /petsc/src/mat/graphops/order/amd/ |
| H A D | amd.c | 23 int status; in MatGetOrdering_AMD() local 52 status = amd_AMD_order(nrow, ia, ja, perm, Control, Info); in MatGetOrdering_AMD() 53 switch (status) { in MatGetOrdering_AMD()
|
| /petsc/src/tao/leastsquares/tutorials/ |
| H A D | chwirut2f.F90 | 252 MPIU_Status status 254 MPIU_Status status(MPI_STATUS_SIZE) 261 PetscCallMPI(MPI_Recv(x, nn, MPIU_SCALAR, zero, MPI_ANY_TAG, PETSC_COMM_WORLD, status, ierr)) 263 tag = status%MPI_TAG 265 tag = status(MPI_TAG) 294 MPIU_Status status 296 MPIU_Status status(MPI_STATUS_SIZE) 305 …CallMPI(MPI_Recv(f, one, MPIU_SCALAR, MPI_ANY_SOURCE, MPI_ANY_TAG, PETSC_COMM_WORLD, status, ierr)) 308 source = status%MPI_SOURCE 310 source = status(MPI_SOURCE) [all …]
|
| H A D | chwirut2.c | 117 MPI_Status status; in EvaluateFunction() local 124 …PetscCallMPI(MPI_Recv(&f_i, 1, MPIU_REAL, MPI_ANY_SOURCE, MPI_ANY_TAG, PETSC_COMM_WORLD, &status)); in EvaluateFunction() 125 if (status.MPI_TAG == IDLE_TAG) { in EvaluateFunction() 128 tag = status.MPI_TAG; in EvaluateFunction() 134 …PetscCallMPI(MPI_Send(x, NPARAMETERS, MPIU_REAL, status.MPI_SOURCE, (PetscMPIInt)next_task, PETSC_… in EvaluateFunction() 139 … PetscCallMPI(MPI_Send(x, NPARAMETERS, MPIU_REAL, status.MPI_SOURCE, IDLE_TAG, PETSC_COMM_WORLD)); in EvaluateFunction() 606 MPI_Status status; in TaskWorker() local 613 PetscCallMPI(MPI_Recv(x, NPARAMETERS, MPIU_REAL, 0, MPI_ANY_TAG, PETSC_COMM_WORLD, &status)); in TaskWorker() 614 tag = status.MPI_TAG; in TaskWorker() 643 MPI_Status status; in StopWorkers() local [all …]
|
| /petsc/src/mat/impls/aij/seq/lusol/ |
| H A D | lusol.c | 195 int i, m, n, nnz, status; in MatSolve_LUSOL() local 206 …c, lusol->indr, lusol->ip, lusol->iq, lusol->lenc, lusol->lenr, lusol->locc, lusol->locr, &status); in MatSolve_LUSOL() 208 PetscCheck(!status, PETSC_COMM_SELF, PETSC_ERR_ARG_SIZ, "solve failed, error code %d", status); in MatSolve_LUSOL() 219 int m, n, nz, nnz, status; in MatLUFactorNumeric_LUSOL() local 271 …->locc, lusol->locr, lusol->iploc, lusol->iqloc, lusol->ipinv, lusol->iqinv, lusol->mnsw, &status); in MatLUFactorNumeric_LUSOL() 273 switch (status) { in MatLUFactorNumeric_LUSOL() 293 } while (status == 7); in MatLUFactorNumeric_LUSOL()
|