Home
last modified time | relevance | path

Searched refs:comm (Results 1 – 25 of 51) sorted by relevance

123

/libCEED/examples/fluids/src/
H A Dmisc.c114 PetscErrorCode LoadFluidsBinaryVec(MPI_Comm comm, PetscViewer viewer, Vec Q, PetscReal *time, Petsc… in LoadFluidsBinaryVec() argument
134 …PetscCheck(length == N, comm, PETSC_ERR_ARG_INCOMP, "File Vec has length %" PetscInt_FMT " but DM … in LoadFluidsBinaryVec()
136 …} else SETERRQ(comm, PETSC_ERR_FILE_UNEXPECTED, "Not a fluids header token or a PETSc Vec in file"… in LoadFluidsBinaryVec()
147 MPI_Comm comm = PetscObjectComm((PetscObject)Q); in RegressionTest() local
152 …PetscCheck(strcmp(app_ctx->test_file_path, "") != 0, comm, PETSC_ERR_FILE_READ, "File for regressi… in RegressionTest()
153 PetscCall(PetscViewerBinaryOpen(comm, app_ctx->test_file_path, FILE_MODE_READ, &viewer)); in RegressionTest()
154 PetscCall(LoadFluidsBinaryVec(comm, viewer, Q_ref, NULL, NULL)); in RegressionTest()
234 PetscErrorCode SetupICsFromBinary(MPI_Comm comm, AppCtx app_ctx, Vec Q) { in SetupICsFromBinary() argument
238 PetscCall(PetscViewerBinaryOpen(comm, app_ctx->cont_file, FILE_MODE_READ, &viewer)); in SetupICsFromBinary()
239 PetscCall(LoadFluidsBinaryVec(comm, viewer, Q, &app_ctx->cont_time, &app_ctx->cont_steps)); in SetupICsFromBinary()
[all …]
/libCEED/examples/petsc/
H A Dbpsswarm.c51 MPI_Comm comm; in main() local
74 comm = PETSC_COMM_WORLD; in main()
75 PetscCall(MPI_Comm_size(comm, &comm_size)); in main()
79 PetscCall(MPI_Comm_split_type(comm, MPI_COMM_TYPE_SHARED, 0, MPI_INFO_NULL, &splitcomm)); in main()
88 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main()
124 …PetscCheck(tmp == dim, comm, PETSC_ERR_USER, "Number of values for -dm_plex_box_faces must match d… in main()
127 …PetscCheck(!user_set_num_points_per_cell || point_swarm_type != SWARM_SINUSOIDAL, comm, PETSC_ERR_… in main()
171 PetscCall(DMPlexCreateFromFile(comm, filename, NULL, PETSC_TRUE, &dm_mesh)); in main()
173 PetscCall(DMCreate(comm, &dm_mesh)); in main()
182 … PetscCheck(!is_simplex, comm, PETSC_ERR_USER, "Only tensor-product background meshes supported"); in main()
[all …]
H A Dbpssphere.c50 MPI_Comm comm; in main() local
71 comm = PETSC_COMM_WORLD; in main()
74 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main()
146 PetscCall(MatCreateShell(comm, l_size, l_size, g_size, g_size, op_apply_ctx, &mat_O)); in main()
154 PetscCall(PetscPrintf(comm, in main()
199 PetscCall(SetupApplyOperatorCtx(comm, dm, ceed, ceed_data, X_loc, op_apply_ctx)); in main()
202 PetscCall(KSPCreate(comm, &ksp)); in main()
230 PetscCall(MPI_Allreduce(MPI_IN_PLACE, &my_rt, 1, MPI_DOUBLE, MPI_MIN, comm)); in main()
266 PetscCall(PetscPrintf(comm, in main()
275 PetscCall(PetscPrintf(comm, " Performance:\n")); in main()
[all …]
H A Dbps.c104 PetscCall(MatCreateShell(rp->comm, l_size, l_size, g_size, g_size, op_apply_ctx, &mat_O)); in RunWithDM()
134 PetscCall(MPI_Comm_size(rp->comm, &comm_size)); in RunWithDM()
135 PetscCall(PetscPrintf(rp->comm, in RunWithDM()
191 PetscCall(SetupApplyOperatorCtx(rp->comm, dm, ceed, ceed_data, X_loc, op_apply_ctx)); in RunWithDM()
192 PetscCall(KSPCreate(rp->comm, &ksp)); in RunWithDM()
218 PetscCall(MPI_Allreduce(MPI_IN_PLACE, &my_rt, 1, MPI_DOUBLE, MPI_MIN, rp->comm)); in RunWithDM()
253 PetscCall(PetscPrintf(rp->comm, in RunWithDM()
262 PetscCall(PetscPrintf(rp->comm, " Performance:\n")); in RunWithDM()
266 … PetscCall(SetupErrorOperatorCtx(rp->comm, dm, ceed, ceed_data, X_loc, op_error, op_error_ctx)); in RunWithDM()
275 PetscCall(MPI_Allreduce(&my_rt, &rt_min, 1, MPI_DOUBLE, MPI_MIN, rp->comm)); in RunWithDM()
[all …]
H A Darea.c59 MPI_Comm comm; in main() local
78 comm = PETSC_COMM_WORLD; in main()
81 PetscOptionsBegin(comm, NULL, "CEED surface area problem with PETSc", NULL); in main()
151 PetscCall(PetscPrintf(comm, in main()
177 PetscCall(PetscPrintf(comm, "Computing the mesh area using the formula: area = 1^T M 1\n")); in main()
205 …PetscCall(PetscPrintf(comm, "Exact mesh surface area : % .14g\n", exact_surfac… in main()
206 PetscCall(PetscPrintf(comm, "Computed mesh surface area : % .14g\n", area)); in main()
207 PetscCall(PetscPrintf(comm, "Area error : % .14g\n", error)); in main()
H A Dmultigrid.c49 MPI_Comm comm; in main() local
75 comm = PETSC_COMM_WORLD; in main()
78 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main()
204 …PetscCall(MatCreateShell(comm, l_size[i], l_size[i], g_size[i], g_size[i], op_apply_ctx[i], &mat_O… in main()
213 …PetscCall(MatCreateShell(comm, l_size[i], l_size[i - 1], g_size[i], g_size[i - 1], pr_restr_ctx[i]… in main()
230 PetscCall(PetscPrintf(comm, in main()
262 PetscCall(PetscPrintf(comm, in main()
328 PetscCall(SetupApplyOperatorCtx(comm, dm[i], ceed, ceed_data[i], X_loc[i], op_apply_ctx[i])); in main()
333 pr_restr_ctx[i]->comm = comm; in main()
380 PetscCall(KSPCreate(comm, &ksp)); in main()
[all …]
H A Dbpsraw.c112 MPI_Comm comm; member
331 PetscCall(MPI_Allreduce(MPI_IN_PLACE, max_error, 1, MPIU_REAL, MPIU_MAX, op_apply_ctx->comm)); in ComputeErrorMax()
339 MPI_Comm comm; in main() local
365 comm = PETSC_COMM_WORLD; in main()
368 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main()
415 PetscCall(MPI_Comm_size(comm, &size)); in main()
425 PetscCall(MPI_Comm_rank(comm, &rank)); in main()
435 PetscCall(VecCreate(comm, &X)); in main()
451 PetscCall(PetscPrintf(comm, in main()
517 PetscCall(ISCreateBlock(comm, num_comp_u, l_size, l_to_g_ind, PETSC_OWN_POINTER, &l_to_g_is)); in main()
[all …]
/libCEED/examples/fluids/problems/
H A Dstg_shur14.c35 PetscErrorCode CalcCholeskyDecomp(MPI_Comm comm, PetscInt nprofs, const CeedScalar Rij[6][nprofs], … in CalcCholeskyDecomp() argument
45 PetscCheck(!isnan(Cij[0][i]) && !isnan(Cij[1][i]) && !isnan(Cij[2][i]), comm, PETSC_ERR_FP, in CalcCholeskyDecomp()
63 static PetscErrorCode ReadStgInflow(const MPI_Comm comm, const char path[PETSC_MAX_PATH_LEN], StgSh… in ReadStgInflow() argument
72 PetscCall(PhastaDatFileOpen(comm, path, char_array_len, dims, &fp)); in ReadStgInflow()
81 PetscCall(PetscSynchronizedFGets(comm, fp, char_array_len, line)); in ReadStgInflow()
83 PetscCheck(ndims == dims[1], comm, PETSC_ERR_FILE_UNEXPECTED, in ReadStgInflow()
99 …PetscCheck(wall_dist[i] >= 0, comm, PETSC_ERR_FILE_UNEXPECTED, "Distance to wall in %s cannot be n… in ReadStgInflow()
100 …PetscCheck(lt[i] >= 0, comm, PETSC_ERR_FILE_UNEXPECTED, "Turbulent length scale in %s cannot be ne… in ReadStgInflow()
101 …PetscCheck(eps[i] >= 0, comm, PETSC_ERR_FILE_UNEXPECTED, "Turbulent dissipation in %s cannot be ne… in ReadStgInflow()
105 PetscCall(CalcCholeskyDecomp(comm, stg_ctx->nprofs, rij, cij)); in ReadStgInflow()
[all …]
H A Dshocktube.c21 MPI_Comm comm = user->comm; in NS_SHOCKTUBE() local
74 PetscOptionsBegin(comm, NULL, "Options for SHOCKTUBE problem", NULL); in NS_SHOCKTUBE()
90 … PetscCall(PetscPrintf(comm, "Warning! -stab supg not implemented for the shocktube problem. \n")); in NS_SHOCKTUBE()
93 PetscCall(PetscPrintf(comm, "Warning! -yzb only implemented for explicit timestepping. \n")); in NS_SHOCKTUBE()
145 MPI_Comm comm = user->comm; in PRINT_SHOCKTUBE() local
148 PetscCall(PetscPrintf(comm, in PRINT_SHOCKTUBE()
H A Deulervortex.c22 MPI_Comm comm = user->comm; in NS_EULER_VORTEX() local
69 PetscOptionsBegin(comm, NULL, "Options for EULER_VORTEX problem", NULL); in NS_EULER_VORTEX()
91 PetscCall(PetscPrintf(comm, "Warning! Use -stab supg only with -implicit\n")); in NS_EULER_VORTEX()
94 …PetscCall(PetscPrintf(comm, "Warning! Background velocity vector for -euler_test t1 and -euler_tes… in NS_EULER_VORTEX()
146 MPI_Comm comm = user->comm; in PRINT_EULER_VORTEX() local
152 PetscCall(PetscPrintf(comm, in PRINT_EULER_VORTEX()
H A Dblasius.c117 static PetscErrorCode GetYNodeLocs(const MPI_Comm comm, const char path[PETSC_MAX_PATH_LEN], PetscR… in GetYNodeLocs() argument
125 PetscCall(PetscFOpen(comm, path, "r", &fp)); in GetYNodeLocs()
126 PetscCall(PetscSynchronizedFGets(comm, fp, char_array_len, line)); in GetYNodeLocs()
142 PetscCall(PetscSynchronizedFGets(comm, fp, char_array_len, line)); in GetYNodeLocs()
144 PetscCheck(ndims == dims[1], comm, PETSC_ERR_FILE_UNEXPECTED, in GetYNodeLocs()
150 PetscCall(PetscFClose(comm, fp)); in GetYNodeLocs()
166 static PetscErrorCode ModifyMesh(MPI_Comm comm, DM dm, PetscInt dim, PetscReal growth, PetscInt N, … in ModifyMesh() argument
218 PetscCheck(*num_node_locs >= faces[1] + 1, comm, PETSC_ERR_FILE_UNEXPECTED, in ModifyMesh()
222 PetscCall(PetscPrintf(comm, in ModifyMesh()
243 MPI_Comm comm = user->comm; in NS_BLASIUS() local
[all …]
H A Dgaussianwave.c21 MPI_Comm comm = user->comm; in NS_GAUSSIAN_WAVE() local
51 PetscOptionsBegin(comm, NULL, "Options for GAUSSIAN_WAVE problem", NULL); in NS_GAUSSIAN_WAVE()
54 …PetscCheck(narray == 3, comm, PETSC_ERR_ARG_SIZ, "-epicenter should recieve array of size 3, inste… in NS_GAUSSIAN_WAVE()
H A Dadvection.c92 MPI_Comm comm = user->comm; in NS_ADVECTION() local
162 PetscOptionsBegin(comm, NULL, "Options for ADVECTION problem", NULL); in NS_ADVECTION()
198 … PetscCall(PetscPrintf(comm, "Warning! Use -wind_translation only with -wind_type translation\n")); in NS_ADVECTION()
202 PetscCall(PetscPrintf(comm, in NS_ADVECTION()
206 PetscCall(PetscPrintf(comm, "Warning! Use -CtauS only with -stab su or -stab supg\n")); in NS_ADVECTION()
277 MPI_Comm comm = user->comm; in PRINT_ADVECTION() local
285 PetscCall(PetscPrintf(comm, in PRINT_ADVECTION()
298 …PetscCall(PetscPrintf(comm, " Background Wind : %f,%f\n", setup_ctx->wind[0]… in PRINT_ADVECTION()
301 …PetscCall(PetscPrintf(comm, " Background Wind : %f,%f,%f\n", setup_ctx->wind… in PRINT_ADVECTION()
/libCEED/examples/solids/src/
H A Dmisc.c23 PetscErrorCode SetupJacobianCtx(MPI_Comm comm, AppCtx app_ctx, DM dm, Vec V, Vec V_loc, CeedData ce… in SetupJacobianCtx() argument
28 jacobian_ctx->comm = comm; in SetupJacobianCtx()
51 PetscErrorCode SetupProlongRestrictCtx(MPI_Comm comm, AppCtx app_ctx, DM dm_c, DM dm_f, Vec V_f, Ve… in SetupProlongRestrictCtx() argument
56 prolong_restr_ctx->comm = comm; in SetupProlongRestrictCtx()
112 PetscErrorCode ViewSolution(MPI_Comm comm, AppCtx app_ctx, Vec U, PetscInt increment, PetscScalar l… in ViewSolution() argument
121 MPI_Comm_rank(comm, &rank); in ViewSolution()
134 PetscCall(PetscViewerVTKOpen(comm, output_filename, FILE_MODE_WRITE, &viewer)); in ViewSolution()
144 PetscErrorCode ViewDiagnosticQuantities(MPI_Comm comm, DM dmU, UserMult user, AppCtx app_ctx, Vec U… in ViewDiagnosticQuantities() argument
216 PetscCall(PetscViewerVTKOpen(comm, output_filename, FILE_MODE_WRITE, &viewer)); in ViewDiagnosticQuantities()
H A Dsetup-dm.c34 PetscErrorCode CreateDistributedDM(MPI_Comm comm, AppCtx app_ctx, DM *dm) { in CreateDistributedDM() argument
46 …PetscCall(DMPlexCreateBoxMesh(comm, dim, PETSC_FALSE, faces, NULL, NULL, NULL, interpolate, 0, PET… in CreateDistributedDM()
48 PetscCall(DMPlexCreateFromFile(comm, filename, NULL, interpolate, dm)); in CreateDistributedDM()
66 MPI_Comm comm; in SetupDMByDegree() local
78 PetscCall(PetscObjectGetComm((PetscObject)dm, &comm)); in SetupDMByDegree()
79 PetscCall(PetscFECreateLagrange(comm, dim, num_comp_u, PETSC_FALSE, order, order, &fe)); in SetupDMByDegree()
88 PetscCall(PetscFECreateLagrange(comm, dim, num_comp_coord, PETSC_FALSE, 1, 1, &fe_coords)); in SetupDMByDegree()
/libCEED/examples/solids/problems/
H A Dneo-hookean.c14 PetscErrorCode PhysicsContext_NH(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx)… in PhysicsContext_NH() argument
21 PetscCall(ProcessPhysics_NH(comm, phys, *units)); in PhysicsContext_NH()
30 PetscErrorCode PhysicsSmootherContext_NH(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF… in PhysicsSmootherContext_NH() argument
37 PetscOptionsBegin(comm, NULL, "Neo-Hookean physical parameters for smoother", NULL); in PhysicsSmootherContext_NH()
62 PetscErrorCode ProcessPhysics_NH(MPI_Comm comm, Physics_NH phys, Units units) { in ProcessPhysics_NH() argument
73 PetscOptionsBegin(comm, NULL, "Neo-Hookean physical parameters", NULL); in ProcessPhysics_NH()
H A Dneo-hookean.h24 PetscErrorCode PhysicsContext_NH(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx);
25 PetscErrorCode PhysicsSmootherContext_NH(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF…
28 PetscErrorCode ProcessPhysics_NH(MPI_Comm comm, Physics_NH phys, Units units);
H A Dmooney-rivlin.h26 PetscErrorCode PhysicsContext_MR(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx);
27 PetscErrorCode PhysicsSmootherContext_MR(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF…
30 PetscErrorCode ProcessPhysics_MR(MPI_Comm comm, Physics_MR phys, Units units);
H A Dmooney-rivlin.c14 PetscErrorCode PhysicsContext_MR(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx)… in PhysicsContext_MR() argument
21 PetscCall(ProcessPhysics_MR(comm, phys, *units)); in PhysicsContext_MR()
30 PetscErrorCode PhysicsSmootherContext_MR(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF… in PhysicsSmootherContext_MR() argument
37 PetscOptionsBegin(comm, NULL, "Mooney Rivlin physical parameters for smoother", NULL); in PhysicsSmootherContext_MR()
65 PetscErrorCode ProcessPhysics_MR(MPI_Comm comm, Physics_MR phys, Units units) { in ProcessPhysics_MR() argument
76 PetscOptionsBegin(comm, NULL, "Mooney Rivlin physical parameters", NULL); in ProcessPhysics_MR()
/libCEED/examples/solids/include/
H A Dmisc.h19 PetscErrorCode SetupJacobianCtx(MPI_Comm comm, AppCtx app_ctx, DM dm, Vec V, Vec V_loc, CeedData ce…
23 PetscErrorCode SetupProlongRestrictCtx(MPI_Comm comm, AppCtx app_ctx, DM dm_c, DM dm_f, Vec V_f, Ve…
34 PetscErrorCode ViewSolution(MPI_Comm comm, AppCtx app_ctx, Vec U, PetscInt increment, PetscScalar l…
36 PetscErrorCode ViewDiagnosticQuantities(MPI_Comm comm, DM dm_U, UserMult user, AppCtx app_ctx, Vec …
/libCEED/examples/solids/
H A Delasticity.c44 MPI_Comm comm; in main() local
81 comm = PETSC_COMM_WORLD; in main()
85 PetscCall(ProcessCommandLineOptions(comm, app_ctx)); in main()
105 PetscCall((*SetupPhysics)(comm, ceed, &units, &ctx_phys)); in main()
109 PetscCall((*SetupSmootherPhysics)(comm, ceed, ctx_phys, &ctx_phys_smoother)); in main()
120 PetscCall(CreateDistributedDM(comm, app_ctx, &dm_orig)); in main()
316 PetscCall(MPI_Comm_size(comm, &comm_size)); in main()
318 PetscCall(PetscPrintf(comm, in main()
330 PetscCall(PetscPrintf(comm, in main()
335 PetscCall(PetscPrintf(comm, in main()
[all …]
/libCEED/examples/fluids/
H A Dnavierstokes.c106 MPI_Comm comm = PETSC_COMM_WORLD; in main() local
107 user->comm = comm; in main()
108 PetscCall(ProcessCommandLineOptions(comm, app_ctx, bc)); in main()
116 …PetscCheck(CeedInit(app_ctx->ceed_resource, &ceed) == CEED_ERROR_SUCCESS, comm, PETSC_ERR_LIB, "Ce… in main()
119 …PetscCheck(CeedSetErrorHandler(ceed, CeedErrorStore) == CEED_ERROR_SUCCESS, comm, PETSC_ERR_LIB, "… in main()
160 PetscCall(CreateDM(comm, problem, mat_type, vec_type, &dm)); in main()
220 MPI_Comm_rank(comm, &rank); in main()
230 PetscCall(SetupICsFromBinary(comm, app_ctx, Q)); in main()
296 …PetscCheck(CeedDestroy(&ceed) == CEED_ERROR_SUCCESS, comm, PETSC_ERR_LIB, "Destroying Ceed object … in main()
/libCEED/examples/petsc/include/
H A Dstructs.h22 MPI_Comm comm; member
33 MPI_Comm comm; member
86 MPI_Comm comm; member
H A Dmatops.h17 PetscErrorCode SetupApplyOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc…
18 PetscErrorCode SetupErrorOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc…
/libCEED/examples/petsc/src/
H A Dmatops.c8 PetscErrorCode SetupApplyOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc… in SetupApplyOperatorCtx() argument
10 op_apply_ctx->comm = comm; in SetupApplyOperatorCtx()
24 PetscErrorCode SetupErrorOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc… in SetupErrorOperatorCtx() argument
27 op_error_ctx->comm = comm; in SetupErrorOperatorCtx()

123