| /libCEED/examples/fluids/src/ |
| H A D | misc.c | 114 PetscErrorCode LoadFluidsBinaryVec(MPI_Comm comm, PetscViewer viewer, Vec Q, PetscReal *time, Petsc… in LoadFluidsBinaryVec() argument 134 …PetscCheck(length == N, comm, PETSC_ERR_ARG_INCOMP, "File Vec has length %" PetscInt_FMT " but DM … in LoadFluidsBinaryVec() 136 …} else SETERRQ(comm, PETSC_ERR_FILE_UNEXPECTED, "Not a fluids header token or a PETSc Vec in file"… in LoadFluidsBinaryVec() 147 MPI_Comm comm = PetscObjectComm((PetscObject)Q); in RegressionTest() local 152 …PetscCheck(strcmp(app_ctx->test_file_path, "") != 0, comm, PETSC_ERR_FILE_READ, "File for regressi… in RegressionTest() 153 PetscCall(PetscViewerBinaryOpen(comm, app_ctx->test_file_path, FILE_MODE_READ, &viewer)); in RegressionTest() 154 PetscCall(LoadFluidsBinaryVec(comm, viewer, Q_ref, NULL, NULL)); in RegressionTest() 234 PetscErrorCode SetupICsFromBinary(MPI_Comm comm, AppCtx app_ctx, Vec Q) { in SetupICsFromBinary() argument 238 PetscCall(PetscViewerBinaryOpen(comm, app_ctx->cont_file, FILE_MODE_READ, &viewer)); in SetupICsFromBinary() 239 PetscCall(LoadFluidsBinaryVec(comm, viewer, Q, &app_ctx->cont_time, &app_ctx->cont_steps)); in SetupICsFromBinary() [all …]
|
| /libCEED/examples/petsc/ |
| H A D | bpsswarm.c | 51 MPI_Comm comm; in main() local 74 comm = PETSC_COMM_WORLD; in main() 75 PetscCall(MPI_Comm_size(comm, &comm_size)); in main() 79 PetscCall(MPI_Comm_split_type(comm, MPI_COMM_TYPE_SHARED, 0, MPI_INFO_NULL, &splitcomm)); in main() 88 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main() 124 …PetscCheck(tmp == dim, comm, PETSC_ERR_USER, "Number of values for -dm_plex_box_faces must match d… in main() 127 …PetscCheck(!user_set_num_points_per_cell || point_swarm_type != SWARM_SINUSOIDAL, comm, PETSC_ERR_… in main() 171 PetscCall(DMPlexCreateFromFile(comm, filename, NULL, PETSC_TRUE, &dm_mesh)); in main() 173 PetscCall(DMCreate(comm, &dm_mesh)); in main() 182 … PetscCheck(!is_simplex, comm, PETSC_ERR_USER, "Only tensor-product background meshes supported"); in main() [all …]
|
| H A D | bpssphere.c | 50 MPI_Comm comm; in main() local 71 comm = PETSC_COMM_WORLD; in main() 74 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main() 146 PetscCall(MatCreateShell(comm, l_size, l_size, g_size, g_size, op_apply_ctx, &mat_O)); in main() 154 PetscCall(PetscPrintf(comm, in main() 199 PetscCall(SetupApplyOperatorCtx(comm, dm, ceed, ceed_data, X_loc, op_apply_ctx)); in main() 202 PetscCall(KSPCreate(comm, &ksp)); in main() 230 PetscCall(MPI_Allreduce(MPI_IN_PLACE, &my_rt, 1, MPI_DOUBLE, MPI_MIN, comm)); in main() 266 PetscCall(PetscPrintf(comm, in main() 275 PetscCall(PetscPrintf(comm, " Performance:\n")); in main() [all …]
|
| H A D | bps.c | 104 PetscCall(MatCreateShell(rp->comm, l_size, l_size, g_size, g_size, op_apply_ctx, &mat_O)); in RunWithDM() 134 PetscCall(MPI_Comm_size(rp->comm, &comm_size)); in RunWithDM() 135 PetscCall(PetscPrintf(rp->comm, in RunWithDM() 191 PetscCall(SetupApplyOperatorCtx(rp->comm, dm, ceed, ceed_data, X_loc, op_apply_ctx)); in RunWithDM() 192 PetscCall(KSPCreate(rp->comm, &ksp)); in RunWithDM() 218 PetscCall(MPI_Allreduce(MPI_IN_PLACE, &my_rt, 1, MPI_DOUBLE, MPI_MIN, rp->comm)); in RunWithDM() 253 PetscCall(PetscPrintf(rp->comm, in RunWithDM() 262 PetscCall(PetscPrintf(rp->comm, " Performance:\n")); in RunWithDM() 266 … PetscCall(SetupErrorOperatorCtx(rp->comm, dm, ceed, ceed_data, X_loc, op_error, op_error_ctx)); in RunWithDM() 275 PetscCall(MPI_Allreduce(&my_rt, &rt_min, 1, MPI_DOUBLE, MPI_MIN, rp->comm)); in RunWithDM() [all …]
|
| H A D | area.c | 59 MPI_Comm comm; in main() local 78 comm = PETSC_COMM_WORLD; in main() 81 PetscOptionsBegin(comm, NULL, "CEED surface area problem with PETSc", NULL); in main() 151 PetscCall(PetscPrintf(comm, in main() 177 PetscCall(PetscPrintf(comm, "Computing the mesh area using the formula: area = 1^T M 1\n")); in main() 205 …PetscCall(PetscPrintf(comm, "Exact mesh surface area : % .14g\n", exact_surfac… in main() 206 PetscCall(PetscPrintf(comm, "Computed mesh surface area : % .14g\n", area)); in main() 207 PetscCall(PetscPrintf(comm, "Area error : % .14g\n", error)); in main()
|
| H A D | multigrid.c | 49 MPI_Comm comm; in main() local 75 comm = PETSC_COMM_WORLD; in main() 78 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main() 204 …PetscCall(MatCreateShell(comm, l_size[i], l_size[i], g_size[i], g_size[i], op_apply_ctx[i], &mat_O… in main() 213 …PetscCall(MatCreateShell(comm, l_size[i], l_size[i - 1], g_size[i], g_size[i - 1], pr_restr_ctx[i]… in main() 230 PetscCall(PetscPrintf(comm, in main() 262 PetscCall(PetscPrintf(comm, in main() 328 PetscCall(SetupApplyOperatorCtx(comm, dm[i], ceed, ceed_data[i], X_loc[i], op_apply_ctx[i])); in main() 333 pr_restr_ctx[i]->comm = comm; in main() 380 PetscCall(KSPCreate(comm, &ksp)); in main() [all …]
|
| H A D | bpsraw.c | 112 MPI_Comm comm; member 331 PetscCall(MPI_Allreduce(MPI_IN_PLACE, max_error, 1, MPIU_REAL, MPIU_MAX, op_apply_ctx->comm)); in ComputeErrorMax() 339 MPI_Comm comm; in main() local 365 comm = PETSC_COMM_WORLD; in main() 368 PetscOptionsBegin(comm, NULL, "CEED BPs in PETSc", NULL); in main() 415 PetscCall(MPI_Comm_size(comm, &size)); in main() 425 PetscCall(MPI_Comm_rank(comm, &rank)); in main() 435 PetscCall(VecCreate(comm, &X)); in main() 451 PetscCall(PetscPrintf(comm, in main() 517 PetscCall(ISCreateBlock(comm, num_comp_u, l_size, l_to_g_ind, PETSC_OWN_POINTER, &l_to_g_is)); in main() [all …]
|
| /libCEED/examples/fluids/problems/ |
| H A D | stg_shur14.c | 35 PetscErrorCode CalcCholeskyDecomp(MPI_Comm comm, PetscInt nprofs, const CeedScalar Rij[6][nprofs], … in CalcCholeskyDecomp() argument 45 PetscCheck(!isnan(Cij[0][i]) && !isnan(Cij[1][i]) && !isnan(Cij[2][i]), comm, PETSC_ERR_FP, in CalcCholeskyDecomp() 63 static PetscErrorCode ReadStgInflow(const MPI_Comm comm, const char path[PETSC_MAX_PATH_LEN], StgSh… in ReadStgInflow() argument 72 PetscCall(PhastaDatFileOpen(comm, path, char_array_len, dims, &fp)); in ReadStgInflow() 81 PetscCall(PetscSynchronizedFGets(comm, fp, char_array_len, line)); in ReadStgInflow() 83 PetscCheck(ndims == dims[1], comm, PETSC_ERR_FILE_UNEXPECTED, in ReadStgInflow() 99 …PetscCheck(wall_dist[i] >= 0, comm, PETSC_ERR_FILE_UNEXPECTED, "Distance to wall in %s cannot be n… in ReadStgInflow() 100 …PetscCheck(lt[i] >= 0, comm, PETSC_ERR_FILE_UNEXPECTED, "Turbulent length scale in %s cannot be ne… in ReadStgInflow() 101 …PetscCheck(eps[i] >= 0, comm, PETSC_ERR_FILE_UNEXPECTED, "Turbulent dissipation in %s cannot be ne… in ReadStgInflow() 105 PetscCall(CalcCholeskyDecomp(comm, stg_ctx->nprofs, rij, cij)); in ReadStgInflow() [all …]
|
| H A D | shocktube.c | 21 MPI_Comm comm = user->comm; in NS_SHOCKTUBE() local 74 PetscOptionsBegin(comm, NULL, "Options for SHOCKTUBE problem", NULL); in NS_SHOCKTUBE() 90 … PetscCall(PetscPrintf(comm, "Warning! -stab supg not implemented for the shocktube problem. \n")); in NS_SHOCKTUBE() 93 PetscCall(PetscPrintf(comm, "Warning! -yzb only implemented for explicit timestepping. \n")); in NS_SHOCKTUBE() 145 MPI_Comm comm = user->comm; in PRINT_SHOCKTUBE() local 148 PetscCall(PetscPrintf(comm, in PRINT_SHOCKTUBE()
|
| H A D | eulervortex.c | 22 MPI_Comm comm = user->comm; in NS_EULER_VORTEX() local 69 PetscOptionsBegin(comm, NULL, "Options for EULER_VORTEX problem", NULL); in NS_EULER_VORTEX() 91 PetscCall(PetscPrintf(comm, "Warning! Use -stab supg only with -implicit\n")); in NS_EULER_VORTEX() 94 …PetscCall(PetscPrintf(comm, "Warning! Background velocity vector for -euler_test t1 and -euler_tes… in NS_EULER_VORTEX() 146 MPI_Comm comm = user->comm; in PRINT_EULER_VORTEX() local 152 PetscCall(PetscPrintf(comm, in PRINT_EULER_VORTEX()
|
| H A D | blasius.c | 117 static PetscErrorCode GetYNodeLocs(const MPI_Comm comm, const char path[PETSC_MAX_PATH_LEN], PetscR… in GetYNodeLocs() argument 125 PetscCall(PetscFOpen(comm, path, "r", &fp)); in GetYNodeLocs() 126 PetscCall(PetscSynchronizedFGets(comm, fp, char_array_len, line)); in GetYNodeLocs() 142 PetscCall(PetscSynchronizedFGets(comm, fp, char_array_len, line)); in GetYNodeLocs() 144 PetscCheck(ndims == dims[1], comm, PETSC_ERR_FILE_UNEXPECTED, in GetYNodeLocs() 150 PetscCall(PetscFClose(comm, fp)); in GetYNodeLocs() 166 static PetscErrorCode ModifyMesh(MPI_Comm comm, DM dm, PetscInt dim, PetscReal growth, PetscInt N, … in ModifyMesh() argument 218 PetscCheck(*num_node_locs >= faces[1] + 1, comm, PETSC_ERR_FILE_UNEXPECTED, in ModifyMesh() 222 PetscCall(PetscPrintf(comm, in ModifyMesh() 243 MPI_Comm comm = user->comm; in NS_BLASIUS() local [all …]
|
| H A D | gaussianwave.c | 21 MPI_Comm comm = user->comm; in NS_GAUSSIAN_WAVE() local 51 PetscOptionsBegin(comm, NULL, "Options for GAUSSIAN_WAVE problem", NULL); in NS_GAUSSIAN_WAVE() 54 …PetscCheck(narray == 3, comm, PETSC_ERR_ARG_SIZ, "-epicenter should recieve array of size 3, inste… in NS_GAUSSIAN_WAVE()
|
| H A D | advection.c | 92 MPI_Comm comm = user->comm; in NS_ADVECTION() local 162 PetscOptionsBegin(comm, NULL, "Options for ADVECTION problem", NULL); in NS_ADVECTION() 198 … PetscCall(PetscPrintf(comm, "Warning! Use -wind_translation only with -wind_type translation\n")); in NS_ADVECTION() 202 PetscCall(PetscPrintf(comm, in NS_ADVECTION() 206 PetscCall(PetscPrintf(comm, "Warning! Use -CtauS only with -stab su or -stab supg\n")); in NS_ADVECTION() 277 MPI_Comm comm = user->comm; in PRINT_ADVECTION() local 285 PetscCall(PetscPrintf(comm, in PRINT_ADVECTION() 298 …PetscCall(PetscPrintf(comm, " Background Wind : %f,%f\n", setup_ctx->wind[0]… in PRINT_ADVECTION() 301 …PetscCall(PetscPrintf(comm, " Background Wind : %f,%f,%f\n", setup_ctx->wind… in PRINT_ADVECTION()
|
| /libCEED/examples/solids/src/ |
| H A D | misc.c | 23 PetscErrorCode SetupJacobianCtx(MPI_Comm comm, AppCtx app_ctx, DM dm, Vec V, Vec V_loc, CeedData ce… in SetupJacobianCtx() argument 28 jacobian_ctx->comm = comm; in SetupJacobianCtx() 51 PetscErrorCode SetupProlongRestrictCtx(MPI_Comm comm, AppCtx app_ctx, DM dm_c, DM dm_f, Vec V_f, Ve… in SetupProlongRestrictCtx() argument 56 prolong_restr_ctx->comm = comm; in SetupProlongRestrictCtx() 112 PetscErrorCode ViewSolution(MPI_Comm comm, AppCtx app_ctx, Vec U, PetscInt increment, PetscScalar l… in ViewSolution() argument 121 MPI_Comm_rank(comm, &rank); in ViewSolution() 134 PetscCall(PetscViewerVTKOpen(comm, output_filename, FILE_MODE_WRITE, &viewer)); in ViewSolution() 144 PetscErrorCode ViewDiagnosticQuantities(MPI_Comm comm, DM dmU, UserMult user, AppCtx app_ctx, Vec U… in ViewDiagnosticQuantities() argument 216 PetscCall(PetscViewerVTKOpen(comm, output_filename, FILE_MODE_WRITE, &viewer)); in ViewDiagnosticQuantities()
|
| H A D | setup-dm.c | 34 PetscErrorCode CreateDistributedDM(MPI_Comm comm, AppCtx app_ctx, DM *dm) { in CreateDistributedDM() argument 46 …PetscCall(DMPlexCreateBoxMesh(comm, dim, PETSC_FALSE, faces, NULL, NULL, NULL, interpolate, 0, PET… in CreateDistributedDM() 48 PetscCall(DMPlexCreateFromFile(comm, filename, NULL, interpolate, dm)); in CreateDistributedDM() 66 MPI_Comm comm; in SetupDMByDegree() local 78 PetscCall(PetscObjectGetComm((PetscObject)dm, &comm)); in SetupDMByDegree() 79 PetscCall(PetscFECreateLagrange(comm, dim, num_comp_u, PETSC_FALSE, order, order, &fe)); in SetupDMByDegree() 88 PetscCall(PetscFECreateLagrange(comm, dim, num_comp_coord, PETSC_FALSE, 1, 1, &fe_coords)); in SetupDMByDegree()
|
| /libCEED/examples/solids/problems/ |
| H A D | neo-hookean.c | 14 PetscErrorCode PhysicsContext_NH(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx)… in PhysicsContext_NH() argument 21 PetscCall(ProcessPhysics_NH(comm, phys, *units)); in PhysicsContext_NH() 30 PetscErrorCode PhysicsSmootherContext_NH(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF… in PhysicsSmootherContext_NH() argument 37 PetscOptionsBegin(comm, NULL, "Neo-Hookean physical parameters for smoother", NULL); in PhysicsSmootherContext_NH() 62 PetscErrorCode ProcessPhysics_NH(MPI_Comm comm, Physics_NH phys, Units units) { in ProcessPhysics_NH() argument 73 PetscOptionsBegin(comm, NULL, "Neo-Hookean physical parameters", NULL); in ProcessPhysics_NH()
|
| H A D | neo-hookean.h | 24 PetscErrorCode PhysicsContext_NH(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx); 25 PetscErrorCode PhysicsSmootherContext_NH(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF… 28 PetscErrorCode ProcessPhysics_NH(MPI_Comm comm, Physics_NH phys, Units units);
|
| H A D | mooney-rivlin.h | 26 PetscErrorCode PhysicsContext_MR(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx); 27 PetscErrorCode PhysicsSmootherContext_MR(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF… 30 PetscErrorCode ProcessPhysics_MR(MPI_Comm comm, Physics_MR phys, Units units);
|
| H A D | mooney-rivlin.c | 14 PetscErrorCode PhysicsContext_MR(MPI_Comm comm, Ceed ceed, Units *units, CeedQFunctionContext *ctx)… in PhysicsContext_MR() argument 21 PetscCall(ProcessPhysics_MR(comm, phys, *units)); in PhysicsContext_MR() 30 PetscErrorCode PhysicsSmootherContext_MR(MPI_Comm comm, Ceed ceed, CeedQFunctionContext ctx, CeedQF… in PhysicsSmootherContext_MR() argument 37 PetscOptionsBegin(comm, NULL, "Mooney Rivlin physical parameters for smoother", NULL); in PhysicsSmootherContext_MR() 65 PetscErrorCode ProcessPhysics_MR(MPI_Comm comm, Physics_MR phys, Units units) { in ProcessPhysics_MR() argument 76 PetscOptionsBegin(comm, NULL, "Mooney Rivlin physical parameters", NULL); in ProcessPhysics_MR()
|
| /libCEED/examples/solids/include/ |
| H A D | misc.h | 19 PetscErrorCode SetupJacobianCtx(MPI_Comm comm, AppCtx app_ctx, DM dm, Vec V, Vec V_loc, CeedData ce… 23 PetscErrorCode SetupProlongRestrictCtx(MPI_Comm comm, AppCtx app_ctx, DM dm_c, DM dm_f, Vec V_f, Ve… 34 PetscErrorCode ViewSolution(MPI_Comm comm, AppCtx app_ctx, Vec U, PetscInt increment, PetscScalar l… 36 PetscErrorCode ViewDiagnosticQuantities(MPI_Comm comm, DM dm_U, UserMult user, AppCtx app_ctx, Vec …
|
| /libCEED/examples/solids/ |
| H A D | elasticity.c | 44 MPI_Comm comm; in main() local 81 comm = PETSC_COMM_WORLD; in main() 85 PetscCall(ProcessCommandLineOptions(comm, app_ctx)); in main() 105 PetscCall((*SetupPhysics)(comm, ceed, &units, &ctx_phys)); in main() 109 PetscCall((*SetupSmootherPhysics)(comm, ceed, ctx_phys, &ctx_phys_smoother)); in main() 120 PetscCall(CreateDistributedDM(comm, app_ctx, &dm_orig)); in main() 316 PetscCall(MPI_Comm_size(comm, &comm_size)); in main() 318 PetscCall(PetscPrintf(comm, in main() 330 PetscCall(PetscPrintf(comm, in main() 335 PetscCall(PetscPrintf(comm, in main() [all …]
|
| /libCEED/examples/fluids/ |
| H A D | navierstokes.c | 106 MPI_Comm comm = PETSC_COMM_WORLD; in main() local 107 user->comm = comm; in main() 108 PetscCall(ProcessCommandLineOptions(comm, app_ctx, bc)); in main() 116 …PetscCheck(CeedInit(app_ctx->ceed_resource, &ceed) == CEED_ERROR_SUCCESS, comm, PETSC_ERR_LIB, "Ce… in main() 119 …PetscCheck(CeedSetErrorHandler(ceed, CeedErrorStore) == CEED_ERROR_SUCCESS, comm, PETSC_ERR_LIB, "… in main() 160 PetscCall(CreateDM(comm, problem, mat_type, vec_type, &dm)); in main() 220 MPI_Comm_rank(comm, &rank); in main() 230 PetscCall(SetupICsFromBinary(comm, app_ctx, Q)); in main() 296 …PetscCheck(CeedDestroy(&ceed) == CEED_ERROR_SUCCESS, comm, PETSC_ERR_LIB, "Destroying Ceed object … in main()
|
| /libCEED/examples/petsc/include/ |
| H A D | structs.h | 22 MPI_Comm comm; member 33 MPI_Comm comm; member 86 MPI_Comm comm; member
|
| H A D | matops.h | 17 PetscErrorCode SetupApplyOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc… 18 PetscErrorCode SetupErrorOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc…
|
| /libCEED/examples/petsc/src/ |
| H A D | matops.c | 8 PetscErrorCode SetupApplyOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc… in SetupApplyOperatorCtx() argument 10 op_apply_ctx->comm = comm; in SetupApplyOperatorCtx() 24 PetscErrorCode SetupErrorOperatorCtx(MPI_Comm comm, DM dm, Ceed ceed, CeedData ceed_data, Vec X_loc… in SetupErrorOperatorCtx() argument 27 op_error_ctx->comm = comm; in SetupErrorOperatorCtx()
|