1c4762a1bSJed Brown static char help[] = "Tests PetscObjectSetOptions() for TS object\n\n";
2c4762a1bSJed Brown
3c4762a1bSJed Brown /* ------------------------------------------------------------------------
4c4762a1bSJed Brown
5c4762a1bSJed Brown This program solves the PDE
6c4762a1bSJed Brown
7c4762a1bSJed Brown u * u_xx
8c4762a1bSJed Brown u_t = ---------
9c4762a1bSJed Brown 2*(t+1)^2
10c4762a1bSJed Brown
11c4762a1bSJed Brown on the domain 0 <= x <= 1, with boundary conditions
12c4762a1bSJed Brown u(t,0) = t + 1, u(t,1) = 2*t + 2,
13c4762a1bSJed Brown and initial condition
14c4762a1bSJed Brown u(0,x) = 1 + x*x.
15c4762a1bSJed Brown
16c4762a1bSJed Brown The exact solution is:
17c4762a1bSJed Brown u(t,x) = (1 + x*x) * (1 + t)
18c4762a1bSJed Brown
19c4762a1bSJed Brown Note that since the solution is linear in time and quadratic in x,
20c4762a1bSJed Brown the finite difference scheme actually computes the "exact" solution.
21c4762a1bSJed Brown
22c4762a1bSJed Brown We use by default the backward Euler method.
23c4762a1bSJed Brown
24c4762a1bSJed Brown ------------------------------------------------------------------------- */
25c4762a1bSJed Brown
26c4762a1bSJed Brown /*
27c4762a1bSJed Brown Include "petscts.h" to use the PETSc timestepping routines. Note that
28c4762a1bSJed Brown this file automatically includes "petscsys.h" and other lower-level
29c4762a1bSJed Brown PETSc include files.
30c4762a1bSJed Brown
31c4762a1bSJed Brown Include the "petscdmda.h" to allow us to use the distributed array data
32c4762a1bSJed Brown structures to manage the parallel grid.
33c4762a1bSJed Brown */
34c4762a1bSJed Brown #include <petscts.h>
35c4762a1bSJed Brown #include <petscdm.h>
36c4762a1bSJed Brown #include <petscdmda.h>
37c4762a1bSJed Brown #include <petscdraw.h>
38c4762a1bSJed Brown
39c4762a1bSJed Brown /*
40c4762a1bSJed Brown User-defined application context - contains data needed by the
41c4762a1bSJed Brown application-provided callback routines.
42c4762a1bSJed Brown */
43c4762a1bSJed Brown typedef struct {
44c4762a1bSJed Brown MPI_Comm comm; /* communicator */
45c4762a1bSJed Brown DM da; /* distributed array data structure */
46c4762a1bSJed Brown Vec localwork; /* local ghosted work vector */
47c4762a1bSJed Brown Vec u_local; /* local ghosted approximate solution vector */
48c4762a1bSJed Brown Vec solution; /* global exact solution vector */
49c4762a1bSJed Brown PetscInt m; /* total number of grid points */
50c4762a1bSJed Brown PetscReal h; /* mesh width: h = 1/(m-1) */
51c4762a1bSJed Brown } AppCtx;
52c4762a1bSJed Brown
53c4762a1bSJed Brown /*
54c4762a1bSJed Brown User-defined routines, provided below.
55c4762a1bSJed Brown */
56c4762a1bSJed Brown extern PetscErrorCode InitialConditions(Vec, AppCtx *);
57c4762a1bSJed Brown extern PetscErrorCode RHSFunction(TS, PetscReal, Vec, Vec, void *);
58c4762a1bSJed Brown extern PetscErrorCode RHSJacobian(TS, PetscReal, Vec, Mat, Mat, void *);
59c4762a1bSJed Brown extern PetscErrorCode ExactSolution(PetscReal, Vec, AppCtx *);
60c4762a1bSJed Brown
main(int argc,char ** argv)61d71ae5a4SJacob Faibussowitsch int main(int argc, char **argv)
62d71ae5a4SJacob Faibussowitsch {
63c4762a1bSJed Brown AppCtx appctx; /* user-defined application context */
64c4762a1bSJed Brown TS ts; /* timestepping context */
65c4762a1bSJed Brown Mat A; /* Jacobian matrix data structure */
66c4762a1bSJed Brown Vec u; /* approximate solution vector */
67c4762a1bSJed Brown PetscInt time_steps_max = 100; /* default max timesteps */
68c4762a1bSJed Brown PetscReal dt;
69c4762a1bSJed Brown PetscReal time_total_max = 100.0; /* default max total time */
70c4762a1bSJed Brown PetscOptions options, optionscopy;
71c4762a1bSJed Brown
72c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
73c4762a1bSJed Brown Initialize program and set problem parameters
74c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
75c4762a1bSJed Brown
76327415f7SBarry Smith PetscFunctionBeginUser;
77c8025a54SPierre Jolivet PetscCall(PetscInitialize(&argc, &argv, NULL, help));
78c4762a1bSJed Brown
799566063dSJacob Faibussowitsch PetscCall(PetscOptionsCreate(&options));
809566063dSJacob Faibussowitsch PetscCall(PetscOptionsSetValue(options, "-ts_monitor", "ascii"));
819566063dSJacob Faibussowitsch PetscCall(PetscOptionsSetValue(options, "-snes_monitor", "ascii"));
829566063dSJacob Faibussowitsch PetscCall(PetscOptionsSetValue(options, "-ksp_monitor", "ascii"));
83c4762a1bSJed Brown
84c4762a1bSJed Brown appctx.comm = PETSC_COMM_WORLD;
85c4762a1bSJed Brown appctx.m = 60;
86c4762a1bSJed Brown
879566063dSJacob Faibussowitsch PetscCall(PetscOptionsGetInt(options, NULL, "-M", &appctx.m, NULL));
88c4762a1bSJed Brown
89c4762a1bSJed Brown appctx.h = 1.0 / (appctx.m - 1.0);
90c4762a1bSJed Brown
91c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
92c4762a1bSJed Brown Create vector data structures
93c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
94c4762a1bSJed Brown
95c4762a1bSJed Brown /*
96c4762a1bSJed Brown Create distributed array (DMDA) to manage parallel grid and vectors
97c4762a1bSJed Brown and to set up the ghost point communication pattern. There are M
98c4762a1bSJed Brown total grid values spread equally among all the processors.
99c4762a1bSJed Brown */
1009566063dSJacob Faibussowitsch PetscCall(DMDACreate1d(PETSC_COMM_WORLD, DM_BOUNDARY_NONE, appctx.m, 1, 1, NULL, &appctx.da));
1019566063dSJacob Faibussowitsch PetscCall(PetscObjectSetOptions((PetscObject)appctx.da, options));
1029566063dSJacob Faibussowitsch PetscCall(DMSetFromOptions(appctx.da));
1039566063dSJacob Faibussowitsch PetscCall(DMSetUp(appctx.da));
104c4762a1bSJed Brown
105c4762a1bSJed Brown /*
106c4762a1bSJed Brown Extract global and local vectors from DMDA; we use these to store the
107c4762a1bSJed Brown approximate solution. Then duplicate these for remaining vectors that
108c4762a1bSJed Brown have the same types.
109c4762a1bSJed Brown */
1109566063dSJacob Faibussowitsch PetscCall(DMCreateGlobalVector(appctx.da, &u));
1119566063dSJacob Faibussowitsch PetscCall(DMCreateLocalVector(appctx.da, &appctx.u_local));
112c4762a1bSJed Brown
113c4762a1bSJed Brown /*
114c4762a1bSJed Brown Create local work vector for use in evaluating right-hand-side function;
115c4762a1bSJed Brown create global work vector for storing exact solution.
116c4762a1bSJed Brown */
1179566063dSJacob Faibussowitsch PetscCall(VecDuplicate(appctx.u_local, &appctx.localwork));
1189566063dSJacob Faibussowitsch PetscCall(VecDuplicate(u, &appctx.solution));
119c4762a1bSJed Brown
120c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
121c4762a1bSJed Brown Create timestepping solver context; set callback routine for
122c4762a1bSJed Brown right-hand-side function evaluation.
123c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
124c4762a1bSJed Brown
1259566063dSJacob Faibussowitsch PetscCall(TSCreate(PETSC_COMM_WORLD, &ts));
1269566063dSJacob Faibussowitsch PetscCall(PetscObjectSetOptions((PetscObject)ts, options));
1279566063dSJacob Faibussowitsch PetscCall(TSSetProblemType(ts, TS_NONLINEAR));
1289566063dSJacob Faibussowitsch PetscCall(TSSetRHSFunction(ts, NULL, RHSFunction, &appctx));
129c4762a1bSJed Brown
130c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
131c4762a1bSJed Brown For nonlinear problems, the user can provide a Jacobian evaluation
132c4762a1bSJed Brown routine (or use a finite differencing approximation).
133c4762a1bSJed Brown
134c4762a1bSJed Brown Create matrix data structure; set Jacobian evaluation routine.
135c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
136c4762a1bSJed Brown
1379566063dSJacob Faibussowitsch PetscCall(MatCreate(PETSC_COMM_WORLD, &A));
1389566063dSJacob Faibussowitsch PetscCall(PetscObjectSetOptions((PetscObject)A, options));
1399566063dSJacob Faibussowitsch PetscCall(MatSetSizes(A, PETSC_DECIDE, PETSC_DECIDE, appctx.m, appctx.m));
1409566063dSJacob Faibussowitsch PetscCall(MatSetFromOptions(A));
1419566063dSJacob Faibussowitsch PetscCall(MatSetUp(A));
1429566063dSJacob Faibussowitsch PetscCall(TSSetRHSJacobian(ts, A, A, RHSJacobian, &appctx));
143c4762a1bSJed Brown
144c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
145c4762a1bSJed Brown Set solution vector and initial timestep
146c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
147c4762a1bSJed Brown
148c4762a1bSJed Brown dt = appctx.h / 2.0;
1499566063dSJacob Faibussowitsch PetscCall(TSSetTimeStep(ts, dt));
150c4762a1bSJed Brown
151c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
152c4762a1bSJed Brown Customize timestepping solver:
153c4762a1bSJed Brown - Set the solution method to be the Backward Euler method.
154c4762a1bSJed Brown - Set timestepping duration info
155c4762a1bSJed Brown Then set runtime options, which can override these defaults.
156c4762a1bSJed Brown For example,
157c4762a1bSJed Brown -ts_max_steps <maxsteps> -ts_max_time <maxtime>
158c4762a1bSJed Brown to override the defaults set by TSSetMaxSteps()/TSSetMaxTime().
159c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
160c4762a1bSJed Brown
1619566063dSJacob Faibussowitsch PetscCall(TSSetType(ts, TSBEULER));
1629566063dSJacob Faibussowitsch PetscCall(TSSetMaxSteps(ts, time_steps_max));
1639566063dSJacob Faibussowitsch PetscCall(TSSetMaxTime(ts, time_total_max));
1649566063dSJacob Faibussowitsch PetscCall(TSSetExactFinalTime(ts, TS_EXACTFINALTIME_STEPOVER));
1659566063dSJacob Faibussowitsch PetscCall(TSSetFromOptions(ts));
166c4762a1bSJed Brown
167c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
168c4762a1bSJed Brown Solve the problem
169c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
170c4762a1bSJed Brown
171c4762a1bSJed Brown /*
172c4762a1bSJed Brown Evaluate initial conditions
173c4762a1bSJed Brown */
1749566063dSJacob Faibussowitsch PetscCall(InitialConditions(u, &appctx));
175c4762a1bSJed Brown
176c4762a1bSJed Brown /*
177c4762a1bSJed Brown Run the timestepping solver
178c4762a1bSJed Brown */
1799566063dSJacob Faibussowitsch PetscCall(TSSolve(ts, u));
180c4762a1bSJed Brown
181c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
182c4762a1bSJed Brown Free work space. All PETSc objects should be destroyed when they
183c4762a1bSJed Brown are no longer needed.
184c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
185c4762a1bSJed Brown
1869566063dSJacob Faibussowitsch PetscCall(PetscObjectGetOptions((PetscObject)ts, &optionscopy));
1873c633725SBarry Smith PetscCheck(options == optionscopy, PETSC_COMM_WORLD, PETSC_ERR_PLIB, "PetscObjectGetOptions() failed");
188c4762a1bSJed Brown
1899566063dSJacob Faibussowitsch PetscCall(TSDestroy(&ts));
1909566063dSJacob Faibussowitsch PetscCall(VecDestroy(&u));
1919566063dSJacob Faibussowitsch PetscCall(MatDestroy(&A));
1929566063dSJacob Faibussowitsch PetscCall(DMDestroy(&appctx.da));
1939566063dSJacob Faibussowitsch PetscCall(VecDestroy(&appctx.localwork));
1949566063dSJacob Faibussowitsch PetscCall(VecDestroy(&appctx.solution));
1959566063dSJacob Faibussowitsch PetscCall(VecDestroy(&appctx.u_local));
1969566063dSJacob Faibussowitsch PetscCall(PetscOptionsDestroy(&options));
197c4762a1bSJed Brown
198c4762a1bSJed Brown /*
199c4762a1bSJed Brown Always call PetscFinalize() before exiting a program. This routine
200c4762a1bSJed Brown - finalizes the PETSc libraries as well as MPI
201c4762a1bSJed Brown - provides summary and diagnostic information if certain runtime
202c4762a1bSJed Brown options are chosen (e.g., -log_view).
203c4762a1bSJed Brown */
2049566063dSJacob Faibussowitsch PetscCall(PetscFinalize());
205b122ec5aSJacob Faibussowitsch return 0;
206c4762a1bSJed Brown }
207c4762a1bSJed Brown /* --------------------------------------------------------------------- */
208c4762a1bSJed Brown /*
209c4762a1bSJed Brown InitialConditions - Computes the solution at the initial time.
210c4762a1bSJed Brown
211c4762a1bSJed Brown Input Parameters:
212c4762a1bSJed Brown u - uninitialized solution vector (global)
213c4762a1bSJed Brown appctx - user-defined application context
214c4762a1bSJed Brown
215c4762a1bSJed Brown Output Parameter:
216c4762a1bSJed Brown u - vector with solution at initial time (global)
217c4762a1bSJed Brown */
InitialConditions(Vec u,AppCtx * appctx)218d71ae5a4SJacob Faibussowitsch PetscErrorCode InitialConditions(Vec u, AppCtx *appctx)
219d71ae5a4SJacob Faibussowitsch {
220c4762a1bSJed Brown PetscScalar *u_localptr, h = appctx->h, x;
221c4762a1bSJed Brown PetscInt i, mybase, myend;
222c4762a1bSJed Brown
2233ba16761SJacob Faibussowitsch PetscFunctionBeginUser;
224c4762a1bSJed Brown /*
225c4762a1bSJed Brown Determine starting point of each processor's range of
226c4762a1bSJed Brown grid values.
227c4762a1bSJed Brown */
2289566063dSJacob Faibussowitsch PetscCall(VecGetOwnershipRange(u, &mybase, &myend));
229c4762a1bSJed Brown
230c4762a1bSJed Brown /*
231c4762a1bSJed Brown Get a pointer to vector data.
232c4762a1bSJed Brown - For default PETSc vectors, VecGetArray() returns a pointer to
233c4762a1bSJed Brown the data array. Otherwise, the routine is implementation dependent.
234c4762a1bSJed Brown - You MUST call VecRestoreArray() when you no longer need access to
235c4762a1bSJed Brown the array.
236c4762a1bSJed Brown - Note that the Fortran interface to VecGetArray() differs from the
237c4762a1bSJed Brown C version. See the users manual for details.
238c4762a1bSJed Brown */
2399566063dSJacob Faibussowitsch PetscCall(VecGetArray(u, &u_localptr));
240c4762a1bSJed Brown
241c4762a1bSJed Brown /*
242c4762a1bSJed Brown We initialize the solution array by simply writing the solution
243c4762a1bSJed Brown directly into the array locations. Alternatively, we could use
244c4762a1bSJed Brown VecSetValues() or VecSetValuesLocal().
245c4762a1bSJed Brown */
246c4762a1bSJed Brown for (i = mybase; i < myend; i++) {
247c4762a1bSJed Brown x = h * (PetscReal)i; /* current location in global grid */
248c4762a1bSJed Brown u_localptr[i - mybase] = 1.0 + x * x;
249c4762a1bSJed Brown }
250c4762a1bSJed Brown
251c4762a1bSJed Brown /*
252c4762a1bSJed Brown Restore vector
253c4762a1bSJed Brown */
2549566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(u, &u_localptr));
2553ba16761SJacob Faibussowitsch PetscFunctionReturn(PETSC_SUCCESS);
256c4762a1bSJed Brown }
257c4762a1bSJed Brown /* --------------------------------------------------------------------- */
258c4762a1bSJed Brown /*
259c4762a1bSJed Brown ExactSolution - Computes the exact solution at a given time.
260c4762a1bSJed Brown
261c4762a1bSJed Brown Input Parameters:
262c4762a1bSJed Brown t - current time
263c4762a1bSJed Brown solution - vector in which exact solution will be computed
264c4762a1bSJed Brown appctx - user-defined application context
265c4762a1bSJed Brown
266c4762a1bSJed Brown Output Parameter:
267c4762a1bSJed Brown solution - vector with the newly computed exact solution
268c4762a1bSJed Brown */
ExactSolution(PetscReal t,Vec solution,AppCtx * appctx)269d71ae5a4SJacob Faibussowitsch PetscErrorCode ExactSolution(PetscReal t, Vec solution, AppCtx *appctx)
270d71ae5a4SJacob Faibussowitsch {
271c4762a1bSJed Brown PetscScalar *s_localptr, h = appctx->h, x;
272c4762a1bSJed Brown PetscInt i, mybase, myend;
273c4762a1bSJed Brown
2743ba16761SJacob Faibussowitsch PetscFunctionBeginUser;
275c4762a1bSJed Brown /*
276c4762a1bSJed Brown Determine starting and ending points of each processor's
277c4762a1bSJed Brown range of grid values
278c4762a1bSJed Brown */
2799566063dSJacob Faibussowitsch PetscCall(VecGetOwnershipRange(solution, &mybase, &myend));
280c4762a1bSJed Brown
281c4762a1bSJed Brown /*
282c4762a1bSJed Brown Get a pointer to vector data.
283c4762a1bSJed Brown */
2849566063dSJacob Faibussowitsch PetscCall(VecGetArray(solution, &s_localptr));
285c4762a1bSJed Brown
286c4762a1bSJed Brown /*
287c4762a1bSJed Brown Simply write the solution directly into the array locations.
288c4762a1bSJed Brown Alternatively, we could use VecSetValues() or VecSetValuesLocal().
289c4762a1bSJed Brown */
290c4762a1bSJed Brown for (i = mybase; i < myend; i++) {
291c4762a1bSJed Brown x = h * (PetscReal)i;
292c4762a1bSJed Brown s_localptr[i - mybase] = (t + 1.0) * (1.0 + x * x);
293c4762a1bSJed Brown }
294c4762a1bSJed Brown
295c4762a1bSJed Brown /*
296c4762a1bSJed Brown Restore vector
297c4762a1bSJed Brown */
2989566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(solution, &s_localptr));
2993ba16761SJacob Faibussowitsch PetscFunctionReturn(PETSC_SUCCESS);
300c4762a1bSJed Brown }
301c4762a1bSJed Brown /* --------------------------------------------------------------------- */
302c4762a1bSJed Brown /*
303c4762a1bSJed Brown RHSFunction - User-provided routine that evalues the right-hand-side
304c4762a1bSJed Brown function of the ODE. This routine is set in the main program by
305c4762a1bSJed Brown calling TSSetRHSFunction(). We compute:
306c4762a1bSJed Brown global_out = F(global_in)
307c4762a1bSJed Brown
308c4762a1bSJed Brown Input Parameters:
309c4762a1bSJed Brown ts - timesteping context
310c4762a1bSJed Brown t - current time
311c4762a1bSJed Brown global_in - vector containing the current iterate
312c4762a1bSJed Brown ctx - (optional) user-provided context for function evaluation.
313c4762a1bSJed Brown In this case we use the appctx defined above.
314c4762a1bSJed Brown
315c4762a1bSJed Brown Output Parameter:
316c4762a1bSJed Brown global_out - vector containing the newly evaluated function
317c4762a1bSJed Brown */
RHSFunction(TS ts,PetscReal t,Vec global_in,Vec global_out,PetscCtx ctx)318*2a8381b2SBarry Smith PetscErrorCode RHSFunction(TS ts, PetscReal t, Vec global_in, Vec global_out, PetscCtx ctx)
319d71ae5a4SJacob Faibussowitsch {
320c4762a1bSJed Brown AppCtx *appctx = (AppCtx *)ctx; /* user-defined application context */
321c4762a1bSJed Brown DM da = appctx->da; /* distributed array */
322c4762a1bSJed Brown Vec local_in = appctx->u_local; /* local ghosted input vector */
323c4762a1bSJed Brown Vec localwork = appctx->localwork; /* local ghosted work vector */
324c4762a1bSJed Brown PetscInt i, localsize;
325c4762a1bSJed Brown PetscMPIInt rank, size;
326c4762a1bSJed Brown PetscScalar *copyptr, sc;
327c4762a1bSJed Brown const PetscScalar *localptr;
328c4762a1bSJed Brown
3293ba16761SJacob Faibussowitsch PetscFunctionBeginUser;
330c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
331c4762a1bSJed Brown Get ready for local function computations
332c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
333c4762a1bSJed Brown /*
334c4762a1bSJed Brown Scatter ghost points to local vector, using the 2-step process
335c4762a1bSJed Brown DMGlobalToLocalBegin(), DMGlobalToLocalEnd().
336c4762a1bSJed Brown By placing code between these two statements, computations can be
337c4762a1bSJed Brown done while messages are in transition.
338c4762a1bSJed Brown */
3399566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalBegin(da, global_in, INSERT_VALUES, local_in));
3409566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalEnd(da, global_in, INSERT_VALUES, local_in));
341c4762a1bSJed Brown
342c4762a1bSJed Brown /*
343c4762a1bSJed Brown Access directly the values in our local INPUT work array
344c4762a1bSJed Brown */
3459566063dSJacob Faibussowitsch PetscCall(VecGetArrayRead(local_in, &localptr));
346c4762a1bSJed Brown
347c4762a1bSJed Brown /*
348c4762a1bSJed Brown Access directly the values in our local OUTPUT work array
349c4762a1bSJed Brown */
3509566063dSJacob Faibussowitsch PetscCall(VecGetArray(localwork, ©ptr));
351c4762a1bSJed Brown
352c4762a1bSJed Brown sc = 1.0 / (appctx->h * appctx->h * 2.0 * (1.0 + t) * (1.0 + t));
353c4762a1bSJed Brown
354c4762a1bSJed Brown /*
355c4762a1bSJed Brown Evaluate our function on the nodes owned by this processor
356c4762a1bSJed Brown */
3579566063dSJacob Faibussowitsch PetscCall(VecGetLocalSize(local_in, &localsize));
358c4762a1bSJed Brown
359c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
360c4762a1bSJed Brown Compute entries for the locally owned part
361c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
362c4762a1bSJed Brown
363c4762a1bSJed Brown /*
364c4762a1bSJed Brown Handle boundary conditions: This is done by using the boundary condition
365c4762a1bSJed Brown u(t,boundary) = g(t,boundary)
366c4762a1bSJed Brown for some function g. Now take the derivative with respect to t to obtain
367c4762a1bSJed Brown u_{t}(t,boundary) = g_{t}(t,boundary)
368c4762a1bSJed Brown
369c4762a1bSJed Brown In our case, u(t,0) = t + 1, so that u_{t}(t,0) = 1
370c4762a1bSJed Brown and u(t,1) = 2t+ 2, so that u_{t}(t,1) = 2
371c4762a1bSJed Brown */
3729566063dSJacob Faibussowitsch PetscCallMPI(MPI_Comm_rank(appctx->comm, &rank));
3739566063dSJacob Faibussowitsch PetscCallMPI(MPI_Comm_size(appctx->comm, &size));
374dd400576SPatrick Sanan if (rank == 0) copyptr[0] = 1.0;
375c4762a1bSJed Brown if (rank == size - 1) copyptr[localsize - 1] = 2.0;
376c4762a1bSJed Brown
377c4762a1bSJed Brown /*
378c4762a1bSJed Brown Handle the interior nodes where the PDE is replace by finite
379c4762a1bSJed Brown difference operators.
380c4762a1bSJed Brown */
381c4762a1bSJed Brown for (i = 1; i < localsize - 1; i++) copyptr[i] = localptr[i] * sc * (localptr[i + 1] + localptr[i - 1] - 2.0 * localptr[i]);
382c4762a1bSJed Brown
383c4762a1bSJed Brown /*
384c4762a1bSJed Brown Restore vectors
385c4762a1bSJed Brown */
3869566063dSJacob Faibussowitsch PetscCall(VecRestoreArrayRead(local_in, &localptr));
3879566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(localwork, ©ptr));
388c4762a1bSJed Brown
389c4762a1bSJed Brown /*
390c4762a1bSJed Brown Insert values from the local OUTPUT vector into the global
391c4762a1bSJed Brown output vector
392c4762a1bSJed Brown */
3939566063dSJacob Faibussowitsch PetscCall(DMLocalToGlobalBegin(da, localwork, INSERT_VALUES, global_out));
3949566063dSJacob Faibussowitsch PetscCall(DMLocalToGlobalEnd(da, localwork, INSERT_VALUES, global_out));
3953ba16761SJacob Faibussowitsch PetscFunctionReturn(PETSC_SUCCESS);
396c4762a1bSJed Brown }
397c4762a1bSJed Brown /* --------------------------------------------------------------------- */
398c4762a1bSJed Brown /*
399c4762a1bSJed Brown RHSJacobian - User-provided routine to compute the Jacobian of
400c4762a1bSJed Brown the nonlinear right-hand-side function of the ODE.
401c4762a1bSJed Brown
402c4762a1bSJed Brown Input Parameters:
403c4762a1bSJed Brown ts - the TS context
404c4762a1bSJed Brown t - current time
405c4762a1bSJed Brown global_in - global input vector
406c4762a1bSJed Brown dummy - optional user-defined context, as set by TSetRHSJacobian()
407c4762a1bSJed Brown
408c4762a1bSJed Brown Output Parameters:
409c4762a1bSJed Brown AA - Jacobian matrix
4107addb90fSBarry Smith BB - optionally different matrix used to construct the preconditioner
411c4762a1bSJed Brown
412c4762a1bSJed Brown Notes:
413c4762a1bSJed Brown RHSJacobian computes entries for the locally owned part of the Jacobian.
414c4762a1bSJed Brown - Currently, all PETSc parallel matrix formats are partitioned by
415c4762a1bSJed Brown contiguous chunks of rows across the processors.
416c4762a1bSJed Brown - Each processor needs to insert only elements that it owns
417c4762a1bSJed Brown locally (but any non-local elements will be sent to the
418c4762a1bSJed Brown appropriate processor during matrix assembly).
419c4762a1bSJed Brown - Always specify global row and columns of matrix entries when
420c4762a1bSJed Brown using MatSetValues().
421c4762a1bSJed Brown - Here, we set all entries for a particular row at once.
422c4762a1bSJed Brown - Note that MatSetValues() uses 0-based row and column numbers
423c4762a1bSJed Brown in Fortran as well as in C.
424c4762a1bSJed Brown */
RHSJacobian(TS ts,PetscReal t,Vec global_in,Mat AA,Mat BB,PetscCtx ctx)425*2a8381b2SBarry Smith PetscErrorCode RHSJacobian(TS ts, PetscReal t, Vec global_in, Mat AA, Mat BB, PetscCtx ctx)
426d71ae5a4SJacob Faibussowitsch {
427c4762a1bSJed Brown AppCtx *appctx = (AppCtx *)ctx; /* user-defined application context */
428c4762a1bSJed Brown Vec local_in = appctx->u_local; /* local ghosted input vector */
429c4762a1bSJed Brown DM da = appctx->da; /* distributed array */
430c4762a1bSJed Brown PetscScalar v[3], sc;
431c4762a1bSJed Brown const PetscScalar *localptr;
432c4762a1bSJed Brown PetscInt i, mstart, mend, mstarts, mends, idx[3], is;
433c4762a1bSJed Brown
4343ba16761SJacob Faibussowitsch PetscFunctionBeginUser;
435c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
436c4762a1bSJed Brown Get ready for local Jacobian computations
437c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
438c4762a1bSJed Brown /*
439c4762a1bSJed Brown Scatter ghost points to local vector, using the 2-step process
440c4762a1bSJed Brown DMGlobalToLocalBegin(), DMGlobalToLocalEnd().
441c4762a1bSJed Brown By placing code between these two statements, computations can be
442c4762a1bSJed Brown done while messages are in transition.
443c4762a1bSJed Brown */
4449566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalBegin(da, global_in, INSERT_VALUES, local_in));
4459566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalEnd(da, global_in, INSERT_VALUES, local_in));
446c4762a1bSJed Brown
447c4762a1bSJed Brown /*
448c4762a1bSJed Brown Get pointer to vector data
449c4762a1bSJed Brown */
4509566063dSJacob Faibussowitsch PetscCall(VecGetArrayRead(local_in, &localptr));
451c4762a1bSJed Brown
452c4762a1bSJed Brown /*
453c4762a1bSJed Brown Get starting and ending locally owned rows of the matrix
454c4762a1bSJed Brown */
4559566063dSJacob Faibussowitsch PetscCall(MatGetOwnershipRange(BB, &mstarts, &mends));
4569371c9d4SSatish Balay mstart = mstarts;
4579371c9d4SSatish Balay mend = mends;
458c4762a1bSJed Brown
459c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
460c4762a1bSJed Brown Compute entries for the locally owned part of the Jacobian.
461c4762a1bSJed Brown - Currently, all PETSc parallel matrix formats are partitioned by
462c4762a1bSJed Brown contiguous chunks of rows across the processors.
463c4762a1bSJed Brown - Each processor needs to insert only elements that it owns
464c4762a1bSJed Brown locally (but any non-local elements will be sent to the
465c4762a1bSJed Brown appropriate processor during matrix assembly).
466c4762a1bSJed Brown - Here, we set all entries for a particular row at once.
467c4762a1bSJed Brown - We can set matrix entries either using either
468c4762a1bSJed Brown MatSetValuesLocal() or MatSetValues().
469c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
470c4762a1bSJed Brown
471c4762a1bSJed Brown /*
472c4762a1bSJed Brown Set matrix rows corresponding to boundary data
473c4762a1bSJed Brown */
474c4762a1bSJed Brown if (mstart == 0) {
475c4762a1bSJed Brown v[0] = 0.0;
4769566063dSJacob Faibussowitsch PetscCall(MatSetValues(BB, 1, &mstart, 1, &mstart, v, INSERT_VALUES));
477c4762a1bSJed Brown mstart++;
478c4762a1bSJed Brown }
479c4762a1bSJed Brown if (mend == appctx->m) {
480c4762a1bSJed Brown mend--;
481c4762a1bSJed Brown v[0] = 0.0;
4829566063dSJacob Faibussowitsch PetscCall(MatSetValues(BB, 1, &mend, 1, &mend, v, INSERT_VALUES));
483c4762a1bSJed Brown }
484c4762a1bSJed Brown
485c4762a1bSJed Brown /*
486c4762a1bSJed Brown Set matrix rows corresponding to interior data. We construct the
487c4762a1bSJed Brown matrix one row at a time.
488c4762a1bSJed Brown */
489c4762a1bSJed Brown sc = 1.0 / (appctx->h * appctx->h * 2.0 * (1.0 + t) * (1.0 + t));
490c4762a1bSJed Brown for (i = mstart; i < mend; i++) {
4919371c9d4SSatish Balay idx[0] = i - 1;
4929371c9d4SSatish Balay idx[1] = i;
4939371c9d4SSatish Balay idx[2] = i + 1;
494c4762a1bSJed Brown is = i - mstart + 1;
495c4762a1bSJed Brown v[0] = sc * localptr[is];
496c4762a1bSJed Brown v[1] = sc * (localptr[is + 1] + localptr[is - 1] - 4.0 * localptr[is]);
497c4762a1bSJed Brown v[2] = sc * localptr[is];
4989566063dSJacob Faibussowitsch PetscCall(MatSetValues(BB, 1, &i, 3, idx, v, INSERT_VALUES));
499c4762a1bSJed Brown }
500c4762a1bSJed Brown
501c4762a1bSJed Brown /*
502c4762a1bSJed Brown Restore vector
503c4762a1bSJed Brown */
5049566063dSJacob Faibussowitsch PetscCall(VecRestoreArrayRead(local_in, &localptr));
505c4762a1bSJed Brown
506c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
507c4762a1bSJed Brown Complete the matrix assembly process and set some options
508c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
509c4762a1bSJed Brown /*
510c4762a1bSJed Brown Assemble matrix, using the 2-step process:
511c4762a1bSJed Brown MatAssemblyBegin(), MatAssemblyEnd()
512c4762a1bSJed Brown Computations can be done while messages are in transition
513c4762a1bSJed Brown by placing code between these two statements.
514c4762a1bSJed Brown */
5159566063dSJacob Faibussowitsch PetscCall(MatAssemblyBegin(BB, MAT_FINAL_ASSEMBLY));
5169566063dSJacob Faibussowitsch PetscCall(MatAssemblyEnd(BB, MAT_FINAL_ASSEMBLY));
517c4762a1bSJed Brown if (BB != AA) {
5189566063dSJacob Faibussowitsch PetscCall(MatAssemblyBegin(AA, MAT_FINAL_ASSEMBLY));
5199566063dSJacob Faibussowitsch PetscCall(MatAssemblyEnd(AA, MAT_FINAL_ASSEMBLY));
520c4762a1bSJed Brown }
521c4762a1bSJed Brown
522c4762a1bSJed Brown /*
523c4762a1bSJed Brown Set and option to indicate that we will never add a new nonzero location
524c4762a1bSJed Brown to the matrix. If we do, it will generate an error.
525c4762a1bSJed Brown */
5269566063dSJacob Faibussowitsch PetscCall(MatSetOption(BB, MAT_NEW_NONZERO_LOCATION_ERR, PETSC_TRUE));
5273ba16761SJacob Faibussowitsch PetscFunctionReturn(PETSC_SUCCESS);
528c4762a1bSJed Brown }
529c4762a1bSJed Brown
530c4762a1bSJed Brown /*TEST
531c4762a1bSJed Brown
532c4762a1bSJed Brown test:
533c4762a1bSJed Brown requires: !single
534c4762a1bSJed Brown
535c4762a1bSJed Brown TEST*/
536