xref: /petsc/doc/manual/getting_started.md (revision d000e8ac2eed99d0cf14796aab94de0aed57fd8b)
1(sec_getting_started)=
2
3# Getting Started
4
5PETSc consists of a collection of classes,
6which are discussed in detail in later parts of the manual ({doc}`programming` and {doc}`additional`).
7The important PETSc classes include
8
9- index sets (`IS`), for indexing into
10  vectors, renumbering, permuting, etc;
11- {any}`ch_vectors` (`Vec`);
12- (generally sparse) {any}`ch_matrices` (`Mat`)
13- {any}`ch_ksp` (`KSP`);
14- preconditioners, including multigrid, block solvers, patch solvers, and
15  sparse direct solvers (`PC`);
16- {any}`ch_snes` (`SNES`);
17- {any}`ch_ts` for solving time-dependent (nonlinear) PDEs, including
18  support for differential-algebraic-equations, and the computation of
19  adjoints (sensitivities/gradients of the solutions) (`TS`);
20- scalable {any}`ch_tao` including a rich set of gradient-based optimizers,
21  Newton-based optimizers and optimization with constraints (`Tao`).
22- {any}`ch_dmbase` code for managing interactions between mesh data structures and vectors,
23  matrices, and solvers (`DM`);
24
25Each class consists of an abstract interface (simply a set of calling
26sequences corresponding to an abstract base class in C++) and an implementation for each algorithm and data structure.
27This design enables easy comparison and use of different
28algorithms (for example, experimenting with different Krylov subspace
29methods, preconditioners, or truncated Newton methods). Hence, PETSc
30provides a rich environment for modeling scientific applications as well
31as for rapid algorithm design and prototyping.
32
33The classes enable easy customization and extension of both algorithms
34and implementations. This approach promotes code reuse and flexibility.
35The PETSc infrastructure creates a foundation for building large-scale
36applications.
37
38It is useful to consider the interrelationships among different pieces
39of PETSc. {any}`fig_library` is a diagram of some
40of these pieces. The figure illustrates the library’s hierarchical
41organization, enabling users to employ the most appropriate solvers for a particular problem.
42
43:::{figure} /images/manual/library_structure.svg
44:alt: PETSc numerical libraries
45:name: fig_library
46
47Numerical Libraries in PETSc
48:::
49
50## Suggested Reading
51
52The manual is divided into four parts:
53
54- {doc}`introduction`
55- {doc}`programming`
56- {doc}`dm`
57- {doc}`additional`
58
59{doc}`introduction` describes the basic procedure for using the PETSc library and
60presents simple examples of solving linear systems with PETSc. This
61section conveys the typical style used throughout the library and
62enables the application programmer to begin using the software
63immediately.
64
65{doc}`programming` explains in detail the use of the various PETSc algebraic objects, such
66as vectors, matrices, index sets, and PETSc solvers, including linear and nonlinear solvers, time integrators,
67and optimization support.
68
69{doc}`dm` details how a user's models and discretizations can easily be interfaced with the
70solvers by using the `DM` construct.
71
72{doc}`additional` describes a variety of useful information, including
73profiling, the options database, viewers, error handling, and some
74details of PETSc design.
75
76[Visual Studio Code](https://code.visualstudio.com/), Eclipse, Emacs, and Vim users may find their development environment's options for
77searching in the source code are
78useful for exploring the PETSc source code. Details of this feature are provided in {any}`sec_developer_environments`.
79
80**Note to Fortran Programmers**: In most of the manual, the examples and calling sequences are given
81for the C/C++ family of programming languages. However, Fortran
82programmers can use all of the functionality of PETSc from Fortran,
83with only minor differences in the user interface.
84{any}`ch_fortran` provides a discussion of the differences between
85using PETSc from Fortran and C, as well as several complete Fortran
86examples.
87
88**Note to Python Programmers**: To program with PETSc in Python, you need to enable Python bindings
89(i.e. petsc4py) with the configure option `--with-petsc4py=1`. See the
90{doc}`PETSc installation guide </install/index>`
91for more details.
92
93(sec_running)=
94
95## Running PETSc Programs
96
97Before using PETSc, the user must first set the environmental variable
98`PETSC_DIR` to indicate the full path of the PETSc home directory. For
99example, under the Unix bash shell, a command of the form
100
101```console
102$ export PETSC_DIR=$HOME/petsc
103```
104
105can be placed in the user’s `.bashrc` or other startup file. In
106addition, the user may need to set the environment variable
107`$PETSC_ARCH` to specify a particular configuration of the PETSc
108libraries. Note that `$PETSC_ARCH` is just a name selected by the
109installer to refer to the libraries compiled for a particular set of
110compiler options and machine type. Using different values of
111`$PETSC_ARCH` allows one to switch between several different sets (say
112debug and optimized versions) of libraries easily. To determine if you need to
113set `$PETSC_ARCH`, look in the directory indicated by `$PETSC_DIR`, if
114there are subdirectories beginning with `arch` then those
115subdirectories give the possible values for `$PETSC_ARCH`.
116
117See {any}`handson` to immediately jump in and run PETSc code.
118
119All PETSc programs use the MPI (Message Passing Interface) standard for
120message-passing communication {cite}`mpi-final`. Thus, to
121execute PETSc programs, users must know the procedure for beginning MPI
122jobs on their selected computer system(s). For instance, when using the
123[MPICH](https://www.mpich.org/) implementation of MPI and many
124others, the following command initiates a program that uses eight
125processors:
126
127```console
128$ mpiexec -n 8 ./petsc_program_name petsc_options
129```
130
131PETSc also provides a script that automatically uses the correct
132`mpiexec` for your configuration.
133
134```console
135$ $PETSC_DIR/lib/petsc/bin/petscmpiexec -n 8 ./petsc_program_name petsc_options
136```
137
138Certain options are supported by all PETSc programs. We list a few
139particularly useful ones below; a complete list can be obtained by
140running any PETSc program with the option `-help`.
141
142- `-log_view` - summarize the program’s performance (see {any}`ch_profiling`)
143- `-fp_trap` - stop on floating-point exceptions; for example divide
144  by zero
145- `-malloc_dump` - enable memory tracing; dump list of unfreed memory
146  at conclusion of the run, see
147  {any}`detecting_memory_problems`,
148- `-malloc_debug` - enable memory debugging (by default, this is
149  activated for the debugging version of PETSc), see
150  {any}`detecting_memory_problems`,
151- `-start_in_debugger` `[noxterm,gdb,lldb]`
152  `[-display name]` - start all (or a subset of the) processes in a debugger. See
153  {any}`sec_debugging`, for more information on
154  debugging PETSc programs.
155- `-on_error_attach_debugger` `[noxterm,gdb,lldb]`
156  `[-display name]` - start debugger only on encountering an error
157- `-info` - print a great deal of information about what the program
158  is doing as it runs
159- `-version` - display the version of PETSc being used
160
161(sec_writing)=
162
163## Writing PETSc Programs
164
165Most PETSc programs begin with a call to
166
167```
168PetscInitialize(int *argc,char ***argv,char *file,char *help);
169```
170
171which initializes PETSc and MPI. The arguments `argc` and `argv` are
172the usual command line arguments in C and C++ programs. The
173argument `file` optionally indicates an alternative name for the PETSc
174options file, `.petscrc`, which resides by default in the user’s home
175directory. {any}`sec_options` provides details
176regarding this file and the PETSc options database, which can be used
177for runtime customization. The final argument, `help`, is an optional
178character string that will be printed if the program is run with the
179`-help` option. In Fortran, the initialization command has the form
180
181```fortran
182call PetscInitialize(character(*) file,integer ierr)
183```
184
185where the file argument is optional.
186
187`PetscInitialize()` automatically calls `MPI_Init()` if MPI has not
188been not previously initialized. In certain circumstances in which MPI
189needs to be initialized directly (or is initialized by some other
190library), the user can first call `MPI_Init()` (or have the other
191library do it), and then call `PetscInitialize()`. By default,
192`PetscInitialize()` sets the PETSc “world” communicator
193`PETSC_COMM_WORLD` to `MPI_COMM_WORLD`.
194
195For those unfamiliar with MPI, a *communicator* indicates
196a collection of processes that will be involved in a
197calculation or communication. Communicators have the variable type
198`MPI_Comm`. In most cases, users can employ the communicator
199`PETSC_COMM_WORLD` to indicate all processes in a given run and
200`PETSC_COMM_SELF` to indicate a single process.
201
202MPI provides routines for generating new communicators consisting of
203subsets of processors, though most users rarely need to use these. The
204book *Using MPI*, by Lusk, Gropp, and Skjellum
205{cite}`using-mpi` provides an excellent introduction to the
206concepts in MPI. See also the [MPI homepage](https://www.mcs.anl.gov/research/projects/mpi/).
207Note that PETSc users
208need not program much message passing directly with MPI, but they must
209be familiar with the basic concepts of message passing and distributed
210memory computing.
211
212All PETSc programs should call `PetscFinalize()` as their final (or
213nearly final) statement. This routine handles options to be called at the conclusion of the
214program and calls `MPI_Finalize()` if `PetscInitialize()` began
215MPI. If MPI was initiated externally from PETSc (by either the user or
216another software package), the user is responsible for calling
217`MPI_Finalize()`.
218
219### Error Checking
220
221Most PETSc functions return a `PetscErrorCode`, an integer
222indicating whether an error occurred during the call. The error code
223is set to be nonzero if an error has been detected; otherwise, it is
224zero. For the C/C++ interface, the error variable is the routine’s
225return value, while for the Fortran version, each PETSc routine has an integer error variable as
226its final argument.
227
228One should always check these routine values as given below in the C/C++
229formats, respectively:
230
231```c
232PetscCall(PetscFunction(Args));
233```
234
235or for Fortran
236
237```fortran
238! within the main program
239PetscCallA(PetscFunction(Args,ierr))
240```
241
242```fortran
243! within any subroutine
244PetscCall(PetscFunction(Args,ierr))
245```
246
247These macros check the returned error code, and if it is nonzero, they call the PETSc error
248handler and then return from the function with the error code. The macros above should be used on all PETSc calls to enable
249a complete error traceback. See {any}`sec_error2` for more details on PETSc error handling.
250
251(sec_simple)=
252
253## Simple PETSc Examples
254
255To help the user use PETSc immediately, we begin with a simple
256uniprocessor example that
257solves the one-dimensional Laplacian problem with finite differences.
258This sequential code illustrates the solution of
259a linear system with `KSP`, the interface to the preconditioners,
260Krylov subspace methods and direct linear solvers of PETSc. Following
261the code, we highlight a few of the most important parts of this example.
262
263:::{admonition} Listing: <a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/src/ksp/ksp/tutorials/ex1.c.html">KSP Tutorial src/ksp/ksp/tutorials/ex1.c</a>
264:name: ksp-ex1
265
266```{literalinclude} /../src/ksp/ksp/tutorials/ex1.c
267:end-before: /*TEST
268```
269:::
270
271### Include Files
272
273The C/C++ include files for PETSc should be used via statements such as
274
275```
276#include <petscksp.h>
277```
278
279where <a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/include/petscksp.h.html">petscksp.h</a>
280is the include file for the linear solver library.
281Each PETSc program must specify an include file corresponding to the
282highest level PETSc objects needed within the program; all of the
283required lower level include files are automatically included within the
284higher level files. For example, <a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/include/petscksp.h.html">petscksp.h</a> includes
285<a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/include/petscmat.h.html">petscmat.h</a>
286(matrices),
287<a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/include/petscvec.h.html">petscvec.h</a>
288(vectors), and
289<a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/include/petscsys.h.html">petscsys.h</a>
290(base PETSc
291file). The PETSc include files are located in the directory
292<a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/include/index.html">\$PETSC_DIR/include</a>.
293See {any}`sec_fortran_includes`
294for a discussion of PETSc include files in Fortran programs.
295
296(the_options_database)=
297
298### The Options Database
299
300As shown in {any}`sec_simple`, the user can
301input control data at run time using the options database. In this
302example the command `PetscOptionsGetInt(NULL,NULL,"-n",&n,NULL);`
303checks whether the user has provided a command line option to set the
304value of `n`, the problem dimension. If so, the variable `n` is set
305accordingly; otherwise, `n` remains unchanged. A complete description
306of the options database may be found in {any}`sec_options`.
307
308(sec_vecintro)=
309
310### Vectors
311
312One creates a new parallel or sequential vector, `x`, of global
313dimension `M` with the commands
314
315```
316VecCreate(MPI_Comm comm,Vec *x);
317VecSetSizes(Vec x, PetscInt m, PetscInt M);
318```
319
320where `comm` denotes the MPI communicator and `m` is the optional
321local size which may be `PETSC_DECIDE`. The type of storage for the
322vector may be set with either calls to `VecSetType()` or
323`VecSetFromOptions()`. Additional vectors of the same type can be
324formed with
325
326```
327VecDuplicate(Vec old,Vec *new);
328```
329
330The commands
331
332```
333VecSet(Vec x,PetscScalar value);
334VecSetValues(Vec x,PetscInt n,PetscInt *indices,PetscScalar *values,INSERT_VALUES);
335```
336
337respectively set all the components of a vector to a particular scalar
338value and assign a different value to each component. More detailed
339information about PETSc vectors, including their basic operations,
340scattering/gathering, index sets, and distributed arrays is available
341in Chapter {any}`ch_vectors`.
342
343Note the use of the PETSc variable type `PetscScalar` in this example.
344`PetscScalar` is defined to be `double` in C/C++ (or
345correspondingly `double precision` in Fortran) for versions of PETSc
346that have *not* been compiled for use with complex numbers. The
347`PetscScalar` data type enables identical code to be used when the
348PETSc libraries have been compiled for use with complex numbers.
349{any}`sec_complex` discusses the use of complex
350numbers in PETSc programs.
351
352(sec_matintro)=
353
354### Matrices
355
356The usage of PETSc matrices and vectors is similar. The user can create a
357new parallel or sequential matrix, `A`, which has `M` global rows
358and `N` global columns, with the routines
359
360```
361MatCreate(MPI_Comm comm,Mat *A);
362MatSetSizes(Mat A,PETSC_DECIDE,PETSC_DECIDE,PetscInt M,PetscInt N);
363```
364
365where the matrix format can be specified at runtime via the options
366database. The user could alternatively specify each processes’ number of
367local rows and columns using `m` and `n`.
368
369```
370MatSetSizes(Mat A,PetscInt m,PetscInt n,PETSC_DETERMINE,PETSC_DETERMINE);
371```
372
373Generally, one then sets the “type” of the matrix, with, for example,
374
375```
376MatSetType(A,MATAIJ);
377```
378
379This causes the matrix `A` to use the compressed sparse row storage
380format to store the matrix entries. See `MatType` for a list of all
381matrix types. Values can then be set with the command
382
383```
384MatSetValues(Mat A,PetscInt m,PetscInt *im,PetscInt n,PetscInt *in,PetscScalar *values,INSERT_VALUES);
385```
386
387After all elements have been inserted into the matrix, it must be
388processed with the pair of commands
389
390```
391MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);
392MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);
393```
394
395{any}`ch_matrices` discusses various matrix formats as
396well as the details of some basic matrix manipulation routines.
397
398### Linear Solvers
399
400After creating the matrix and vectors that define a linear system,
401`Ax` $=$ `b`, the user can then use `KSP` to solve the
402system with the following sequence of commands:
403
404```
405KSPCreate(MPI_Comm comm,KSP *ksp);
406KSPSetOperators(KSP ksp,Mat Amat,Mat Pmat);
407KSPSetFromOptions(KSP ksp);
408KSPSolve(KSP ksp,Vec b,Vec x);
409KSPDestroy(KSP ksp);
410```
411
412The user first creates the `KSP` context and sets the operators
413associated with the system (matrix that defines the linear system,
414`Amat` and matrix from which the preconditioner is constructed,
415`Pmat` ). The user then sets various options for customized solutions,
416solves the linear system, and finally destroys the `KSP` context. The command `KSPSetFromOptions()` enables the user to
417customize the linear solution method at runtime using the options
418database, which is discussed in {any}`sec_options`. Through this database, the
419user not only can select an iterative method and preconditioner, but
420can also prescribe the convergence tolerance, set various monitoring
421routines, etc. (see, e.g., {any}`sec_profiling_programs`).
422
423{any}`ch_ksp` describes in detail the `KSP` package,
424including the `PC` and `KSP` packages for preconditioners and Krylov
425subspace methods.
426
427### Nonlinear Solvers
428
429PETSc provides
430an interface to tackle nonlinear problems called `SNES`.
431{any}`ch_snes` describes the nonlinear
432solvers in detail. We highly recommend most PETSc users work directly with
433`SNES`, rather than using PETSc for the linear problem and writing their own
434nonlinear solver. Similarly, users should use `TS` rather than rolling their own time integrators.
435
436(sec_error2)=
437
438### Error Checking
439
440As noted above, PETSc functions return a `PetscErrorCode`, which is an integer
441indicating whether an error has occurred during the call. Below, we indicate a traceback
442generated by error detection within a sample PETSc program. The error
443occurred on line 3618 of the file
444`$PETSC_DIR/src/mat/impls/aij/seq/aij.c` and was caused by trying to
445allocate too large an array in memory. The routine was called in the
446program `ex3.c` on line 66. See
447{any}`sec_fortran_errors` for details regarding error checking
448when using the PETSc Fortran interface.
449
450```none
451$ cd $PETSC_DIR/src/ksp/ksp/tutorials
452$ make ex3
453$ mpiexec -n 1 ./ex3 -m 100000
454[0]PETSC ERROR: --------------------- Error Message --------------------------------
455[0]PETSC ERROR: Out of memory. This could be due to allocating
456[0]PETSC ERROR: too large an object or bleeding by not properly
457[0]PETSC ERROR: destroying unneeded objects.
458[0]PETSC ERROR: Memory allocated 11282182704 Memory used by process 7075897344
459[0]PETSC ERROR: Try running with -malloc_dump or -malloc_view for info.
460[0]PETSC ERROR: Memory requested 18446744072169447424
461[0]PETSC ERROR: PETSc Development Git Revision: v3.7.1-224-g9c9a9c5 Git Date: 2016-05-18 22:43:00 -0500
462[0]PETSC ERROR: ./ex3 on a arch-darwin-double-debug named Patricks-MacBook-Pro-2.local by patrick Mon Jun 27 18:04:03 2016
463[0]PETSC ERROR: Configure options PETSC_DIR=/Users/patrick/petsc PETSC_ARCH=arch-darwin-double-debug --download-mpich --download-f2cblaslapack --with-cc=clang --with-cxx=clang++ --with-fc=gfortran --with-debugging=1 --with-precision=double --with-scalar-type=real --with-viennacl=0 --download-c2html -download-sowing
464[0]PETSC ERROR: #1 MatSeqAIJSetPreallocation_SeqAIJ() line 3618 in /Users/patrick/petsc/src/mat/impls/aij/seq/aij.c
465[0]PETSC ERROR: #2 PetscTrMallocDefault() line 188 in /Users/patrick/petsc/src/sys/memory/mtr.c
466[0]PETSC ERROR: #3 MatSeqAIJSetPreallocation_SeqAIJ() line 3618 in /Users/patrick/petsc/src/mat/impls/aij/seq/aij.c
467[0]PETSC ERROR: #4 MatSeqAIJSetPreallocation() line 3562 in /Users/patrick/petsc/src/mat/impls/aij/seq/aij.c
468[0]PETSC ERROR: #5 main() line 66 in /Users/patrick/petsc/src/ksp/ksp/tutorials/ex3.c
469[0]PETSC ERROR: PETSc Option Table entries:
470[0]PETSC ERROR: -m 100000
471[0]PETSC ERROR: ----------------End of Error Message ------- send entire error message to petsc-maint@mcs.anl.gov----------
472```
473
474When running the debug version [^debug-footnote] of the PETSc libraries, it checks for memory corruption (writing outside of array bounds
475, etc.). The macro `CHKMEMQ` can be called anywhere in the code to check
476the current status of the memory for corruption. By putting several (or
477many) of these macros into your code, you can usually easily track down
478in what small segment of your code the corruption has occurred. One can
479also use Valgrind to track down memory errors; see the [FAQ](https://petsc.org/release/faq/).
480
481For complete error handling, calls to MPI functions should be made with `PetscCallMPI(MPI_Function(Args))`.
482In Fortran subroutines use `PetscCallMPI(MPI_Function(Args, ierr))` and in Fortran main use
483`PetscCallMPIA(MPI_Function(Args, ierr))`.
484
485PETSc has a small number of C/C++-only macros that do not explicitly return error codes. These are used in the style
486
487```c
488XXXBegin(Args);
489other code
490XXXEnd();
491```
492
493and include `PetscOptionsBegin()`, `PetscOptionsEnd()`, `PetscObjectOptionsBegin()`,
494`PetscOptionsHeadBegin()`, `PetscOptionsHeadEnd()`, `PetscDrawCollectiveBegin()`, `PetscDrawCollectiveEnd()`,
495`MatPreallocateEnd()`, and `MatPreallocateBegin()`. These should not be checked for error codes.
496Another class of functions with the `Begin()` and `End()` paradigm
497including `MatAssemblyBegin()`, and `MatAssemblyEnd()` do return error codes that should be checked.
498
499PETSc also has a set of C/C++-only macros that return an object, or `NULL` if an error has been detected. These include
500`PETSC_VIEWER_STDOUT_WORLD`, `PETSC_VIEWER_DRAW_WORLD`, `PETSC_VIEWER_STDOUT_(MPI_Comm)`, and `PETSC_VIEWER_DRAW_(MPI_Comm)`.
501
502Finally `PetscObjectComm((PetscObject)x)` returns the communicator associated with the object `x` or `MPI_COMM_NULL` if an
503error was detected.
504
505(sec_parallel)=
506
507# Parallel and GPU Programming
508
509Numerical computing today has multiple levels of parallelism (concurrency).
510
511- Low-level, single instruction multiple data (SIMD) parallelism or, somewhat similar, on-GPU parallelism,
512- medium-level, multiple instruction multiple data shared memory parallelism (thread parallelism), and
513- high-level, distributed memory parallelism.
514
515Traditional CPUs support the lower two levels via, for example, Intel AVX-like instructions ({any}`sec_cpu_simd`) and Unix threads, often managed by using OpenMP pragmas ({any}`sec_cpu_openmp`),
516(or multiple processes). GPUs also support the lower two levels via kernel functions ({any}`sec_gpu_kernels`) and streams ({any}`sec_gpu_streams`).
517Distributed memory parallelism is created by combining multiple
518CPUs and/or GPUs and using MPI for communication ({any}`sec_mpi`).
519
520In addition, there is also concurrency between computations (floating point operations) and data movement (from memory to caches and registers
521and via MPI between distinct memory nodes).
522
523PETSc supports all these parallelism levels, but its strongest support is for MPI-based distributed memory parallelism.
524
525(sec_mpi)=
526
527## MPI Parallelism
528
529Since PETSc uses the message-passing model for parallel programming and
530employs MPI for all interprocessor communication, the user can
531employ MPI routines as needed throughout an application code. However,
532by default, the user is shielded from many of the details of message
533passing within PETSc since these are hidden within parallel objects,
534such as vectors, matrices, and solvers. In addition, PETSc provides
535tools such as vector scatter and gather to assist in the
536management of parallel data.
537
538Recall that the user must specify a communicator upon creation of any
539PETSc object (such as a vector, matrix, or solver) to indicate the
540processors over which the object is to be distributed. For example, as
541mentioned above, some commands for matrix, vector, and linear solver
542creation are:
543
544```
545MatCreate(MPI_Comm comm,Mat *A);
546VecCreate(MPI_Comm comm,Vec *x);
547KSPCreate(MPI_Comm comm,KSP *ksp);
548```
549
550The creation routines are collective on all processes in the
551communicator; thus, all processors in the communicator *must* call the
552creation routine. In addition, if a sequence of collective routines is
553being used, they *must* be called in the same order on each process.
554
555The next example, given below,
556illustrates the solution of a linear system in parallel. This code,
557corresponding to
558<a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/src/ksp/ksp/tutorials/ex2.c.html">KSP Tutorial ex2</a>,
559handles the two-dimensional Laplacian discretized with finite
560differences, where the linear system is again solved with KSP. The code
561performs the same tasks as the sequential version within
562{any}`sec_simple`. Note that the user interface
563for initiating the program, creating vectors and matrices, and solving
564the linear system is *exactly* the same for the uniprocessor and
565multiprocessor examples. The primary difference between the examples in
566{any}`sec_simple` and
567here is each processor forms only its
568local part of the matrix and vectors in the parallel case.
569
570:::{admonition} Listing: <a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/src/ksp/ksp/tutorials/ex2.c.html">KSP Tutorial src/ksp/ksp/tutorials/ex2.c</a>
571:name: ksp-ex2
572
573```{literalinclude} /../src/ksp/ksp/tutorials/ex2.c
574:end-before: /*TEST
575```
576:::
577
578(sec_cpu_simd)=
579
580## CPU SIMD parallelism
581
582SIMD parallelism occurs most commonly in the Intel advanced vector extensions (AVX) families of instructions (see [Wikipedia](https://en.wikipedia.org/wiki/Advanced_Vector_Extensions)).
583It may be automatically used by the optimizing compiler or in low-level libraries that PETSc uses, such as BLAS
584(see [BLIS](https://github.com/flame/blis)), or rarely,
585directly in PETSc C/C++ code, as in [MatMult_SeqSELL](https://petsc.org/main/src/mat/impls/sell/seq/sell.c.html#MatMult_SeqSELL).
586
587(sec_cpu_openmp)=
588
589## CPU OpenMP parallelism
590
591OpenMP parallelism is thread parallelism. Multiple threads (independent streams of instructions) process data and perform computations on different
592parts of memory that is
593shared (accessible) to all of the threads. The OpenMP model is based on inserting pragmas into code, indicating that a series of instructions
594(often within a loop) can be run in parallel. This is also called a fork-join model of parallelism since much of the code remains sequential and only the
595computationally expensive parts in the 'parallel region' are parallel. Thus, OpenMP makes it relatively easy to add some
596parallelism to a conventional sequential code in a shared memory environment.
597
598POSIX threads (pthreads) is a library that may be called from C/C++. The library contains routines to create, join, and remove threads, plus manage communications and
599synchronizations between threads. Pthreads is rarely used directly in numerical libraries and applications. Sometimes OpenMP is implemented on top of pthreads.
600
601If one adds
602OpenMP parallelism to an MPI code, one must not over-subscribe the hardware resources. For example, if MPI already has one MPI process (rank)
603per hardware core, then
604using four OpenMP threads per MPI process will slow the code down since now one core must switch back and forth between four OpenMP threads.
605
606For application codes that use certain external packages, including BLAS/LAPACK, SuperLU_DIST, MUMPS, MKL, and SuiteSparse, one can build PETSc and these
607packages to take advantage of OpenMP by using the configure option `--with-openmp`. The number of OpenMP threads used in the application can be controlled with
608the PETSc command line option `-omp_num_threads <num>` or the environmental variable `OMP_NUM_THREADS`. Running a PETSc program with `-omp_view` will display the
609number of threads used. The default number is often absurdly high for the given hardware, so we recommend always setting it appropriately.
610
611Users can also put OpenMP pragmas into their own code. However, since standard PETSc is not thread-safe, they should not, in general,
612call PETSc routines from inside the parallel regions.
613
614There is an OpenMP thread-safe subset of PETSc that may be configured for using `--with-threadsafety` (often used along with `--with-openmp` or
615`--download-concurrencykit`). <a href="PETSC_DOC_OUT_ROOT_PLACEHOLDER/src/ksp/ksp/tutorials/ex61f.F90.html">KSP Tutorial ex61f</a> demonstrates
616how this may be used with OpenMP. In this mode, one may have individual OpenMP threads that each manage their own
617(sequential) PETSc objects (each thread can interact only with its own objects). This
618is useful when one has many small systems (or sets of ODEs) that must be integrated in an
619"embarrassingly parallel" fashion on multicore systems.
620
621The ./configure option `--with-openmp-kernels` causes some PETSc numerical kernels to be compiled using OpenMP pragmas to take advantage of multiple cores.
622One must be careful to ensure the number of threads used by each MPI process **times** the number of MPI processes is less than the number of
623cores on the system; otherwise the code will slow down dramatically.
624
625PETSc's MPI-based linear solvers may be accessed from a sequential or non-MPI OpenMP program, see {any}`sec_pcmpi`.
626
627:::{seealso}
628Edward A. Lee, [The Problem with Threads](https://digitalassets.lib.berkeley.edu/techreports/ucb/text/EECS-2006-1.pdf), Technical Report No. UCB/EECS-2006-1 January [[DOI]](https://doi.org/10.1109/MC.2006.180)
62910, 2006
630:::
631
632(sec_gpu_kernels)=
633
634## GPU kernel parallelism
635
636GPUs offer at least two levels of clearly defined parallelism. Kernel-level parallelism is much like SIMD parallelism applied to loops;
637many "iterations" of the loop index run on different hardware in "lock-step".
638PETSc utilizes this parallelism with three similar but slightly different models:
639
640- CUDA, which is provided by NVIDIA and runs on NVIDIA GPUs
641- HIP, provided by AMD, which can, in theory, run on both AMD and NVIDIA GPUs
642- and Kokkos, an open-source package that provides a slightly higher-level programming model to utilize GPU kernels.
643
644To utilize this one configures PETSc with either `--with-cuda` or `--with-hip` and, if they plan to use Kokkos, also `--download-kokkos --download-kokkos-kernels`.
645
646In the GPU programming model that PETSc uses, the GPU memory is distinct from the CPU memory. This means that data that resides on the CPU
647memory must be copied to the GPU (often, this copy is done automatically by the libraries, and the user does not need to manage it)
648if one wishes to use the GPU computational power on it. This memory copy is slow compared to the GPU speed; hence, it is crucial to minimize these copies. This often
649translates to trying to do almost all the computation on the GPU and not constantly switching between computations on the CPU and the GPU on the same data.
650
651PETSc utilizes GPUs by providing vector and matrix classes (Vec and Mat) specifically written to run on the GPU. However, since it is difficult to
652write an entire PETSc code that runs only on the GPU, one can also access and work with (for example, put entries into) the vectors and matrices
653on the CPU. The vector classes
654are `VECCUDA`, `MATAIJCUSPARSE`, `VECKOKKOS`, `MATAIJKOKKOS`, and `VECHIP` (matrices are not yet supported by PETSc with HIP).
655
656More details on using GPUs from PETSc will follow in this document.
657
658(sec_gpu_streams)=
659
660## GPU stream parallelism
661
662Please contribute to this document.
663
664```{raw} latex
665\newpage
666```
667
668# Compiling and Running Programs
669
670The output below illustrates compiling and running a
671PETSc program using MPICH on a macOS laptop. Note that different
672machines will have compilation commands as determined by the
673configuration process. See {any}`sec_writing_application_codes` for
674a discussion about how to compile your PETSc programs. Users who are
675experiencing difficulties linking PETSc programs should refer to the [FAQ](https://petsc.org/release/faq/).
676
677```none
678$ cd $PETSC_DIR/src/ksp/ksp/tutorials
679$ make ex2
680/Users/patrick/petsc/arch-debug/bin/mpicc -o ex2.o -c -g3   -I/Users/patrick/petsc/include -I/Users/patrick/petsc/arch-debug/include `pwd`/ex2.c
681/Users/patrick/petsc/arch-debug/bin/mpicc -g3  -o ex2 ex2.o  -Wl,-rpath,/Users/patrick/petsc/arch-debug/lib -L/Users/patrick/petsc/arch-debug/lib  -lpetsc -lf2clapack -lf2cblas -lmpifort -lgfortran -lgcc_ext.10.5 -lquadmath -lm -lclang_rt.osx -lmpicxx -lc++ -ldl -lmpi -lpmpi -lSystem
682/bin/rm -f ex2.o
683$ $PETSC_DIR/lib/petsc/bin/petscmpiexec -n 1 ./ex2
684Norm of error 0.000156044 iterations 6
685$ $PETSC_DIR/lib/petsc/bin/petscmpiexec -n 2 ./ex2
686Norm of error 0.000411674 iterations 7
687```
688
689(sec_profiling_programs)=
690
691# Profiling Programs
692
693The option
694`-log_view` activates printing of a performance summary, including
695times, floating point operation (flop) rates, and message-passing
696activity. {any}`ch_profiling` provides details about
697profiling, including the interpretation of the output data below.
698This particular example involves
699the solution of a linear system on one processor using GMRES and ILU.
700The low floating point operation (flop) rates in this example are because the code solved a tiny system. We include this example
701merely to demonstrate the ease of extracting performance information.
702
703(listing_exprof)=
704
705```none
706$ $PETSC_DIR/lib/petsc/bin/petscmpiexec -n 1 ./ex1 -n 1000 -pc_type ilu -ksp_type gmres -ksp_rtol 1.e-7 -log_view
707...
708------------------------------------------------------------------------------------------------------------------------
709Event                Count      Time (sec)     Flops                             --- Global ---  --- Stage ----  Total
710                   Max Ratio  Max     Ratio   Max  Ratio  Mess   AvgLen  Reduct  %T %F %M %L %R  %T %F %M %L %R Mflop/s
711------------------------------------------------------------------------------------------------------------------------
712
713VecMDot                1 1.0 3.2830e-06 1.0 2.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  0  5  0  0  0   0  5  0  0  0   609
714VecNorm                3 1.0 4.4550e-06 1.0 6.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  0 14  0  0  0   0 14  0  0  0  1346
715VecScale               2 1.0 4.0110e-06 1.0 2.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  0  5  0  0  0   0  5  0  0  0   499
716VecCopy                1 1.0 3.2280e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
717VecSet                11 1.0 2.5537e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  2  0  0  0  0   2  0  0  0  0     0
718VecAXPY                2 1.0 2.0930e-06 1.0 4.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  0 10  0  0  0   0 10  0  0  0  1911
719VecMAXPY               2 1.0 1.1280e-06 1.0 4.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  0 10  0  0  0   0 10  0  0  0  3546
720VecNormalize           2 1.0 9.3970e-06 1.0 6.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  1 14  0  0  0   1 14  0  0  0   638
721MatMult                2 1.0 1.1177e-05 1.0 9.99e+03 1.0 0.0e+00 0.0e+00 0.0e+00  1 24  0  0  0   1 24  0  0  0   894
722MatSolve               2 1.0 1.9933e-05 1.0 9.99e+03 1.0 0.0e+00 0.0e+00 0.0e+00  1 24  0  0  0   1 24  0  0  0   501
723MatLUFactorNum         1 1.0 3.5081e-05 1.0 4.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  2 10  0  0  0   2 10  0  0  0   114
724MatILUFactorSym        1 1.0 4.4259e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  3  0  0  0  0   3  0  0  0  0     0
725MatAssemblyBegin       1 1.0 8.2015e-08 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
726MatAssemblyEnd         1 1.0 3.3536e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  2  0  0  0  0   2  0  0  0  0     0
727MatGetRowIJ            1 1.0 1.5960e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
728MatGetOrdering         1 1.0 3.9791e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  3  0  0  0  0   3  0  0  0  0     0
729MatView                2 1.0 6.7909e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  5  0  0  0  0   5  0  0  0  0     0
730KSPGMRESOrthog         1 1.0 7.5970e-06 1.0 4.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00  1 10  0  0  0   1 10  0  0  0   526
731KSPSetUp               1 1.0 3.4424e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  2  0  0  0  0   2  0  0  0  0     0
732KSPSolve               1 1.0 2.7264e-04 1.0 3.30e+04 1.0 0.0e+00 0.0e+00 0.0e+00 19 79  0  0  0  19 79  0  0  0   121
733PCSetUp                1 1.0 1.5234e-04 1.0 4.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00 11 10  0  0  0  11 10  0  0  0    26
734PCApply                2 1.0 2.1022e-05 1.0 9.99e+03 1.0 0.0e+00 0.0e+00 0.0e+00  1 24  0  0  0   1 24  0  0  0   475
735------------------------------------------------------------------------------------------------------------------------
736
737Memory usage is given in bytes:
738
739Object Type          Creations   Destructions     Memory  Descendants' Mem.
740Reports information only for process 0.
741
742--- Event Stage 0: Main Stage
743
744              Vector     8              8        76224     0.
745              Matrix     2              2       134212     0.
746       Krylov Solver     1              1        18400     0.
747      Preconditioner     1              1         1032     0.
748           Index Set     3              3        10328     0.
749              Viewer     1              0            0     0.
750========================================================================================================================
751...
752```
753
754(sec_writing_application_codes)=
755
756# Writing C/C++ or Fortran Applications
757
758The examples throughout the library demonstrate the software usage and
759can serve as templates for developing custom applications. We suggest
760that new PETSc users examine programs in the directories
761`$PETSC_DIR/src/<library>/tutorials` where `<library>` denotes any
762of the PETSc libraries (listed in the following section), such as
763`SNES` or `KSP`, `TS`, or `TAO`. The manual pages at
764<https://petsc.org/release/documentation/> provide links (organized by
765routine names and concepts) to the tutorial examples.
766
767To develop an application program that uses PETSc, we suggest the following:
768
769- {ref}`Download <doc_download>` and {ref}`install <doc_install>` PETSc.
770
771- For completely new applications
772
773  > 1. Make a directory for your source code: for example, `mkdir $HOME/application`
774  >
775  > 2. Change to that directory, for
776  >    example, `cd $HOME/application`
777  >
778  > 3. Copy an example in the directory that corresponds to the
779  >    problems of interest into your directory, for
780  >    example, `cp $PETSC_DIR/src/snes/tutorials/ex19.c app.c`
781  >
782  > 4. Select an application build process. The `PETSC_DIR` (and `PETSC_ARCH` if the `--prefix=directoryname`
783  >    option was not used when configuring PETSc) environmental variable(s) must be
784  >    set for any of these approaches.
785  >
786  >    - make (recommended). It uses the [pkg-config](https://en.wikipedia.org/wiki/Pkg-config) tool
787  >      and is the recommended approach. Copy \$PETSC_DIR/share/petsc/Makefile.user or \$PETSC_DIR/share/petsc/Makefile.basic.user
788  >      to your directory, for example, `cp $PETSC_DIR/share/petsc/Makefile.user makefile`
789  >
790  >      Examine the comments in this makefile.
791  >
792  >      Makefile.user uses the [pkg-config](https://en.wikipedia.org/wiki/Pkg-config) tool and is the recommended approach.
793  >
794  >      Use `make app` to compile your program.
795  >
796  >    - CMake. Copy \$PETSC_DIR/share/petsc/CMakeLists.txt to your directory, for example, `cp $PETSC_DIR/share/petsc/CMakeLists.txt CMakeLists.txt`
797  >
798  >      Edit CMakeLists.txt, read the comments on usage, and change the name of the application from ex1 to your application executable name.
799  >
800  > 5. Run the program, for example,
801  >    `./app`
802  >
803  > 6. Start to modify the program to develop your application.
804
805- For adding PETSc to an existing application
806
807  > 1. Start with a working version of your code that you build and run to confirm that it works.
808  >
809  > 2. Upgrade your build process. The `PETSC_DIR` (and `PETSC_ARCH` if the `--prefix=directoryname`
810  >    option was not used when configuring PETSc) environmental variable(s) must be
811  >    set for any of these approaches.
812  >
813  >    - Using make. Update the application makefile to add the appropriate PETSc include
814  >      directories and libraries.
815  >
816  >      - Recommended approach. Examine the comments in \$PETSC_DIR/share/petsc/Makefile.user and transfer selected portions of
817  >        that file to your makefile.
818  >
819  >      - Minimalist. Add the line
820  >
821  >        ```console
822  >        include ${PETSC_DIR}/lib/petsc/conf/variables
823  >        ```
824  >
825  >        to the bottom of your makefile. This will provide a set of PETSc-specific make variables you may use in your makefile. See
826  >        the comments in the file \$PETSC_DIR/share/petsc/Makefile.basic.user for details on the usage.
827  >
828  >      - Simple, but hands the build process over to PETSc's control. Add the lines
829  >
830  >        ```console
831  >        include ${PETSC_DIR}/lib/petsc/conf/variables
832  >        include ${PETSC_DIR}/lib/petsc/conf/rules
833  >        ```
834  >
835  >        to the bottom of your makefile. See the comments in the file \$PETSC_DIR/share/petsc/Makefile.basic.user for details on the usage.
836  >        Since PETSc's rules now control the build process, you will likely need to simplify and remove much of the material that is in
837  >        your makefile.
838  >
839  >      - Not recommended since you must change your makefile for each new configuration/computing system. This approach does not require
840  >        the environmental variable `PETSC_DIR` to be set when building your application since the information will be hardwired in your
841  >        makefile. Run the following command in the PETSc root directory to get the information needed by your makefile:
842  >
843  >        ```console
844  >        $ make getlinklibs getincludedirs getcflags getcxxflags getfortranflags getccompiler getfortrancompiler getcxxcompiler
845  >        ```
846  >
847  >        All the libraries listed need to be linked into your executable, and the
848  >        include directories and flags need to be passed to the compiler(s). Usually,
849  >        this is done by setting `LDFLAGS=<list of library flags and libraries>` and
850  >        `CFLAGS=<list of -I and other flags>` and `FFLAGS=<list of -I and other flags>` etc in your makefile.
851  >
852  >    - Using CMake. Update the application CMakeLists.txt by examining the code and comments in
853  >      \$PETSC_DIR/share/petsc/CMakeLists.txt
854  >
855  > 3. Rebuild your application and ensure it still runs correctly.
856  >
857  > 4. Add a `PetscInitialize()` near the beginning of your code and `PetscFinalize()` near the end with appropriate include commands
858  >    (and use statements in Fortran).
859  >
860  > 5. Rebuild your application and ensure it still runs correctly.
861  >
862  > 6. Slowly start utilizing PETSc functionality in your code, and ensure that your code continues to build and run correctly.
863
864(sec_oo)=
865
866# PETSc's Object-Oriented Design
867
868Though PETSc has a large API, conceptually, it's rather simple.
869There are three abstract basic data objects (classes): index sets, `IS`, vectors, `Vec`, and matrices, `Mat`.
870Plus, a larger number of abstract algorithm objects (classes) starting with: preconditioners, `PC`, Krylov solvers, `KSP`, and so forth.
871
872Let `Object`
873represent any of these objects. Objects are created with
874
875```
876Object obj;
877ObjectCreate(MPI_Comm, &obj);
878```
879
880The object is initially empty, and little can be done with it. A particular implementation of the class is associated with the object by setting the object's "type", where type
881is merely a string name of an implementation class using
882
883```
884ObjectSetType(obj,"ImplementationName");
885```
886
887Some objects support subclasses, which are specializations of the type. These are set with
888
889```
890ObjectNameSetType(obj,"ImplementationSubName");
891```
892
893For example, within `TS` one may do
894
895```
896TS ts;
897TSCreate(PETSC_COMM_WORLD,&ts);
898TSSetType(ts,TSARKIMEX);
899TSARKIMEXSetType(ts,TSARKIMEX3);
900```
901
902The abstract class `TS` can embody any ODE/DAE integrator scheme.
903This example creates an additive Runge-Kutta ODE/DAE IMEX integrator, whose type name is `TSARKIMEX`, using a 3rd-order scheme with an L-stable implicit part,
904whose subtype name is `TSARKIMEX3`.
905
906To allow PETSc objects to be runtime configurable, PETSc objects provide a universal way of selecting types (classes) and subtypes at runtime from
907what is referred to as the PETSc "options database". The code above can be replaced with
908
909```
910TS obj;
911TSCreate(PETSC_COMM_WORLD,&obj);
912TSSetFromOptions(obj);
913```
914
915now, both the type and subtype can be conveniently set from the command line
916
917```console
918$ ./app -ts_type arkimex -ts_arkimex_type 3
919```
920
921The object's type (implementation class) or subclass can also be changed at any time simply by calling `TSSetType()` again (though to override command line options, the call to `TSSetType()` must be made \_after\_ `TSSetFromOptions()`). For example:
922
923```
924// (if set) command line options "override" TSSetType()
925TSSetType(ts, TSGLLE);
926TSSetFromOptions(ts);
927
928// TSSetType() overrides command line options
929TSSetFromOptions(ts);
930TSSetType(ts, TSGLLE);
931```
932
933Since the later call always overrides the earlier call, the second form shown is rarely -- if ever -- used, as it is less flexible than configuring command line settings.
934
935The standard methods on an object are of the general form.
936
937```
938ObjectSetXXX(obj,...);
939ObjectGetXXX(obj,...);
940ObjectYYY(obj,...);
941```
942
943For example
944
945```
946TSSetRHSFunction(obj,...)
947```
948
949Particular types and subtypes of objects may have their own methods, which are given in the form
950
951```
952ObjectNameSetXXX(obj,...);
953ObjectNameGetXXX(obj,...);
954ObjectNameYYY(obj,...);
955```
956
957and
958
959```
960ObjectNameSubNameSetXXX(obj,...);
961ObjectNameSubNameGetXXX(obj,...);
962ObjectNameSubNameYYY(obj,...);
963```
964
965where Name and SubName are the type and subtype names (for example, as above `TSARKIMEX` and `3`. Most "set" operations have options database versions with the same
966names in lower case, separated by underscores, and with the word "set" removed. For example,
967
968```
969KSPGMRESSetRestart(obj,30);
970```
971
972can be set at the command line with
973
974```console
975$ ./app -ksp_gmres_restart 30
976```
977
978A special subset of type-specific methods is ignored if the type does not match the function name. These are usually setter functions that control some aspect specific to the subtype.
979Note that we leveraged this functionality in the MPI example above ({any}`sec_mpi`) by calling `Mat*SetPreallocation()` for a number of different matrix types. As another example,
980
981```
982KSPGMRESSetRestart(obj,30);   // ignored if the type is not KSPGMRES
983```
984
985These allow cleaner application code since it does not have many if statements to avoid inactive methods. That is, one does not need to write code like
986
987```
988if (type == KSPGMRES) {     // unneeded clutter
989  KSPGMRESSetRestart(obj,30);
990}
991```
992
993Many "get" routines give one temporary access to an object's internal data. They are used in the style
994
995```
996XXX xxx;
997ObjectGetXXX(obj,&xxx);
998// use xxx
999ObjectRestoreXXX(obj,&xxx);
1000```
1001
1002Objects obtained with a "get" routine should be returned with a "restore" routine, generally within the same function. Objects obtained with a "create" routine should be freed
1003with a "destroy" routine.
1004
1005There may be variants of the "get" routines that give more limited access to the obtained object. For example,
1006
1007```
1008const PetscScalar *x;
1009
1010// specialized variant of VecGetArray()
1011VecGetArrayRead(vec, &x);
1012// one can read but not write with x[]
1013PetscReal y = 2*x[0];
1014// don't forget to restore x after you are done with it
1015VecRestoreArrayRead(vec, &x);
1016```
1017
1018Objects can be displayed (in a large number of ways) with
1019
1020```
1021ObjectView(obj,PetscViewer viewer);
1022ObjectViewFromOptions(obj,...);
1023```
1024
1025Where `PetscViewer` is an abstract object that can represent standard output, an ASCII or binary file, a graphical window, etc. The second
1026variant allows the user to delay until runtime the decision of what viewer and format to use to view the object or if to view the object at all.
1027
1028Objects are destroyed with
1029
1030```
1031ObjectDestroy(&obj)
1032```
1033
1034:::{figure} /images/manual/objectlife.svg
1035:name: fig_objectlife
1036
1037Sample lifetime of a PETSc object
1038:::
1039
1040## User Callbacks
1041
1042The user may wish to override or provide custom functionality in many situations. This is handled via callbacks, which the library will call at the appropriate time. The most general way to apply a callback has this form:
1043
1044```
1045ObjectCallbackSetter(obj, callbackfunction(), void *ctx, contextdestroy(void *ctx));
1046```
1047
1048where `ObjectCallbackSetter()` is a callback setter such as `SNESSetFunction()`. `callbackfunction()` is what will be called
1049by the library, `ctx` is an optional data structure (array, struct, PETSc object) that is used by `callbackfunction()`
1050and `contextdestroy(void *ctx)` is an optional function that will be called when `obj` is destroyed to free
1051anything in `ctx`. The use of the `contextdestroy()` allows users to "set and forget"
1052data structures that will not be needed elsewhere but still need to be deleted when no longer needed. Here is an example of the use of a full-fledged callback
1053
1054```
1055TS              ts;
1056TSMonitorLGCtx *ctx;
1057
1058TSMonitorLGCtxCreate(..., &ctx)
1059TSMonitorSet(ts, TSMonitorLGTimeStep, ctx, (PetscCtxDestroyFn *)TSMonitorLGCtxDestroy);
1060TSSolve(ts);
1061```
1062
1063Occasionally, routines to set callback functions take additional data objects that will be used by the object but are not context data for the function. For example,
1064
1065```
1066SNES obj;
1067Vec  r;
1068void *ctx;
1069
1070SNESSetFunction(snes, r, UserApplyFunction(SNES,Vec,Vec,void *ctx), ctx);
1071```
1072
1073The `r` vector is an optional argument provided by the user, which will be used as work-space by `SNES`. Note that this callback does not provide a way for the user
1074to have the `ctx` destroyed when the `SNES` object is destroyed; the users must ensure that they free it at an appropriate time. There is no logic to the various ways
1075PETSc accepts callback functions in different places in the code.
1076
1077See {any}`fig_taocallbacks` for a cartoon on callbacks in `Tao`.
1078
1079(sec_directory)=
1080
1081# Directory Structure
1082
1083We conclude this introduction with an overview of the organization of
1084the PETSc software. The root directory of PETSc contains the following
1085directories:
1086
1087- `doc` The source code and Python scripts for building the website and documentation
1088
1089- `lib/petsc/conf` - Base PETSc configuration files that define the standard
1090  make variables and rules used by PETSc
1091
1092- `include` - All include files for PETSc that are visible to the
1093  user.
1094
1095- `include/petsc/finclude` - PETSc Fortran include files.
1096
1097- `include/petsc/private` - Private PETSc include files that should
1098  *not* need to be used by application programmers.
1099
1100- `share` - Some small test matrices and other data files
1101
1102- `src` - The source code for all PETSc libraries, which currently
1103  includes
1104
1105  - `vec` - vectors,
1106
1107    - `is` - index sets,
1108
1109  - `mat` - matrices,
1110
1111  - `ksp` - complete linear equations solvers,
1112
1113    - `ksp` - Krylov subspace accelerators,
1114    - `pc` - preconditioners,
1115
1116  - `snes` - nonlinear solvers
1117
1118  - `ts` - ODE/DAE solvers and timestepping,
1119
1120  - `tao` - optimizers,
1121
1122  - `dm` - data management between meshes and solvers, vectors, and
1123    matrices,
1124
1125  - `sys` - general system-related routines,
1126
1127    - `logging` - PETSc logging and profiling routines,
1128
1129    - `classes` - low-level classes
1130
1131      - `draw` - simple graphics,
1132      - `viewer` - a mechanism for printing and visualizing PETSc
1133        objects,
1134      - `bag` - mechanism for saving and loading from disk user
1135        data stored in C structs.
1136      - `random` - random number generators.
1137
1138Each PETSc source code library directory has the following subdirectories:
1139
1140- `tutorials` - Programs designed to teach users about PETSc.
1141  These codes can serve as templates for applications.
1142- `tests` - Programs designed for thorough testing of PETSc. As
1143  such, these codes are not intended for examination by users.
1144- `interface` - Provides the abstract base classes for the objects.
1145  The code here does not know about particular implementations and does not perform
1146  operations on the underlying numerical data.
1147- `impls` - Source code for one or more implementations of the class for particular
1148  data structures or algorithms.
1149- `utils` - Utility routines. The source here may know about the
1150  implementations, but ideally, will not know about implementations for
1151  other components.
1152
1153```{rubric} Footnotes
1154```
1155
1156[^debug-footnote]: Configure PETSc with `--with-debugging`.
1157
1158```{eval-rst}
1159.. bibliography:: /petsc.bib
1160   :filter: docname in docnames
1161```
1162