xref: /libCEED/README.md (revision 53ee81eed7fe6a54158b819b478dad66ca478dc3)
1# libCEED: Efficient Extensible Discretization
2
3[![GitHub Actions][github-badge]][github-link]
4[![GitLab-CI][gitlab-badge]][gitlab-link]
5[![Code coverage][codecov-badge]][codecov-link]
6[![BSD-2-Clause][license-badge]][license-link]
7[![Documentation][doc-badge]][doc-link]
8[![JOSS paper][joss-badge]][joss-link]
9[![Binder][binder-badge]][binder-link]
10
11## Summary and Purpose
12
13libCEED provides fast algebra for element-based discretizations, designed for performance portability, run-time flexibility, and clean embedding in higher level libraries and applications.
14It offers a C99 interface as well as bindings for Fortran, Python, Julia, and Rust.
15While our focus is on high-order finite elements, the approach is mostly algebraic and thus applicable to other discretizations in factored form, as explained in the [user manual](https://libceed.org/en/latest/) and API implementation portion of the [documentation](https://libceed.org/en/latest/api/).
16
17One of the challenges with high-order methods is that a global sparse matrix is no longer a good representation of a high-order linear operator, both with respect to the FLOPs needed for its evaluation, as well as the memory transfer needed for a matvec.
18Thus, high-order methods require a new "format" that still represents a linear (or more generally non-linear) operator, but not through a sparse matrix.
19
20The goal of libCEED is to propose such a format, as well as supporting implementations and data structures, that enable efficient operator evaluation on a variety of computational device types (CPUs, GPUs, etc.).
21This new operator description is based on algebraically [factored form](https://libceed.org/en/latest/libCEEDapi/#finite-element-operator-decomposition), which is easy to incorporate in a wide variety of applications, without significant refactoring of their own discretization infrastructure.
22
23The repository is part of the [CEED software suite](http://ceed.exascaleproject.org/software/), a collection of software benchmarks, miniapps, libraries and APIs for efficient exascale discretizations based on high-order finite element and spectral element methods.
24See <http://github.com/ceed> for more information and source code availability.
25
26The CEED research is supported by the [Exascale Computing Project](https://exascaleproject.org/exascale-computing-project) (17-SC-20-SC), a collaborative effort of two U.S. Department of Energy organizations (Office of Science and the National Nuclear Security Administration) responsible for the planning and preparation of a [capable exascale ecosystem](https://exascaleproject.org/what-is-exascale), including software, applications, hardware, advanced system engineering and early testbed platforms, in support of the nation’s exascale computing imperative.
27
28For more details on the CEED API see the [user manual](https://libceed.org/en/latest/).
29
30<!-- getting-started-inclusion -->
31
32## Building
33
34The CEED library, `libceed`, is a C99 library with no required dependencies, and with Fortran, Python, Julia, and Rust interfaces.
35It can be built using:
36
37```console
38$ make
39```
40
41or, with optimization flags:
42
43```console
44$ make OPT='-O3 -march=skylake-avx512 -ffp-contract=fast'
45```
46
47These optimization flags are used by all languages (C, C++, Fortran) and this makefile variable can also be set for testing and examples (below).
48
49The library attempts to automatically detect support for the AVX instruction set using gcc-style compiler options for the host.
50Support may need to be manually specified via:
51
52```console
53$ make AVX=1
54```
55
56or:
57
58```console
59$ make AVX=0
60```
61
62if your compiler does not support gcc-style options, if you are cross compiling, etc.
63
64To enable CUDA support, add `CUDA_DIR=/opt/cuda` or an appropriate directory to your `make` invocation.
65To enable HIP support, add `ROCM_DIR=/opt/rocm` or an appropriate directory.
66To enable SYCL support, add `SYCL_DIR=/opt/sycl` or an appropriate directory.
67Note that SYCL backends require building with oneAPI compilers as well:
68
69```console
70$ . /opt/intel/oneapi/setvars.sh
71$ make SYCL_DIR=/opt/intel/oneapi/compiler/latest/linux SYCLCXX=icpx CC=icx CXX=icpx
72```
73
74The library can be configured for host applications which use OpenMP paralellism via:
75
76```console
77$ make OPENMP=1
78```
79
80which will allow operators created and applied from different threads inside an `omp parallel` region.
81
82To store these or other arguments as defaults for future invocations of `make`, use:
83
84```console
85$ make configure CUDA_DIR=/usr/local/cuda ROCM_DIR=/opt/rocm OPT='-O3 -march=znver2'
86```
87
88which stores these variables in `config.mk`.
89
90### WebAssembly
91
92libCEED can be built for WASM using [Emscripten](https://emscripten.org). For example, one can build the library and run a standalone WASM executable using
93
94``` console
95$ emmake make build/ex2-surface.wasm
96$ wasmer build/ex2-surface.wasm -- -s 200000
97```
98
99## Additional Language Interfaces
100
101The Fortran interface is built alongside the library automatically.
102
103Python users can install using:
104
105```console
106$ pip install libceed
107```
108
109or in a clone of the repository via `pip install .`.
110
111Julia users can install using:
112
113```console
114$ julia
115julia> ]
116pkg> add LibCEED
117```
118
119See the [LibCEED.jl documentation](http://ceed.exascaleproject.org/libCEED-julia-docs/dev/) for more information.
120
121Rust users can include libCEED via `Cargo.toml`:
122
123```toml
124[dependencies]
125libceed = "0.12.0"
126```
127
128See the [Cargo documentation](https://doc.rust-lang.org/cargo/reference/specifying-dependencies.html#specifying-dependencies-from-git-repositories) for details.
129
130## Testing
131
132The test suite produces [TAP](https://testanything.org) output and is run by:
133
134```console
135$ make test
136```
137
138or, using the `prove` tool distributed with Perl (recommended):
139
140```console
141$ make prove
142```
143
144## Backends
145
146There are multiple supported backends, which can be selected at runtime in the examples:
147
148| CEED resource              | Backend                                           | Deterministic Capable |
149| :---                       | :---                                              | :---:                 |
150||
151| **CPU Native**             |
152| `/cpu/self/ref/serial`     | Serial reference implementation                   | Yes                   |
153| `/cpu/self/ref/blocked`    | Blocked reference implementation                  | Yes                   |
154| `/cpu/self/opt/serial`     | Serial optimized C implementation                 | Yes                   |
155| `/cpu/self/opt/blocked`    | Blocked optimized C implementation                | Yes                   |
156| `/cpu/self/avx/serial`     | Serial AVX implementation                         | Yes                   |
157| `/cpu/self/avx/blocked`    | Blocked AVX implementation                        | Yes                   |
158||
159| **CPU Valgrind**           |
160| `/cpu/self/memcheck/*`     | Memcheck backends, undefined value checks         | Yes                   |
161||
162| **CPU LIBXSMM**            |
163| `/cpu/self/xsmm/serial`    | Serial LIBXSMM implementation                     | Yes                   |
164| `/cpu/self/xsmm/blocked`   | Blocked LIBXSMM implementation                    | Yes                   |
165||
166| **CUDA Native**            |
167| `/gpu/cuda/ref`            | Reference pure CUDA kernels                       | Yes                   |
168| `/gpu/cuda/shared`         | Optimized pure CUDA kernels using shared memory   | Yes                   |
169| `/gpu/cuda/gen`            | Optimized pure CUDA kernels using code generation | No                    |
170||
171| **HIP Native**             |
172| `/gpu/hip/ref`             | Reference pure HIP kernels                        | Yes                   |
173| `/gpu/hip/shared`          | Optimized pure HIP kernels using shared memory    | Yes                   |
174| `/gpu/hip/gen`             | Optimized pure HIP kernels using code generation  | No                    |
175||
176| **SYCL Native**            |
177| `/gpu/sycl/ref`            | Reference pure SYCL kernels                       | Yes                   |
178| `/gpu/sycl/shared`         | Optimized pure SYCL kernels using shared memory   | Yes                   |
179||
180| **MAGMA**                  |
181| `/gpu/cuda/magma`          | CUDA MAGMA kernels                                | No                    |
182| `/gpu/cuda/magma/det`      | CUDA MAGMA kernels                                | Yes                   |
183| `/gpu/hip/magma`           | HIP MAGMA kernels                                 | No                    |
184| `/gpu/hip/magma/det`       | HIP MAGMA kernels                                 | Yes                   |
185||
186
187The `/cpu/self/*/serial` backends process one element at a time and are intended for meshes with a smaller number of high order elements.
188The `/cpu/self/*/blocked` backends process blocked batches of eight interlaced elements and are intended for meshes with higher numbers of elements.
189
190The `/cpu/self/ref/*` backends are written in pure C and provide basic functionality.
191
192The `/cpu/self/opt/*` backends are written in pure C and use partial e-vectors to improve performance.
193
194The `/cpu/self/avx/*` backends rely upon AVX instructions to provide vectorized CPU performance.
195
196The `/cpu/self/memcheck/*` backends rely upon the [Valgrind](https://valgrind.org/) Memcheck tool to help verify that user QFunctions have no undefined values.
197To use, run your code with Valgrind and the Memcheck backends, e.g. `valgrind ./build/ex1 -ceed /cpu/self/ref/memcheck`.
198A 'development' or 'debugging' version of Valgrind with headers is required to use this backend.
199This backend can be run in serial or blocked mode and defaults to running in the serial mode if `/cpu/self/memcheck` is selected at runtime.
200
201The `/cpu/self/xsmm/*` backends rely upon the [LIBXSMM](https://github.com/libxsmm/libxsmm) package to provide vectorized CPU performance.
202If linking MKL and LIBXSMM is desired but the Makefile is not detecting `MKLROOT`, linking libCEED against MKL can be forced by setting the environment variable `MKL=1`.
203The LIBXSMM `main` development branch from 7 April 2024 or newer is required.
204
205The `/gpu/cuda/*` backends provide GPU performance strictly using CUDA.
206
207The `/gpu/hip/*` backends provide GPU performance strictly using HIP.
208They are based on the `/gpu/cuda/*` backends.
209ROCm version 4.2 or newer is required.
210
211The `/gpu/sycl/*` backends provide GPU performance strictly using SYCL.
212They are based on the `/gpu/cuda/*` and `/gpu/hip/*` backends.
213
214The `/gpu/*/magma/*` backends rely upon the [MAGMA](https://bitbucket.org/icl/magma) package.
215To enable the MAGMA backends, the environment variable `MAGMA_DIR` must point to the top-level MAGMA directory, with the MAGMA library located in `$(MAGMA_DIR)/lib/`.
216By default, `MAGMA_DIR` is set to `../magma`; to build the MAGMA backends with a MAGMA installation located elsewhere, create a link to `magma/` in libCEED's parent directory, or set `MAGMA_DIR` to the proper location.
217MAGMA version 2.5.0 or newer is required.
218Currently, each MAGMA library installation is only built for either CUDA or HIP.
219The corresponding set of libCEED backends (`/gpu/cuda/magma/*` or `/gpu/hip/magma/*`) will automatically be built for the version of the MAGMA library found in `MAGMA_DIR`.
220
221Users can specify a device for all CUDA, HIP, and MAGMA backends through adding `:device_id=#` after the resource name.
222For example:
223
224> - `/gpu/cuda/gen:device_id=1`
225
226Bit-for-bit reproducibility is important in some applications.
227However, some libCEED backends use non-deterministic operations, such as `atomicAdd` for increased performance.
228The backends which are capable of generating reproducible results, with the proper compilation options, are highlighted in the list above.
229
230<!-- getting-started-exclusion -->
231
232## Examples
233
234libCEED comes with several examples of its usage, ranging from standalone C codes in the `/examples/ceed` directory to examples based on external packages, such as MFEM, PETSc, and Nek5000.
235Nek5000 v18.0 or greater is required.
236
237To build the examples, set the `MFEM_DIR`, `PETSC_DIR` (and optionally `PETSC_ARCH`), and `NEK5K_DIR` variables and run:
238
239```console
240$ cd examples/
241```
242
243<!-- running-examples-inclusion -->
244
245```console
246# libCEED examples on CPU and GPU
247$ cd ceed/
248$ make
249$ ./ex1-volume -ceed /cpu/self
250$ ./ex1-volume -ceed /gpu/cuda
251$ ./ex2-surface -ceed /cpu/self
252$ ./ex2-surface -ceed /gpu/cuda
253$ cd ..
254
255# MFEM+libCEED examples on CPU and GPU
256$ cd mfem/
257$ make
258$ ./bp1 -ceed /cpu/self -no-vis
259$ ./bp3 -ceed /gpu/cuda -no-vis
260$ cd ..
261
262# Nek5000+libCEED examples on CPU and GPU
263$ cd nek/
264$ make
265$ ./nek-examples.sh -e bp1 -ceed /cpu/self -b 3
266$ ./nek-examples.sh -e bp3 -ceed /gpu/cuda -b 3
267$ cd ..
268
269# PETSc+libCEED examples on CPU and GPU
270$ cd petsc/
271$ make
272$ ./bps -problem bp1 -ceed /cpu/self
273$ ./bps -problem bp2 -ceed /gpu/cuda
274$ ./bps -problem bp3 -ceed /cpu/self
275$ ./bps -problem bp4 -ceed /gpu/cuda
276$ ./bps -problem bp5 -ceed /cpu/self
277$ ./bps -problem bp6 -ceed /gpu/cuda
278$ cd ..
279
280$ cd petsc/
281$ make
282$ ./bpsraw -problem bp1 -ceed /cpu/self
283$ ./bpsraw -problem bp2 -ceed /gpu/cuda
284$ ./bpsraw -problem bp3 -ceed /cpu/self
285$ ./bpsraw -problem bp4 -ceed /gpu/cuda
286$ ./bpsraw -problem bp5 -ceed /cpu/self
287$ ./bpsraw -problem bp6 -ceed /gpu/cuda
288$ cd ..
289
290$ cd petsc/
291$ make
292$ ./bpssphere -problem bp1 -ceed /cpu/self
293$ ./bpssphere -problem bp2 -ceed /gpu/cuda
294$ ./bpssphere -problem bp3 -ceed /cpu/self
295$ ./bpssphere -problem bp4 -ceed /gpu/cuda
296$ ./bpssphere -problem bp5 -ceed /cpu/self
297$ ./bpssphere -problem bp6 -ceed /gpu/cuda
298$ cd ..
299
300$ cd petsc/
301$ make
302$ ./area -problem cube -ceed /cpu/self -degree 3
303$ ./area -problem cube -ceed /gpu/cuda -degree 3
304$ ./area -problem sphere -ceed /cpu/self -degree 3 -dm_refine 2
305$ ./area -problem sphere -ceed /gpu/cuda -degree 3 -dm_refine 2
306
307$ cd fluids/
308$ make
309$ ./navierstokes -ceed /cpu/self -degree 1
310$ ./navierstokes -ceed /gpu/cuda -degree 1
311$ cd ..
312
313$ cd solids/
314$ make
315$ ./elasticity -ceed /cpu/self -mesh [.exo file] -degree 2 -E 1 -nu 0.3 -problem Linear -forcing mms
316$ ./elasticity -ceed /gpu/cuda -mesh [.exo file] -degree 2 -E 1 -nu 0.3 -problem Linear -forcing mms
317$ cd ..
318```
319
320For the last example shown, sample meshes to be used in place of `[.exo file]` can be found at <https://github.com/jeremylt/ceedSampleMeshes>
321
322The above code assumes a GPU-capable machine with the CUDA backends enabled.
323Depending on the available backends, other CEED resource specifiers can be provided with the `-ceed` option.
324Other command line arguments can be found in [examples/petsc](https://github.com/CEED/libCEED/blob/main/examples/petsc/README.md).
325
326<!-- running-examples-exclusion -->
327
328## Benchmarks
329
330A sequence of benchmarks for all enabled backends can be run using:
331
332```console
333$ make benchmarks
334```
335
336The results from the benchmarks are stored inside the `benchmarks/` directory and they can be viewed using the commands (requires python with matplotlib):
337
338```console
339$ cd benchmarks
340$ python postprocess-plot.py petsc-bps-bp1-*-output.txt
341$ python postprocess-plot.py petsc-bps-bp3-*-output.txt
342```
343
344Using the `benchmarks` target runs a comprehensive set of benchmarks which may take some time to run.
345Subsets of the benchmarks can be run using the scripts in the `benchmarks` folder.
346
347For more details about the benchmarks, see the `benchmarks/README.md` file.
348
349## Install
350
351To install libCEED, run:
352
353```console
354$ make install prefix=/path/to/install/dir
355```
356
357or (e.g., if creating packages):
358
359```console
360$ make install prefix=/usr DESTDIR=/packaging/path
361```
362
363To build and install in separate steps, run:
364
365```console
366$ make for_install=1 prefix=/path/to/install/dir
367$ make install prefix=/path/to/install/dir
368```
369
370The usual variables like `CC` and `CFLAGS` are used, and optimization flags for all languages can be set using the likes of `OPT='-O3 -march=native'`.
371Use `STATIC=1` to build static libraries (`libceed.a`).
372
373To install libCEED for Python, run:
374
375```console
376$ pip install libceed
377```
378
379with the desired setuptools options, such as `--user`.
380
381### pkg-config
382
383In addition to library and header, libCEED provides a [pkg-config](https://en.wikipedia.org/wiki/Pkg-config) file that can be used to easily compile and link.
384[For example](https://people.freedesktop.org/~dbn/pkg-config-guide.html#faq), if `$prefix` is a standard location or you set the environment variable `PKG_CONFIG_PATH`:
385
386```console
387$ cc `pkg-config --cflags --libs ceed` -o myapp myapp.c
388```
389
390will build `myapp` with libCEED.
391This can be used with the source or installed directories.
392Most build systems have support for pkg-config.
393
394## Contact
395
396You can reach the libCEED team by emailing [ceed-users@llnl.gov](mailto:ceed-users@llnl.gov) or by leaving a comment in the [issue tracker](https://github.com/CEED/libCEED/issues).
397
398## How to Cite
399
400If you utilize libCEED please cite:
401
402```bibtex
403@article{libceed-joss-paper,
404  author       = {
405    Brown, Jed and
406    Abdelfattah, Ahmad and
407    Barra, Valeria and
408    Beams, Natalie and
409    Camier, Jean-Sylvain and
410    Dobrev, Veselin and
411    Dudouit, Yohann and
412    Ghaffari, Leila and
413    Kolev, Tzanio and
414    Medina, David and
415    Pazner, Will and
416    Ratnayaka, Thilina and
417    Thompson, Jeremy L. and
418    Tomov, Stan
419  },
420  title        = {{libCEED}: Fast algebra for high-order element-based discretizations},
421  journal      = {Journal of Open Source Software},
422  year         = {2021},
423  publisher    = {The Open Journal},
424  volume       = {6},
425  number       = {63},
426  pages        = {2945},
427  doi          = {10.21105/joss.02945}
428}
429```
430
431The archival copy of the libCEED user manual is maintained on [Zenodo](https://doi.org/10.5281/zenodo.4302736).
432To cite the user manual:
433
434```bibtex
435@misc{libceed-user-manual,
436  author       = {
437    Abdelfattah, Ahmad and
438    Barra, Valeria and
439    Beams, Natalie and
440    Brown, Jed and
441    Camier, Jean-Sylvain and
442    Dobrev, Veselin and
443    Dudouit, Yohann and
444    Ghaffari, Leila and
445    Grimberg, Sebastian and
446    Kolev, Tzanio and
447    Medina, David and
448    Pazner, Will and
449    Ratnayaka, Thilina and
450    Shakeri, Rezgar and
451    Thompson, Jeremy L. and
452    Tomov, Stanimire and
453    Wright III, James
454  },
455  title        = {{libCEED} User Manual},
456  month        = nov,
457  year         = 2023,
458  publisher    = {Zenodo},
459  version      = {0.12.0},
460  doi          = {10.5281/zenodo.10062388}
461}
462```
463
464For libCEED's Python interface please cite:
465
466```bibtex
467@InProceedings{libceed-scipy,
468  author    = {
469    Barra, Valeria and
470    Brown, Jed and
471    Thompson, Jeremy L. and
472    Dudouit, Yohann
473  },
474  title     = {{H}igh-performance operator evaluations with ease of use: {libCEED}'s {P}ython interface},
475  booktitle = {{P}roceedings of the 19th {P}ython in {S}cience {C}onference},
476  pages     = {85 - 90},
477  year      = {2020},
478  editor    = {{M}eghann {A}garwal and {C}hris {C}alloway and {D}illon {N}iederhut and {D}avid {S}hupe},
479  doi       = {10.25080/Majora-342d178e-00c}
480}
481```
482
483The BibTeX entries for these references can be found in the `doc/bib/references.bib` file.
484
485## Copyright
486
487The following copyright applies to each file in the CEED software suite, unless otherwise stated in the file:
488
489> Copyright (c) 2017-2026, Lawrence Livermore National Security, LLC and other CEED contributors.
490> All rights reserved.
491
492See files LICENSE and NOTICE for details.
493
494[github-badge]: https://github.com/CEED/libCEED/workflows/C/Fortran/badge.svg
495[github-link]: https://github.com/CEED/libCEED/actions
496[gitlab-badge]: https://gitlab.com/libceed/libCEED/badges/main/pipeline.svg?key_text=GitLab-CI
497[gitlab-link]: https://gitlab.com/libceed/libCEED/-/pipelines?page=1&scope=all&ref=main
498[codecov-badge]: https://codecov.io/gh/CEED/libCEED/branch/main/graphs/badge.svg
499[codecov-link]: https://codecov.io/gh/CEED/libCEED/
500[license-badge]: https://img.shields.io/badge/License-BSD%202--Clause-orange.svg
501[license-link]: https://opensource.org/licenses/BSD-2-Clause
502[doc-badge]: https://readthedocs.org/projects/libceed/badge/?version=latest
503[doc-link]: https://libceed.org/en/latest/?badge=latest
504[joss-badge]: https://joss.theoj.org/papers/10.21105/joss.02945/status.svg
505[joss-link]: https://doi.org/10.21105/joss.02945
506[binder-badge]: http://mybinder.org/badge_logo.svg
507[binder-link]: https://mybinder.org/v2/gh/CEED/libCEED/main?urlpath=lab/tree/examples/python/tutorial-0-ceed.ipynb
508