xref: /libCEED/include/ceed/jit-source/cuda/cuda-ref-restriction-offset.h (revision fd831f258b694b9857c348ebef72fcfdbf8a8f6b)
1 // Copyright (c) 2017-2024, Lawrence Livermore National Security, LLC and other CEED contributors.
2 // All Rights Reserved. See the top-level LICENSE and NOTICE files for details.
3 //
4 // SPDX-License-Identifier: BSD-2-Clause
5 //
6 // This file is part of CEED:  http://github.com/ceed
7 
8 /// @file
9 /// Internal header for CUDA offset element restriction kernels
10 #include <ceed/types.h>
11 
12 //------------------------------------------------------------------------------
13 // L-vector -> E-vector, standard (with offsets)
14 //------------------------------------------------------------------------------
15 extern "C" __global__ void OffsetNoTranspose(const CeedInt *__restrict__ indices, const CeedScalar *__restrict__ u, CeedScalar *__restrict__ v) {
16   for (CeedInt node = blockIdx.x * blockDim.x + threadIdx.x; node < RSTR_NUM_ELEM * RSTR_ELEM_SIZE; node += blockDim.x * gridDim.x) {
17     const CeedInt ind      = indices[node];
18     const CeedInt loc_node = node % RSTR_ELEM_SIZE;
19     const CeedInt elem     = node / RSTR_ELEM_SIZE;
20 
21     for (CeedInt comp = 0; comp < RSTR_NUM_COMP; comp++) {
22       v[loc_node + comp * RSTR_ELEM_SIZE * RSTR_NUM_ELEM + elem * RSTR_ELEM_SIZE] = u[ind + comp * RSTR_COMP_STRIDE];
23     }
24   }
25 }
26 
27 //------------------------------------------------------------------------------
28 // E-vector -> L-vector, standard (with offsets)
29 //------------------------------------------------------------------------------
30 #if !USE_DETERMINISTIC
31 extern "C" __global__ void OffsetTranspose(const CeedInt *__restrict__ indices, const CeedScalar *__restrict__ u, CeedScalar *__restrict__ v) {
32   for (CeedInt node = blockIdx.x * blockDim.x + threadIdx.x; node < RSTR_NUM_ELEM * RSTR_ELEM_SIZE; node += blockDim.x * gridDim.x) {
33     const CeedInt ind      = indices[node];
34     const CeedInt loc_node = node % RSTR_ELEM_SIZE;
35     const CeedInt elem     = node / RSTR_ELEM_SIZE;
36 
37     for (CeedInt comp = 0; comp < RSTR_NUM_COMP; comp++) {
38       atomicAdd(&v[ind + comp * RSTR_COMP_STRIDE], u[loc_node + comp * RSTR_ELEM_SIZE * RSTR_NUM_ELEM + elem * RSTR_ELEM_SIZE]);
39     }
40   }
41 }
42 #else
43 extern "C" __global__ void OffsetTranspose(const CeedInt *__restrict__ l_vec_indices, const CeedInt *__restrict__ t_indices,
44                                            const CeedInt *__restrict__ t_offsets, const CeedScalar *__restrict__ u, CeedScalar *__restrict__ v) {
45   CeedScalar value[RSTR_NUM_COMP];
46 
47   for (CeedInt i = blockIdx.x * blockDim.x + threadIdx.x; i < RSTR_NUM_NODES; i += blockDim.x * gridDim.x) {
48     const CeedInt ind     = l_vec_indices[i];
49     const CeedInt range_1 = t_offsets[i];
50     const CeedInt range_N = t_offsets[i + 1];
51 
52     for (CeedInt comp = 0; comp < RSTR_NUM_COMP; comp++) value[comp] = 0.0;
53 
54     for (CeedInt j = range_1; j < range_N; j++) {
55       const CeedInt t_ind    = t_indices[j];
56       const CeedInt loc_node = t_ind % RSTR_ELEM_SIZE;
57       const CeedInt elem     = t_ind / RSTR_ELEM_SIZE;
58 
59       for (CeedInt comp = 0; comp < RSTR_NUM_COMP; comp++) {
60         value[comp] += u[loc_node + comp * RSTR_ELEM_SIZE * RSTR_NUM_ELEM + elem * RSTR_ELEM_SIZE];
61       }
62     }
63 
64     for (CeedInt comp = 0; comp < RSTR_NUM_COMP; comp++) v[ind + comp * RSTR_COMP_STRIDE] += value[comp];
65   }
66 }
67 #endif
68