xref: /libCEED/backends/cuda/ceed-cuda-compile.cpp (revision a491a57ee1186d4a57a443cd84e0df1f56768bf7)
1 // Copyright (c) 2017-2024, Lawrence Livermore National Security, LLC and other CEED contributors.
2 // All Rights Reserved. See the top-level LICENSE and NOTICE files for details.
3 //
4 // SPDX-License-Identifier: BSD-2-Clause
5 //
6 // This file is part of CEED:  http://github.com/ceed
7 
8 #include "ceed-cuda-compile.h"
9 
10 #include <ceed.h>
11 #include <ceed/backend.h>
12 #include <ceed/jit-tools.h>
13 #include <cuda_runtime.h>
14 #include <nvrtc.h>
15 #include <stdarg.h>
16 #include <string.h>
17 
18 #include <sstream>
19 
20 #include "ceed-cuda-common.h"
21 
22 #define CeedChk_Nvrtc(ceed, x)                                                                              \
23   do {                                                                                                      \
24     nvrtcResult result = static_cast<nvrtcResult>(x);                                                       \
25     if (result != NVRTC_SUCCESS) return CeedError((ceed), CEED_ERROR_BACKEND, nvrtcGetErrorString(result)); \
26   } while (0)
27 
28 #define CeedCallNvrtc(ceed, ...)  \
29   do {                            \
30     int ierr_q_ = __VA_ARGS__;    \
31     CeedChk_Nvrtc(ceed, ierr_q_); \
32   } while (0)
33 
34 //------------------------------------------------------------------------------
35 // Compile CUDA kernel
36 //------------------------------------------------------------------------------
37 int CeedCompile_Cuda(Ceed ceed, const char *source, CUmodule *module, const CeedInt num_defines, ...) {
38   size_t                ptx_size;
39   char                 *ptx;
40   const char           *jit_defs_path, *jit_defs_source;
41   const int             num_opts            = 4;
42   CeedInt               num_jit_source_dirs = 0;
43   const char          **opts;
44   nvrtcProgram          prog;
45   struct cudaDeviceProp prop;
46   Ceed_Cuda            *ceed_data;
47 
48   cudaFree(0);  // Make sure a Context exists for nvrtc
49 
50   std::ostringstream code;
51 
52   // Get kernel specific options, such as kernel constants
53   if (num_defines > 0) {
54     va_list args;
55     va_start(args, num_defines);
56     char *name;
57     int   val;
58 
59     for (int i = 0; i < num_defines; i++) {
60       name = va_arg(args, char *);
61       val  = va_arg(args, int);
62       code << "#define " << name << " " << val << "\n";
63     }
64     va_end(args);
65   }
66 
67   // Standard libCEED definitions for CUDA backends
68   CeedCallBackend(CeedGetJitAbsolutePath(ceed, "ceed/jit-source/cuda/cuda-jit.h", &jit_defs_path));
69   {
70     char *source;
71 
72     CeedCallBackend(CeedLoadSourceToBuffer(ceed, jit_defs_path, &source));
73     jit_defs_source = source;
74   }
75   code << jit_defs_source;
76   code << "\n\n";
77   CeedCallBackend(CeedFree(&jit_defs_path));
78   CeedCallBackend(CeedFree(&jit_defs_source));
79 
80   // Non-macro options
81   CeedCallBackend(CeedCalloc(num_opts, &opts));
82   opts[0] = "-default-device";
83   CeedCallBackend(CeedGetData(ceed, &ceed_data));
84   CeedCallCuda(ceed, cudaGetDeviceProperties(&prop, ceed_data->device_id));
85   std::string arch_arg =
86 #if CUDA_VERSION >= 11010
87       // NVRTC used to support only virtual architectures through the option
88       // -arch, since it was only emitting PTX. It will now support actual
89       // architectures as well to emit SASS.
90       // https://docs.nvidia.com/cuda/cuda-c-best-practices-guide/index.html#dynamic-code-generation
91       "-arch=sm_"
92 #else
93       "-arch=compute_"
94 #endif
95       + std::to_string(prop.major) + std::to_string(prop.minor);
96   opts[1] = arch_arg.c_str();
97   opts[2] = "-Dint32_t=int";
98   opts[3] = "-DCEED_RUNNING_JIT_PASS=1";
99   {
100     const char **jit_source_dirs;
101 
102     CeedCallBackend(CeedGetJitSourceRoots(ceed, &num_jit_source_dirs, &jit_source_dirs));
103     CeedCallBackend(CeedRealloc(num_opts + num_jit_source_dirs, &opts));
104     for (CeedInt i = 0; i < num_jit_source_dirs; i++) {
105       std::ostringstream include_dirs_arg;
106 
107       include_dirs_arg << "-I" << jit_source_dirs[i];
108       CeedCallBackend(CeedStringAllocCopy(include_dirs_arg.str().c_str(), (char **)&opts[num_opts + i]));
109     }
110     CeedCallBackend(CeedRestoreJitSourceRoots(ceed, &jit_source_dirs));
111   }
112 
113   // Add string source argument provided in call
114   code << source;
115 
116   // Create Program
117   CeedCallNvrtc(ceed, nvrtcCreateProgram(&prog, code.str().c_str(), NULL, 0, NULL, NULL));
118 
119   // Compile kernel
120   nvrtcResult result = nvrtcCompileProgram(prog, num_opts + num_jit_source_dirs, opts);
121 
122   for (CeedInt i = 0; i < num_jit_source_dirs; i++) {
123     CeedCallBackend(CeedFree(&opts[num_opts + i]));
124   }
125   CeedCallBackend(CeedFree(&opts));
126   if (result != NVRTC_SUCCESS) {
127     char  *log;
128     size_t log_size;
129 
130     CeedDebug256(ceed, CEED_DEBUG_COLOR_ERROR, "---------- CEED JIT SOURCE FAILED TO COMPILE ----------\n");
131     CeedDebug(ceed, "Source:\n%s\n", code.str().c_str());
132     CeedDebug256(ceed, CEED_DEBUG_COLOR_ERROR, "---------- CEED JIT SOURCE FAILED TO COMPILE ----------\n");
133     CeedCallNvrtc(ceed, nvrtcGetProgramLogSize(prog, &log_size));
134     CeedCallBackend(CeedMalloc(log_size, &log));
135     CeedCallNvrtc(ceed, nvrtcGetProgramLog(prog, log));
136     return CeedError(ceed, CEED_ERROR_BACKEND, "%s\n%s", nvrtcGetErrorString(result), log);
137   }
138 
139 #if CUDA_VERSION >= 11010
140   CeedCallNvrtc(ceed, nvrtcGetCUBINSize(prog, &ptx_size));
141   CeedCallBackend(CeedMalloc(ptx_size, &ptx));
142   CeedCallNvrtc(ceed, nvrtcGetCUBIN(prog, ptx));
143 #else
144   CeedCallNvrtc(ceed, nvrtcGetPTXSize(prog, &ptx_size));
145   CeedCallBackend(CeedMalloc(ptx_size, &ptx));
146   CeedCallNvrtc(ceed, nvrtcGetPTX(prog, ptx));
147 #endif
148   CeedCallNvrtc(ceed, nvrtcDestroyProgram(&prog));
149 
150   CeedCallCuda(ceed, cuModuleLoadData(module, ptx));
151   CeedCallBackend(CeedFree(&ptx));
152   return CEED_ERROR_SUCCESS;
153 }
154 
155 //------------------------------------------------------------------------------
156 // Get CUDA kernel
157 //------------------------------------------------------------------------------
158 int CeedGetKernel_Cuda(Ceed ceed, CUmodule module, const char *name, CUfunction *kernel) {
159   CeedCallCuda(ceed, cuModuleGetFunction(kernel, module, name));
160   return CEED_ERROR_SUCCESS;
161 }
162 
163 //------------------------------------------------------------------------------
164 // Run CUDA kernel with block size selected automatically based on the kernel
165 //     (which may use enough registers to require a smaller block size than the
166 //      hardware is capable)
167 //------------------------------------------------------------------------------
168 int CeedRunKernelAutoblockCuda(Ceed ceed, CUfunction kernel, size_t points, void **args) {
169   int min_grid_size, max_block_size;
170 
171   CeedCallCuda(ceed, cuOccupancyMaxPotentialBlockSize(&min_grid_size, &max_block_size, kernel, NULL, 0, 0x10000));
172   CeedCallBackend(CeedRunKernel_Cuda(ceed, kernel, CeedDivUpInt(points, max_block_size), max_block_size, args));
173   return CEED_ERROR_SUCCESS;
174 }
175 
176 //------------------------------------------------------------------------------
177 // Run CUDA kernel
178 //------------------------------------------------------------------------------
179 int CeedRunKernel_Cuda(Ceed ceed, CUfunction kernel, const int grid_size, const int block_size, void **args) {
180   CeedCallBackend(CeedRunKernelDimShared_Cuda(ceed, kernel, grid_size, block_size, 1, 1, 0, args));
181   return CEED_ERROR_SUCCESS;
182 }
183 
184 //------------------------------------------------------------------------------
185 // Run CUDA kernel for spatial dimension
186 //------------------------------------------------------------------------------
187 int CeedRunKernelDim_Cuda(Ceed ceed, CUfunction kernel, const int grid_size, const int block_size_x, const int block_size_y, const int block_size_z,
188                           void **args) {
189   CeedCallBackend(CeedRunKernelDimShared_Cuda(ceed, kernel, grid_size, block_size_x, block_size_y, block_size_z, 0, args));
190   return CEED_ERROR_SUCCESS;
191 }
192 
193 //------------------------------------------------------------------------------
194 // Run CUDA kernel for spatial dimension with shared memory
195 //------------------------------------------------------------------------------
196 int CeedRunKernelDimShared_Cuda(Ceed ceed, CUfunction kernel, const int grid_size, const int block_size_x, const int block_size_y,
197                                 const int block_size_z, const int shared_mem_size, void **args) {
198 #if CUDA_VERSION >= 9000
199   cuFuncSetAttribute(kernel, CU_FUNC_ATTRIBUTE_MAX_DYNAMIC_SHARED_SIZE_BYTES, shared_mem_size);
200 #endif
201   CUresult result = cuLaunchKernel(kernel, grid_size, 1, 1, block_size_x, block_size_y, block_size_z, shared_mem_size, NULL, args, NULL);
202 
203   if (result == CUDA_ERROR_LAUNCH_OUT_OF_RESOURCES) {
204     int max_threads_per_block, shared_size_bytes, num_regs;
205 
206     cuFuncGetAttribute(&max_threads_per_block, CU_FUNC_ATTRIBUTE_MAX_THREADS_PER_BLOCK, kernel);
207     cuFuncGetAttribute(&shared_size_bytes, CU_FUNC_ATTRIBUTE_SHARED_SIZE_BYTES, kernel);
208     cuFuncGetAttribute(&num_regs, CU_FUNC_ATTRIBUTE_NUM_REGS, kernel);
209     return CeedError(ceed, CEED_ERROR_BACKEND,
210                      "CUDA_ERROR_LAUNCH_OUT_OF_RESOURCES: max_threads_per_block %d on block size (%d,%d,%d), shared_size %d, num_regs %d",
211                      max_threads_per_block, block_size_x, block_size_y, block_size_z, shared_size_bytes, num_regs);
212   } else CeedChk_Cu(ceed, result);
213   return CEED_ERROR_SUCCESS;
214 }
215 
216 //------------------------------------------------------------------------------
217