1 #include <petsc/private/cpp/memory.hpp> // make_unique 2 3 #include "cupmdevice.hpp" 4 5 #include <algorithm> 6 #include <csetjmp> // for cuda mpi awareness 7 #include <csignal> // SIGSEGV 8 #include <iterator> 9 #include <type_traits> 10 11 namespace Petsc 12 { 13 14 namespace device 15 { 16 17 namespace cupm 18 { 19 20 // internal "impls" class for CUPMDevice. Each instance represents a single cupm device 21 template <DeviceType T> 22 class Device<T>::DeviceInternal { 23 const int id_; 24 bool devInitialized_ = false; 25 cupmDeviceProp_t dprop_{}; // cudaDeviceProp appears to be an actual struct, i.e. you can't 26 // initialize it with nullptr or NULL (i've tried) 27 28 static PetscErrorCode CUPMAwareMPI_(bool *) noexcept; 29 30 public: 31 // default constructor 32 explicit constexpr DeviceInternal(int dev) noexcept : id_(dev) { } 33 34 // gather all relevant information for a particular device, a cupmDeviceProp_t is 35 // usually sufficient here 36 PetscErrorCode initialize() noexcept; 37 PetscErrorCode configure() noexcept; 38 PetscErrorCode view(PetscViewer) const noexcept; 39 PetscErrorCode getattribute(PetscDeviceAttribute, void *) const noexcept; 40 41 PETSC_NODISCARD auto id() const -> decltype(id_) { return id_; } 42 PETSC_NODISCARD auto initialized() const -> decltype(devInitialized_) { return devInitialized_; } 43 PETSC_NODISCARD auto prop() const -> const decltype(dprop_) & { return dprop_; } 44 }; 45 46 // the goal here is simply to get the cupm backend to create its context, not to do any type of 47 // modification of it, or create objects (since these may be affected by subsequent 48 // configuration changes) 49 template <DeviceType T> 50 PetscErrorCode Device<T>::DeviceInternal::initialize() noexcept 51 { 52 PetscFunctionBegin; 53 if (initialized()) PetscFunctionReturn(PETSC_SUCCESS); 54 devInitialized_ = true; 55 // need to do this BEFORE device has been set, although if the user 56 // has already done this then we just ignore it 57 if (cupmSetDeviceFlags(cupmDeviceMapHost) == cupmErrorSetOnActiveProcess) { 58 // reset the error if it was cupmErrorSetOnActiveProcess 59 const auto PETSC_UNUSED unused = cupmGetLastError(); 60 } else PetscCallCUPM(cupmGetLastError()); 61 // cuda 5.0+ will create a context when cupmSetDevice is called 62 if (cupmSetDevice(id()) != cupmErrorDeviceAlreadyInUse) PetscCallCUPM(cupmGetLastError()); 63 // and in case it doesn't, explicitly call init here 64 PetscCallCUPM(cupmInit(0)); 65 // where is this variable defined and when is it set? who knows! but it is defined and set 66 // at this point. either way, each device must make this check since I guess MPI might not be 67 // aware of all of them? 68 if (use_gpu_aware_mpi) { 69 bool aware; 70 71 PetscCall(CUPMAwareMPI_(&aware)); 72 // For OpenMPI, we could do a compile time check with 73 // "defined(PETSC_HAVE_OMPI_MAJOR_VERSION) && defined(MPIX_CUDA_AWARE_SUPPORT) && 74 // MPIX_CUDA_AWARE_SUPPORT" to see if it is CUDA-aware. However, recent versions of IBM 75 // Spectrum MPI (e.g., 10.3.1) on Summit meet above conditions, but one has to use jsrun 76 // --smpiargs=-gpu to really enable GPU-aware MPI. So we do the check at runtime with a 77 // code that works only with GPU-aware MPI. 78 if (PetscUnlikely(!aware)) { 79 PetscCall((*PetscErrorPrintf)("PETSc is configured with GPU support, but your MPI is not GPU-aware. For better performance, please use a GPU-aware MPI.\n")); 80 PetscCall((*PetscErrorPrintf)("If you do not care, add option -use_gpu_aware_mpi 0. To not see the message again, add the option to your .petscrc, OR add it to the env var PETSC_OPTIONS.\n")); 81 PetscCall((*PetscErrorPrintf)("If you do care, for IBM Spectrum MPI on OLCF Summit, you may need jsrun --smpiargs=-gpu.\n")); 82 PetscCall((*PetscErrorPrintf)("For OpenMPI, you need to configure it --with-cuda (https://www.open-mpi.org/faq/?category=buildcuda)\n")); 83 PetscCall((*PetscErrorPrintf)("For MVAPICH2-GDR, you need to set MV2_USE_CUDA=1 (http://mvapich.cse.ohio-state.edu/userguide/gdr/)\n")); 84 PetscCall((*PetscErrorPrintf)("For Cray-MPICH, you need to set MPICH_GPU_SUPPORT_ENABLED=1 (man mpi to see manual of cray-mpich)\n")); 85 PETSCABORT(PETSC_COMM_SELF, PETSC_ERR_LIB); 86 } 87 } 88 PetscFunctionReturn(PETSC_SUCCESS); 89 } 90 91 template <DeviceType T> 92 PetscErrorCode Device<T>::DeviceInternal::configure() noexcept 93 { 94 PetscFunctionBegin; 95 PetscAssert(initialized(), PETSC_COMM_SELF, PETSC_ERR_COR, "Device %d being configured before it was initialized", id()); 96 // why on EARTH nvidia insists on making otherwise informational states into 97 // fully-fledged error codes is beyond me. Why couldn't a pointer to bool argument have 98 // sufficed?!?!?! 99 if (cupmSetDevice(id_) != cupmErrorDeviceAlreadyInUse) PetscCallCUPM(cupmGetLastError()); 100 // need to update the device properties 101 PetscCallCUPM(cupmGetDeviceProperties(&dprop_, id_)); 102 PetscCall(PetscInfo(nullptr, "Configured device %d\n", id_)); 103 PetscFunctionReturn(PETSC_SUCCESS); 104 } 105 106 template <DeviceType T> 107 PetscErrorCode Device<T>::DeviceInternal::view(PetscViewer viewer) const noexcept 108 { 109 PetscBool iascii; 110 111 PetscFunctionBegin; 112 PetscAssert(initialized(), PETSC_COMM_SELF, PETSC_ERR_COR, "Device %d being viewed before it was initialized or configured", id()); 113 // we don't print device-specific info in CI-mode 114 if (PetscUnlikely(PetscCIEnabled)) PetscFunctionReturn(PETSC_SUCCESS); 115 PetscCall(PetscObjectTypeCompare(PetscObjectCast(viewer), PETSCVIEWERASCII, &iascii)); 116 if (iascii) { 117 MPI_Comm comm; 118 PetscMPIInt rank; 119 PetscViewer sviewer; 120 121 PetscCall(PetscObjectGetComm(PetscObjectCast(viewer), &comm)); 122 PetscCallMPI(MPI_Comm_rank(comm, &rank)); 123 PetscCall(PetscViewerGetSubViewer(viewer, PETSC_COMM_SELF, &sviewer)); 124 PetscCall(PetscViewerASCIIPrintf(sviewer, "[%d] name: %s\n", rank, dprop_.name)); 125 PetscCall(PetscViewerASCIIPushTab(sviewer)); 126 PetscCall(PetscViewerASCIIPrintf(sviewer, "Compute capability: %d.%d\n", dprop_.major, dprop_.minor)); 127 PetscCall(PetscViewerASCIIPrintf(sviewer, "Multiprocessor Count: %d\n", dprop_.multiProcessorCount)); 128 PetscCall(PetscViewerASCIIPrintf(sviewer, "Maximum Grid Dimensions: %d x %d x %d\n", dprop_.maxGridSize[0], dprop_.maxGridSize[1], dprop_.maxGridSize[2])); 129 PetscCall(PetscViewerASCIIPrintf(sviewer, "Maximum Block Dimensions: %d x %d x %d\n", dprop_.maxThreadsDim[0], dprop_.maxThreadsDim[1], dprop_.maxThreadsDim[2])); 130 PetscCall(PetscViewerASCIIPrintf(sviewer, "Maximum Threads Per Block: %d\n", dprop_.maxThreadsPerBlock)); 131 PetscCall(PetscViewerASCIIPrintf(sviewer, "Warp Size: %d\n", dprop_.warpSize)); 132 PetscCall(PetscViewerASCIIPrintf(sviewer, "Total Global Memory (bytes): %zu\n", dprop_.totalGlobalMem)); 133 PetscCall(PetscViewerASCIIPrintf(sviewer, "Total Constant Memory (bytes): %zu\n", dprop_.totalConstMem)); 134 PetscCall(PetscViewerASCIIPrintf(sviewer, "Shared Memory Per Block (bytes): %zu\n", dprop_.sharedMemPerBlock)); 135 PetscCall(PetscViewerASCIIPrintf(sviewer, "Multiprocessor Clock Rate (KHz): %d\n", dprop_.clockRate)); 136 PetscCall(PetscViewerASCIIPrintf(sviewer, "Memory Clock Rate (KHz): %d\n", dprop_.memoryClockRate)); 137 PetscCall(PetscViewerASCIIPrintf(sviewer, "Memory Bus Width (bits): %d\n", dprop_.memoryBusWidth)); 138 PetscCall(PetscViewerASCIIPrintf(sviewer, "Peak Memory Bandwidth (GB/s): %f\n", 2.0 * dprop_.memoryClockRate * (dprop_.memoryBusWidth / 8) / 1.0e6)); 139 PetscCall(PetscViewerASCIIPrintf(sviewer, "Can map host memory: %s\n", dprop_.canMapHostMemory ? "PETSC_TRUE" : "PETSC_FALSE")); 140 PetscCall(PetscViewerASCIIPrintf(sviewer, "Can execute multiple kernels concurrently: %s\n", dprop_.concurrentKernels ? "PETSC_TRUE" : "PETSC_FALSE")); 141 PetscCall(PetscViewerASCIIPopTab(sviewer)); 142 PetscCall(PetscViewerRestoreSubViewer(viewer, PETSC_COMM_SELF, &sviewer)); 143 PetscCall(PetscViewerFlush(viewer)); 144 } 145 PetscFunctionReturn(PETSC_SUCCESS); 146 } 147 148 template <DeviceType T> 149 PetscErrorCode Device<T>::DeviceInternal::getattribute(PetscDeviceAttribute attr, void *value) const noexcept 150 { 151 PetscFunctionBegin; 152 PetscAssert(initialized(), PETSC_COMM_SELF, PETSC_ERR_COR, "Device %d was not initialized", id()); 153 switch (attr) { 154 case PETSC_DEVICE_ATTR_SIZE_T_SHARED_MEM_PER_BLOCK: 155 *static_cast<std::size_t *>(value) = prop().sharedMemPerBlock; 156 case PETSC_DEVICE_ATTR_MAX: 157 break; 158 } 159 PetscFunctionReturn(PETSC_SUCCESS); 160 } 161 162 static std::jmp_buf cupmMPIAwareJumpBuffer; 163 static bool cupmMPIAwareJumpBufferSet; 164 165 // godspeed to anyone that attempts to call this function 166 void SilenceVariableIsNotNeededAndWillNotBeEmittedWarning_ThisFunctionShouldNeverBeCalled() 167 { 168 PETSCABORT(MPI_COMM_NULL, (PetscErrorCode)INT_MAX); 169 if (cupmMPIAwareJumpBufferSet) (void)cupmMPIAwareJumpBuffer; 170 } 171 172 template <DeviceType T> 173 PetscErrorCode Device<T>::DeviceInternal::CUPMAwareMPI_(bool *awareness) noexcept 174 { 175 constexpr int hbuf[] = {1, 0}; 176 int *dbuf = nullptr; 177 const auto cupmSignalHandler = [](int signal, void *ptr) -> PetscErrorCode { 178 if ((signal == SIGSEGV) && cupmMPIAwareJumpBufferSet) std::longjmp(cupmMPIAwareJumpBuffer, 1); 179 return PetscSignalHandlerDefault(signal, ptr); 180 }; 181 182 PetscFunctionBegin; 183 *awareness = false; 184 PetscCallCUPM(cupmMalloc(reinterpret_cast<void **>(&dbuf), sizeof(hbuf))); 185 PetscCallCUPM(cupmMemcpy(dbuf, hbuf, sizeof(hbuf), cupmMemcpyHostToDevice)); 186 PetscCallCUPM(cupmDeviceSynchronize()); 187 PetscCall(PetscPushSignalHandler(cupmSignalHandler, nullptr)); 188 cupmMPIAwareJumpBufferSet = true; 189 if (!setjmp(cupmMPIAwareJumpBuffer) && !MPI_Allreduce(dbuf, dbuf + 1, 1, MPI_INT, MPI_SUM, PETSC_COMM_SELF)) *awareness = true; 190 cupmMPIAwareJumpBufferSet = false; 191 PetscCall(PetscPopSignalHandler()); 192 PetscCallCUPM(cupmFree(dbuf)); 193 PetscFunctionReturn(PETSC_SUCCESS); 194 } 195 196 template <DeviceType T> 197 PetscErrorCode Device<T>::finalize_() noexcept 198 { 199 PetscFunctionBegin; 200 if (PetscUnlikely(!initialized_)) PetscFunctionReturn(PETSC_SUCCESS); 201 for (auto &&device : devices_) device.reset(); 202 defaultDevice_ = PETSC_CUPM_DEVICE_NONE; // disabled by default 203 initialized_ = false; 204 PetscFunctionReturn(PETSC_SUCCESS); 205 } 206 207 template <DeviceType T> 208 PETSC_NODISCARD static PETSC_CONSTEXPR_14 const char *CUPM_VISIBLE_DEVICES() noexcept 209 { 210 switch (T) { 211 case DeviceType::CUDA: 212 return "CUDA_VISIBLE_DEVICES"; 213 case DeviceType::HIP: 214 return "HIP_VISIBLE_DEVICES"; 215 } 216 PetscUnreachable(); 217 return "PETSC_ERROR_PLIB"; 218 } 219 220 template <DeviceType T> 221 PetscErrorCode Device<T>::initialize(MPI_Comm comm, PetscInt *defaultDeviceId, PetscBool *defaultView, PetscDeviceInitType *defaultInitType) noexcept 222 { 223 auto initId = std::make_pair(*defaultDeviceId, PETSC_FALSE); 224 auto initView = std::make_pair(*defaultView, PETSC_FALSE); 225 auto initType = std::make_pair(*defaultInitType, PETSC_FALSE); 226 int ndev = 0; 227 228 PetscFunctionBegin; 229 if (initialized_) PetscFunctionReturn(PETSC_SUCCESS); 230 initialized_ = true; 231 PetscCall(PetscRegisterFinalize(finalize_)); 232 PetscCall(base_type::PetscOptionDeviceAll(comm, initType, initId, initView)); 233 234 if (initType.first == PETSC_DEVICE_INIT_NONE) { 235 initId.first = PETSC_CUPM_DEVICE_NONE; 236 } else if (const auto cerr = cupmGetDeviceCount(&ndev)) { 237 auto PETSC_UNUSED ignored = cupmGetLastError(); 238 239 PetscCheck((initType.first != PETSC_DEVICE_INIT_EAGER) && !initView.first, comm, PETSC_ERR_USER_INPUT, "Cannot eagerly initialize %s, as doing so results in %s error %d (%s) : %s", cupmName(), cupmName(), static_cast<PetscErrorCode>(cerr), cupmGetErrorName(cerr), cupmGetErrorString(cerr)); 240 // we won't be initializing anything anyways 241 initType.first = PETSC_DEVICE_INIT_NONE; 242 // save the error code for later 243 initId.first = -static_cast<decltype(initId.first)>(cerr); 244 } 245 246 // check again for init type, since the device count may have changed it 247 if (initType.first == PETSC_DEVICE_INIT_NONE) { 248 // id < 0 (excluding PETSC_DECIDE) indicates an error has occurred during setup 249 if ((initId.first > 0) || (initId.first == PETSC_DECIDE)) initId.first = PETSC_CUPM_DEVICE_NONE; 250 // initType overrides initView 251 initView.first = PETSC_FALSE; 252 } else { 253 PetscCall(PetscDeviceCheckDeviceCount_Internal(ndev)); 254 if (initId.first == PETSC_DECIDE) { 255 if (ndev) { 256 PetscMPIInt rank; 257 258 PetscCallMPI(MPI_Comm_rank(comm, &rank)); 259 initId.first = rank % ndev; 260 } else initId.first = 0; 261 } 262 if (initView.first) initType.first = PETSC_DEVICE_INIT_EAGER; 263 } 264 265 static_assert(std::is_same<PetscMPIInt, decltype(defaultDevice_)>::value, ""); 266 // initId.first is PetscInt, _defaultDevice is int 267 PetscCall(PetscMPIIntCast(initId.first, &defaultDevice_)); 268 // record the results of the initialization 269 *defaultDeviceId = initId.first; 270 *defaultView = initView.first; 271 *defaultInitType = initType.first; 272 PetscFunctionReturn(PETSC_SUCCESS); 273 } 274 275 template <DeviceType T> 276 PetscErrorCode Device<T>::init_device_id_(PetscInt *inid) const noexcept 277 { 278 const auto id = *inid == PETSC_DECIDE ? defaultDevice_ : *inid; 279 const auto cerr = static_cast<cupmError_t>(-defaultDevice_); 280 281 PetscFunctionBegin; 282 PetscCheck(defaultDevice_ != PETSC_CUPM_DEVICE_NONE, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "Trying to retrieve a %s PetscDevice when it has been disabled", cupmName()); 283 PetscCheck(defaultDevice_ >= 0, PETSC_COMM_SELF, PETSC_ERR_GPU, "Cannot lazily initialize PetscDevice: %s error %d (%s) : %s", cupmName(), static_cast<PetscErrorCode>(cerr), cupmGetErrorName(cerr), cupmGetErrorString(cerr)); 284 PetscAssert(static_cast<decltype(devices_.size())>(id) < devices_.size(), PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Only supports %zu number of devices but trying to get device with id %" PetscInt_FMT, devices_.size(), id); 285 286 if (!devices_[id]) devices_[id] = util::make_unique<DeviceInternal>(id); 287 PetscAssert(id == devices_[id]->id(), PETSC_COMM_SELF, PETSC_ERR_PLIB, "Entry %" PetscInt_FMT " contains device with mismatching id %d", id, devices_[id]->id()); 288 PetscCall(devices_[id]->initialize()); 289 *inid = id; 290 PetscFunctionReturn(PETSC_SUCCESS); 291 } 292 293 template <DeviceType T> 294 PetscErrorCode Device<T>::configure_device_(PetscDevice device) noexcept 295 { 296 PetscFunctionBegin; 297 PetscCall(devices_[device->deviceId]->configure()); 298 PetscFunctionReturn(PETSC_SUCCESS); 299 } 300 301 template <DeviceType T> 302 PetscErrorCode Device<T>::view_device_(PetscDevice device, PetscViewer viewer) noexcept 303 { 304 PetscFunctionBegin; 305 // now this __shouldn't__ reconfigure the device, but there is a petscinfo call to indicate 306 // it is being reconfigured 307 PetscCall(devices_[device->deviceId]->configure()); 308 PetscCall(devices_[device->deviceId]->view(viewer)); 309 PetscFunctionReturn(PETSC_SUCCESS); 310 } 311 312 template <DeviceType T> 313 PetscErrorCode Device<T>::get_attribute_(PetscInt id, PetscDeviceAttribute attr, void *value) noexcept 314 { 315 PetscFunctionBegin; 316 PetscCall(devices_[id]->getattribute(attr, value)); 317 PetscFunctionReturn(PETSC_SUCCESS); 318 } 319 320 // explicitly instantiate the classes 321 #if PetscDefined(HAVE_CUDA) 322 template class Device<DeviceType::CUDA>; 323 #endif 324 #if PetscDefined(HAVE_HIP) 325 template class Device<DeviceType::HIP>; 326 #endif 327 328 } // namespace cupm 329 330 } // namespace device 331 332 } // namespace Petsc 333