#include // make_unique #include "cupmdevice.hpp" #include #include // for cuda mpi awareness #include // SIGSEGV #include #include namespace Petsc { namespace device { namespace cupm { // internal "impls" class for CUPMDevice. Each instance represents a single cupm device template class Device::DeviceInternal { const int id_; bool devInitialized_ = false; cupmDeviceProp_t dprop_{}; // cudaDeviceProp appears to be an actual struct, i.e. you can't // initialize it with nullptr or NULL (i've tried) static PetscErrorCode CUPMAwareMPI_(bool *) noexcept; public: // default constructor explicit constexpr DeviceInternal(int dev) noexcept : id_(dev) { } // gather all relevant information for a particular device, a cupmDeviceProp_t is // usually sufficient here PetscErrorCode initialize() noexcept; PetscErrorCode configure() noexcept; PetscErrorCode view(PetscViewer) const noexcept; PetscErrorCode getattribute(PetscDeviceAttribute, void *) const noexcept; PETSC_NODISCARD auto id() const -> decltype(id_) { return id_; } PETSC_NODISCARD auto initialized() const -> decltype(devInitialized_) { return devInitialized_; } PETSC_NODISCARD auto prop() const -> const decltype(dprop_) & { return dprop_; } }; // the goal here is simply to get the cupm backend to create its context, not to do any type of // modification of it, or create objects (since these may be affected by subsequent // configuration changes) template PetscErrorCode Device::DeviceInternal::initialize() noexcept { PetscFunctionBegin; if (initialized()) PetscFunctionReturn(PETSC_SUCCESS); devInitialized_ = true; // need to do this BEFORE device has been set, although if the user // has already done this then we just ignore it if (cupmSetDeviceFlags(cupmDeviceMapHost) == cupmErrorSetOnActiveProcess) { // reset the error if it was cupmErrorSetOnActiveProcess const auto PETSC_UNUSED unused = cupmGetLastError(); } else PetscCallCUPM(cupmGetLastError()); // cuda 5.0+ will create a context when cupmSetDevice is called if (cupmSetDevice(id()) != cupmErrorDeviceAlreadyInUse) PetscCallCUPM(cupmGetLastError()); // and in case it doesn't, explicitly call init here PetscCallCUPM(cupmInit(0)); // where is this variable defined and when is it set? who knows! but it is defined and set // at this point. either way, each device must make this check since I guess MPI might not be // aware of all of them? if (use_gpu_aware_mpi) { bool aware; PetscCall(CUPMAwareMPI_(&aware)); // For Open MPI, we could do a compile time check with // "defined(PETSC_HAVE_OMPI_MAJOR_VERSION) && defined(MPIX_CUDA_AWARE_SUPPORT) && // MPIX_CUDA_AWARE_SUPPORT" to see if it is CUDA-aware. However, recent versions of IBM // Spectrum MPI (e.g., 10.3.1) on Summit meet above conditions, but one has to use jsrun // --smpiargs=-gpu to really enable GPU-aware MPI. So we do the check at runtime with a // code that works only with GPU-aware MPI. if (PetscUnlikely(!aware)) { PetscCall((*PetscErrorPrintf)("PETSc is configured with GPU support, but your MPI is not GPU-aware. For better performance, please use a GPU-aware MPI.\n")); PetscCall((*PetscErrorPrintf)("If you do not care, add option -use_gpu_aware_mpi 0. To not see the message again, add the option to your .petscrc, OR add it to the env var PETSC_OPTIONS.\n")); PetscCall((*PetscErrorPrintf)("If you do care, for IBM Spectrum MPI on OLCF Summit, you may need jsrun --smpiargs=-gpu.\n")); PetscCall((*PetscErrorPrintf)("For Open MPI, you need to configure it --with-cuda (https://www.open-mpi.org/faq/?category=buildcuda)\n")); PetscCall((*PetscErrorPrintf)("For MVAPICH2-GDR, you need to set MV2_USE_CUDA=1 (http://mvapich.cse.ohio-state.edu/userguide/gdr/)\n")); PetscCall((*PetscErrorPrintf)("For Cray-MPICH, you need to set MPICH_GPU_SUPPORT_ENABLED=1 (man mpi to see manual of cray-mpich)\n")); PETSCABORT(PETSC_COMM_SELF, PETSC_ERR_LIB); } } PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::DeviceInternal::configure() noexcept { PetscFunctionBegin; PetscAssert(initialized(), PETSC_COMM_SELF, PETSC_ERR_COR, "Device %d being configured before it was initialized", id()); // why on EARTH nvidia insists on making otherwise informational states into // fully-fledged error codes is beyond me. Why couldn't a pointer to bool argument have // sufficed?!?!?! if (cupmSetDevice(id_) != cupmErrorDeviceAlreadyInUse) PetscCallCUPM(cupmGetLastError()); // need to update the device properties PetscCallCUPM(cupmGetDeviceProperties(&dprop_, id_)); PetscDeviceCUPMRuntimeArch = dprop_.major * 10 + dprop_.minor; PetscCall(PetscInfo(nullptr, "Configured device %d\n", id_)); PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::DeviceInternal::view(PetscViewer viewer) const noexcept { PetscBool iascii; PetscFunctionBegin; PetscAssert(initialized(), PETSC_COMM_SELF, PETSC_ERR_COR, "Device %d being viewed before it was initialized or configured", id()); // we don't print device-specific info in CI-mode if (PetscUnlikely(PetscCIEnabled)) PetscFunctionReturn(PETSC_SUCCESS); PetscCall(PetscObjectTypeCompare(PetscObjectCast(viewer), PETSCVIEWERASCII, &iascii)); if (iascii) { MPI_Comm comm; PetscMPIInt rank; PetscViewer sviewer; PetscCall(PetscObjectGetComm(PetscObjectCast(viewer), &comm)); PetscCallMPI(MPI_Comm_rank(comm, &rank)); PetscCall(PetscViewerGetSubViewer(viewer, PETSC_COMM_SELF, &sviewer)); PetscCall(PetscViewerASCIIPrintf(sviewer, "[%d] name: %s\n", rank, dprop_.name)); PetscCall(PetscViewerASCIIPushTab(sviewer)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Compute capability: %d.%d\n", dprop_.major, dprop_.minor)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Multiprocessor Count: %d\n", dprop_.multiProcessorCount)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Maximum Grid Dimensions: %d x %d x %d\n", dprop_.maxGridSize[0], dprop_.maxGridSize[1], dprop_.maxGridSize[2])); PetscCall(PetscViewerASCIIPrintf(sviewer, "Maximum Block Dimensions: %d x %d x %d\n", dprop_.maxThreadsDim[0], dprop_.maxThreadsDim[1], dprop_.maxThreadsDim[2])); PetscCall(PetscViewerASCIIPrintf(sviewer, "Maximum Threads Per Block: %d\n", dprop_.maxThreadsPerBlock)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Warp Size: %d\n", dprop_.warpSize)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Total Global Memory (bytes): %zu\n", dprop_.totalGlobalMem)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Total Constant Memory (bytes): %zu\n", dprop_.totalConstMem)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Shared Memory Per Block (bytes): %zu\n", dprop_.sharedMemPerBlock)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Multiprocessor Clock Rate (KHz): %d\n", dprop_.clockRate)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Memory Clock Rate (KHz): %d\n", dprop_.memoryClockRate)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Memory Bus Width (bits): %d\n", dprop_.memoryBusWidth)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Peak Memory Bandwidth (GB/s): %f\n", 2.0 * dprop_.memoryClockRate * (dprop_.memoryBusWidth / 8) / 1.0e6)); PetscCall(PetscViewerASCIIPrintf(sviewer, "Can map host memory: %s\n", dprop_.canMapHostMemory ? "PETSC_TRUE" : "PETSC_FALSE")); PetscCall(PetscViewerASCIIPrintf(sviewer, "Can execute multiple kernels concurrently: %s\n", dprop_.concurrentKernels ? "PETSC_TRUE" : "PETSC_FALSE")); PetscCall(PetscViewerASCIIPopTab(sviewer)); PetscCall(PetscViewerRestoreSubViewer(viewer, PETSC_COMM_SELF, &sviewer)); } PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::DeviceInternal::getattribute(PetscDeviceAttribute attr, void *value) const noexcept { PetscFunctionBegin; PetscAssert(initialized(), PETSC_COMM_SELF, PETSC_ERR_COR, "Device %d was not initialized", id()); switch (attr) { case PETSC_DEVICE_ATTR_SIZE_T_SHARED_MEM_PER_BLOCK: *static_cast(value) = prop().sharedMemPerBlock; case PETSC_DEVICE_ATTR_MAX: break; } PetscFunctionReturn(PETSC_SUCCESS); } static std::jmp_buf cupmMPIAwareJumpBuffer; static bool cupmMPIAwareJumpBufferSet; // godspeed to anyone that attempts to call this function void SilenceVariableIsNotNeededAndWillNotBeEmittedWarning_ThisFunctionShouldNeverBeCalled() { PETSCABORT(MPI_COMM_NULL, (PetscErrorCode)INT_MAX); if (cupmMPIAwareJumpBufferSet) (void)cupmMPIAwareJumpBuffer; } template PetscErrorCode Device::DeviceInternal::CUPMAwareMPI_(bool *awareness) noexcept { constexpr int hbuf[] = {1, 0}; int *dbuf = nullptr; const auto cupmSignalHandler = [](int signal, void *ptr) -> PetscErrorCode { if ((signal == SIGSEGV) && cupmMPIAwareJumpBufferSet) std::longjmp(cupmMPIAwareJumpBuffer, 1); return PetscSignalHandlerDefault(signal, ptr); }; PetscFunctionBegin; *awareness = false; PetscCallCUPM(cupmMalloc(reinterpret_cast(&dbuf), sizeof(hbuf))); PetscCallCUPM(cupmMemcpy(dbuf, hbuf, sizeof(hbuf), cupmMemcpyHostToDevice)); PetscCallCUPM(cupmDeviceSynchronize()); PetscCall(PetscPushSignalHandler(cupmSignalHandler, nullptr)); cupmMPIAwareJumpBufferSet = true; if (!setjmp(cupmMPIAwareJumpBuffer) && !MPI_Allreduce(dbuf, dbuf + 1, 1, MPI_INT, MPI_SUM, PETSC_COMM_SELF)) *awareness = true; cupmMPIAwareJumpBufferSet = false; PetscCall(PetscPopSignalHandler()); PetscCallCUPM(cupmFree(dbuf)); PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::finalize_() noexcept { PetscFunctionBegin; if (PetscUnlikely(!initialized_)) PetscFunctionReturn(PETSC_SUCCESS); for (auto &&device : devices_) device.reset(); defaultDevice_ = PETSC_CUPM_DEVICE_NONE; // disabled by default initialized_ = false; PetscFunctionReturn(PETSC_SUCCESS); } template PETSC_NODISCARD static PETSC_CONSTEXPR_14 const char *CUPM_VISIBLE_DEVICES() noexcept { switch (T) { case DeviceType::CUDA: return "CUDA_VISIBLE_DEVICES"; case DeviceType::HIP: return "HIP_VISIBLE_DEVICES"; } PetscUnreachable(); return "PETSC_ERROR_PLIB"; } template PetscErrorCode Device::initialize(MPI_Comm comm, PetscInt *defaultDeviceId, PetscBool *defaultView, PetscDeviceInitType *defaultInitType) noexcept { auto initId = std::make_pair(*defaultDeviceId, PETSC_FALSE); auto initView = std::make_pair(*defaultView, PETSC_FALSE); auto initType = std::make_pair(*defaultInitType, PETSC_FALSE); int ndev = 0; PetscFunctionBegin; if (initialized_) PetscFunctionReturn(PETSC_SUCCESS); initialized_ = true; PetscCall(PetscRegisterFinalize(finalize_)); PetscCall(base_type::PetscOptionDeviceAll(comm, initType, initId, initView)); if (initType.first == PETSC_DEVICE_INIT_NONE) { initId.first = PETSC_CUPM_DEVICE_NONE; } else if (const auto cerr = cupmGetDeviceCount(&ndev)) { auto PETSC_UNUSED ignored = cupmGetLastError(); PetscCheck((initType.first != PETSC_DEVICE_INIT_EAGER) && !initView.first, comm, PETSC_ERR_USER_INPUT, "Cannot eagerly initialize %s, as doing so results in %s error %d (%s) : %s", cupmName(), cupmName(), static_cast(cerr), cupmGetErrorName(cerr), cupmGetErrorString(cerr)); // we won't be initializing anything anyways initType.first = PETSC_DEVICE_INIT_NONE; // save the error code for later initId.first = -static_cast(cerr); } // check again for init type, since the device count may have changed it if (initType.first == PETSC_DEVICE_INIT_NONE) { // id < 0 (excluding PETSC_DECIDE) indicates an error has occurred during setup if ((initId.first > 0) || (initId.first == PETSC_DECIDE)) initId.first = PETSC_CUPM_DEVICE_NONE; // initType overrides initView initView.first = PETSC_FALSE; } else { PetscCall(PetscDeviceCheckDeviceCount_Internal(ndev)); if (initId.first == PETSC_DECIDE) { if (ndev) { PetscMPIInt rank; PetscCallMPI(MPI_Comm_rank(comm, &rank)); initId.first = rank % ndev; } else initId.first = 0; } if (initView.first) initType.first = PETSC_DEVICE_INIT_EAGER; } static_assert(std::is_same::value, ""); // initId.first is PetscInt, _defaultDevice is int PetscCall(PetscMPIIntCast(initId.first, &defaultDevice_)); // record the results of the initialization *defaultDeviceId = initId.first; *defaultView = initView.first; *defaultInitType = initType.first; PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::init_device_id_(PetscInt *inid) const noexcept { const auto id = *inid == PETSC_DECIDE ? defaultDevice_ : *inid; const auto cerr = static_cast(-defaultDevice_); PetscFunctionBegin; PetscCheck(defaultDevice_ != PETSC_CUPM_DEVICE_NONE, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "Trying to retrieve a %s PetscDevice when it has been disabled", cupmName()); PetscCheck(defaultDevice_ >= 0, PETSC_COMM_SELF, PETSC_ERR_GPU, "Cannot lazily initialize PetscDevice: %s error %d (%s) : %s", cupmName(), static_cast(cerr), cupmGetErrorName(cerr), cupmGetErrorString(cerr)); PetscAssert(static_cast(id) < devices_.size(), PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Only supports %zu number of devices but trying to get device with id %" PetscInt_FMT, devices_.size(), id); if (!devices_[id]) devices_[id] = util::make_unique(id); PetscAssert(id == devices_[id]->id(), PETSC_COMM_SELF, PETSC_ERR_PLIB, "Entry %" PetscInt_FMT " contains device with mismatching id %d", id, devices_[id]->id()); PetscCall(devices_[id]->initialize()); *inid = id; PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::configure_device_(PetscDevice device) noexcept { PetscFunctionBegin; PetscCall(devices_[device->deviceId]->configure()); PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::view_device_(PetscDevice device, PetscViewer viewer) noexcept { PetscFunctionBegin; // now this __shouldn't__ reconfigure the device, but there is a petscinfo call to indicate // it is being reconfigured PetscCall(devices_[device->deviceId]->configure()); PetscCall(devices_[device->deviceId]->view(viewer)); PetscFunctionReturn(PETSC_SUCCESS); } template PetscErrorCode Device::get_attribute_(PetscInt id, PetscDeviceAttribute attr, void *value) noexcept { PetscFunctionBegin; PetscCall(devices_[id]->getattribute(attr, value)); PetscFunctionReturn(PETSC_SUCCESS); } // explicitly instantiate the classes #if PetscDefined(HAVE_CUDA) template class Device; #endif #if PetscDefined(HAVE_HIP) template class Device; #endif } // namespace cupm } // namespace device } // namespace Petsc