ExaGO: add v1.4.0 and v1.4.1 (#29712)
* ExaGO: Handling of cuda architectures and amdgpu targets changed to effectively handle multiple targets. See #28441. * Add ROCm support to ExaGO and update ROCm support in HiOp * ExaGO+rocm requires HiOp+rocm * Newer versions of CMake may set HIP_CLANG_INCLUDE_PATH incorrectly: add comments to the ExaGO/HiOp packages explaining how to address this problem if it occurs.
This commit is contained in:
parent
9516fa9447
commit
078ee48c4b
@ -6,7 +6,7 @@
|
||||
from spack import *
|
||||
|
||||
|
||||
class Exago(CMakePackage, CudaPackage):
|
||||
class Exago(CMakePackage, CudaPackage, ROCmPackage):
|
||||
"""ExaGO is a package for solving large-scale power grid optimization
|
||||
problems on parallel and distributed architectures, particularly targeted
|
||||
for exascale machines."""
|
||||
@ -15,7 +15,9 @@ class Exago(CMakePackage, CudaPackage):
|
||||
git = 'https://gitlab.pnnl.gov/exasgd/frameworks/exago.git'
|
||||
maintainers = ['ashermancinelli', 'CameronRutherford']
|
||||
|
||||
version('1.3.0', commit='58b039d746a6eac8e84b0afc01354cd58caec485', submodules=True, preferred=True)
|
||||
version('1.4.1', commit='ea607c685444b5f345bfdc9a59c345f0f30adde2', submodules=True, preferred=True)
|
||||
version('1.4.0', commit='4f4c3fdb40b52ace2d6ba000e7f24b340ec8e886', submodules=True)
|
||||
version('1.3.0', commit='58b039d746a6eac8e84b0afc01354cd58caec485', submodules=True)
|
||||
version('1.2.0', commit='255a214e', submodules=True)
|
||||
version('1.1.2', commit='db3bb16e', submodules=True)
|
||||
version('1.1.1', commit='0e0a3f27', submodules=True)
|
||||
@ -30,6 +32,7 @@ class Exago(CMakePackage, CudaPackage):
|
||||
variant('mpi', default=True, description='Enable/Disable MPI')
|
||||
variant('raja', default=False, description='Enable/Disable RAJA')
|
||||
variant('python', default=True, description='Enable/Disable Python bindings')
|
||||
conflicts('+python', when='@:1.3.0', msg='Python bindings require ExaGO 1.4')
|
||||
|
||||
# Solver options
|
||||
variant('hiop', default=False, description='Enable/Disable HiOp')
|
||||
@ -44,6 +47,7 @@ class Exago(CMakePackage, CudaPackage):
|
||||
depends_on('raja', when='+raja')
|
||||
|
||||
depends_on('raja+cuda', when='+raja+cuda')
|
||||
depends_on('raja+rocm', when='+raja+rocm')
|
||||
depends_on('raja@0.14.0:', when='@1.1.0: +raja')
|
||||
depends_on('umpire', when='+raja')
|
||||
depends_on('umpire@6.0.0:', when='@1.1.0: +raja')
|
||||
@ -65,37 +69,61 @@ class Exago(CMakePackage, CudaPackage):
|
||||
depends_on('hiop@0.5.3:', when='@1.3.0:+hiop')
|
||||
|
||||
depends_on('hiop+cuda', when='+hiop+cuda')
|
||||
depends_on('hiop+rocm', when='+hiop+rocm')
|
||||
depends_on('hiop~mpi', when='+hiop~mpi')
|
||||
depends_on('hiop+mpi', when='+hiop+mpi')
|
||||
|
||||
depends_on('petsc@3.13:3.14', when='@:1.2.99')
|
||||
depends_on('petsc@3.16.0', when='@1.3.0:')
|
||||
depends_on('petsc@3.16.0:', when='@1.3.0:')
|
||||
depends_on('petsc~mpi', when='~mpi')
|
||||
|
||||
depends_on('ipopt', when='+ipopt')
|
||||
|
||||
depends_on('py-mpi4py', when='@1.3.0:+mpi+python')
|
||||
|
||||
flag_handler = build_system_flags
|
||||
|
||||
def cmake_args(self):
|
||||
args = []
|
||||
spec = self.spec
|
||||
|
||||
args.append("-DEXAGO_RUN_TESTS=ON")
|
||||
# NOTE: If building with spack develop on a cluster, you may want to
|
||||
# change the ctest launch command to use your job scheduler like so:
|
||||
#
|
||||
# args.append(
|
||||
# self.define('EXAGO_CTEST_LAUNCH_COMMAND', 'srun -t 10:00'))
|
||||
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_MPI', 'mpi'))
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_RAJA', 'raja'))
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_HIOP', 'hiop'))
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_IPOPT', 'ipopt'))
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_GPU', 'cuda'))
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_CUDA', 'cuda'))
|
||||
args.append(self.define_from_variant('EXAGO_ENABLE_PYTHON', 'python'))
|
||||
args.append("-DPETSC_DIR='{0}'".format(spec['petsc'].prefix))
|
||||
args.extend([
|
||||
self.define('EXAGO_ENABLE_GPU', '+cuda' in spec or '+rocm' in spec),
|
||||
self.define_from_variant('EXAGO_ENABLE_CUDA', 'cuda'),
|
||||
self.define_from_variant('EXAGO_ENABLE_HIP', 'rocm'),
|
||||
self.define('PETSC_DIR', spec['petsc'].prefix),
|
||||
self.define('EXAGO_RUN_TESTS', True),
|
||||
self.define_from_variant('EXAGO_ENABLE_MPI', 'mpi'),
|
||||
self.define_from_variant('EXAGO_ENABLE_RAJA', 'raja'),
|
||||
self.define_from_variant('EXAGO_ENABLE_HIOP', 'hiop'),
|
||||
self.define_from_variant('EXAGO_ENABLE_IPOPT', 'ipopt'),
|
||||
self.define_from_variant('EXAGO_ENABLE_PYTHON', 'python'),
|
||||
])
|
||||
|
||||
if '+cuda' in spec:
|
||||
cuda_arch_list = spec.variants['cuda_arch'].value
|
||||
cuda_arch = cuda_arch_list[0]
|
||||
if cuda_arch != 'none':
|
||||
if cuda_arch_list[0] != 'none':
|
||||
args.append(
|
||||
"-DCMAKE_CUDA_ARCHITECTURES={0}".format(cuda_arch))
|
||||
self.define('CMAKE_CUDA_ARCHITECTURES', cuda_arch_list))
|
||||
|
||||
# NOTE: if +rocm, some HIP CMake variables may not be set correctly.
|
||||
# Namely, HIP_CLANG_INCLUDE_PATH. If the configure phase fails due to
|
||||
# this variable being undefined, adding the following line typically
|
||||
# resolves this issue:
|
||||
#
|
||||
# args.append(
|
||||
# self.define('HIP_CLANG_INCLUDE_PATH',
|
||||
# '/opt/rocm-X.Y.Z/llvm/lib/clang/14.0.0/include/'))
|
||||
if '+rocm' in spec:
|
||||
rocm_arch_list = spec.variants['amdgpu_target'].value
|
||||
if rocm_arch_list[0] != 'none':
|
||||
args.append(self.define('GPU_TARGETS', rocm_arch_list))
|
||||
args.append(self.define('AMDGPU_TARGETS', rocm_arch_list))
|
||||
|
||||
return args
|
||||
|
@ -76,7 +76,8 @@ class Hiop(CMakePackage, CudaPackage, ROCmPackage):
|
||||
depends_on('magma@{0}:'.format(magma_v), when='@{0}:+cuda'.format(hiop_v))
|
||||
depends_on('magma@{0}:'.format(magma_v), when='@{0}:+rocm'.format(hiop_v))
|
||||
|
||||
depends_on('raja+openmp', when='+raja')
|
||||
depends_on('raja', when='+raja')
|
||||
depends_on('raja+openmp', when='+raja~cuda~rocm')
|
||||
depends_on('raja@0.14.0:', when='@0.5.0:+raja')
|
||||
depends_on('raja+cuda', when='+raja+cuda')
|
||||
depends_on('raja+rocm', when='+raja+rocm')
|
||||
@ -105,15 +106,19 @@ def cmake_args(self):
|
||||
args = []
|
||||
spec = self.spec
|
||||
|
||||
if spec.satisfies('+rocm') or spec.satisfies('+cuda'):
|
||||
args.append('-DHIOP_USE_GPU=ON')
|
||||
args.append('-DHIOP_USE_MAGMA=ON')
|
||||
use_gpu = '+cuda' in spec or '+rocm' in spec
|
||||
|
||||
if use_gpu:
|
||||
args.extend([
|
||||
self.define('HIOP_USE_GPU', True),
|
||||
self.define('HIOP_USE_MAGMA', True),
|
||||
self.define('HIOP_MAGMA_DIR', spec['magma'].prefix),
|
||||
])
|
||||
|
||||
args.extend([
|
||||
self.define('HIOP_BUILD_STATIC', True),
|
||||
self.define('LAPACK_FOUND', True),
|
||||
self.define('LAPACK_LIBRARIES', spec['lapack'].libs + spec['blas'].libs),
|
||||
self.define('HIOP_USE_HIP', False),
|
||||
self.define_from_variant('HIOP_BUILD_SHARED', 'shared'),
|
||||
self.define_from_variant('HIOP_USE_MPI', 'mpi'),
|
||||
self.define_from_variant('HIOP_DEEPCHECKS', 'deepchecking'),
|
||||
@ -127,6 +132,12 @@ def cmake_args(self):
|
||||
self.define_from_variant('HIOP_TEST_WITH_BSUB', 'jsrun'),
|
||||
])
|
||||
|
||||
# NOTE: If building with spack develop on a cluster, you may want to
|
||||
# change the ctest launch command to use your job scheduler like so:
|
||||
#
|
||||
# args.append(
|
||||
# self.define('HIOP_CTEST_LAUNCH_COMMAND', 'srun -t 10:00'))
|
||||
|
||||
if '+mpi' in spec:
|
||||
args.extend([
|
||||
self.define('MPI_HOME', spec['mpi'].prefix),
|
||||
@ -134,17 +145,34 @@ def cmake_args(self):
|
||||
self.define('MPI_CXX_COMPILER', spec['mpi'].mpicxx),
|
||||
self.define('MPI_Fortran_COMPILER', spec['mpi'].mpifc),
|
||||
])
|
||||
# NOTE: On Cray platforms, libfabric is occasionally not picked up
|
||||
# by Spack, causing HiOp's CMake code to fail to find MPI Fortran
|
||||
# libraries. If this is the case, adding the following lines may
|
||||
# resolve the issue. Searching <builddir>/CMakeFiles/CMakeError.log
|
||||
# for MPI Fortran errors is the fastest way to check for this error.
|
||||
#
|
||||
# args.append(
|
||||
# self.define('MPI_Fortran_LINK_FLAGS',
|
||||
# '-L/path/to/libfabric/lib64/ -lfabric'))
|
||||
|
||||
if '+cuda' in spec:
|
||||
cuda_arch_list = spec.variants['cuda_arch'].value
|
||||
cuda_arch = cuda_arch_list[0]
|
||||
if cuda_arch != 'none':
|
||||
args.extend([
|
||||
self.define('HIOP_NVCC_ARCH', 'sm_{0}'.format(cuda_arch)),
|
||||
self.define('CMAKE_CUDA_ARCHITECTURES', cuda_arch),
|
||||
])
|
||||
if '+magma' in spec:
|
||||
args.append(self.define('HIOP_MAGMA_DIR', spec['magma'].prefix))
|
||||
if cuda_arch_list[0] != 'none':
|
||||
args.append(self.define('CMAKE_CUDA_ARCHITECTURES', cuda_arch_list))
|
||||
|
||||
# NOTE: if +rocm, some HIP CMake variables may not be set correctly.
|
||||
# Namely, HIP_CLANG_INCLUDE_PATH. If the configure phase fails due to
|
||||
# this variable being undefined, adding the following line typically
|
||||
# resolves this issue:
|
||||
#
|
||||
# args.append(
|
||||
# self.define('HIP_CLANG_INCLUDE_PATH',
|
||||
# '/opt/rocm-X.Y.Z/llvm/lib/clang/14.0.0/include/'))
|
||||
if '+rocm' in spec:
|
||||
rocm_arch_list = spec.variants['amdgpu_target'].value
|
||||
if rocm_arch_list[0] != 'none':
|
||||
args.append(self.define('GPU_TARGETS', rocm_arch_list))
|
||||
args.append(self.define('AMDGPU_TARGETS', rocm_arch_list))
|
||||
|
||||
if '+kron' in spec:
|
||||
args.append(self.define('HIOP_UMFPACK_DIR', spec['suite-sparse'].prefix))
|
||||
|
Loading…
Reference in New Issue
Block a user