py-torch: overhaul package (#24294)
* py-torch: patch no longer needed on master * Overhaul PyTorch package * py-torch: add v1.9.0 * Change defaults on macOS * Submodules still needed... * Add ONNX dependency * System libs don't work for many submodules * Silence CMake warning * Add conflict for +cuda+rocm * Add more deps * Add more BLAS options * Disable some broken system libs options * Add patches to build older versions * +mkldnn requires mkl * Fix BLAS settings
This commit is contained in:
parent
7753c816f0
commit
4f9b539644
@ -0,0 +1,53 @@
|
||||
diff --git a/aten/src/ATen/native/sparse/cuda/SparseCUDABlas.cu b/aten/src/ATen/native/sparse/cuda/SparseCUDABlas.cu
|
||||
index 1cee04c200..f46003d9a9 100644
|
||||
--- a/aten/src/ATen/native/sparse/cuda/SparseCUDABlas.cu
|
||||
+++ b/aten/src/ATen/native/sparse/cuda/SparseCUDABlas.cu
|
||||
@@ -10,48 +10,6 @@
|
||||
namespace at { namespace native { namespace sparse { namespace cuda {
|
||||
|
||||
|
||||
-std::string cusparseGetErrorString(cusparseStatus_t status) {
|
||||
- switch(status)
|
||||
- {
|
||||
- case CUSPARSE_STATUS_SUCCESS:
|
||||
- return "success";
|
||||
-
|
||||
- case CUSPARSE_STATUS_NOT_INITIALIZED:
|
||||
- return "library not initialized";
|
||||
-
|
||||
- case CUSPARSE_STATUS_ALLOC_FAILED:
|
||||
- return "resource allocation failed";
|
||||
-
|
||||
- case CUSPARSE_STATUS_INVALID_VALUE:
|
||||
- return "an invalid numeric value was used as an argument";
|
||||
-
|
||||
- case CUSPARSE_STATUS_ARCH_MISMATCH:
|
||||
- return "an absent device architectural feature is required";
|
||||
-
|
||||
- case CUSPARSE_STATUS_MAPPING_ERROR:
|
||||
- return "an access to GPU memory space failed";
|
||||
-
|
||||
- case CUSPARSE_STATUS_EXECUTION_FAILED:
|
||||
- return "the GPU program failed to execute";
|
||||
-
|
||||
- case CUSPARSE_STATUS_INTERNAL_ERROR:
|
||||
- return "an internal operation failed";
|
||||
-
|
||||
- case CUSPARSE_STATUS_MATRIX_TYPE_NOT_SUPPORTED:
|
||||
- return "the matrix type is not supported by this function";
|
||||
-
|
||||
- case CUSPARSE_STATUS_ZERO_PIVOT:
|
||||
- return "an entry of the matrix is either structural zero or numerical zero (singular block)";
|
||||
-
|
||||
- default:
|
||||
- {
|
||||
- std::ostringstream oss;
|
||||
- oss << "unknown error " << static_cast<int64_t>(status);
|
||||
- return oss.str();
|
||||
- }
|
||||
- }
|
||||
-}
|
||||
-
|
||||
inline void CUSPARSE_CHECK(cusparseStatus_t status)
|
||||
{
|
||||
if (status != CUSPARSE_STATUS_SUCCESS) {
|
@ -3,8 +3,10 @@
|
||||
#
|
||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||
|
||||
from spack import *
|
||||
import os
|
||||
import sys
|
||||
|
||||
from spack import *
|
||||
|
||||
|
||||
class PyTorch(PythonPackage, CudaPackage):
|
||||
@ -21,6 +23,7 @@ class PyTorch(PythonPackage, CudaPackage):
|
||||
import_modules = ['torch', 'torch.autograd', 'torch.nn', 'torch.utils']
|
||||
|
||||
version('master', branch='master', submodules=True)
|
||||
version('1.9.0', tag='v1.9.0', submodules=True)
|
||||
version('1.8.1', tag='v1.8.1', submodules=True)
|
||||
version('1.8.0', tag='v1.8.0', submodules=True)
|
||||
version('1.7.1', tag='v1.7.1', submodules=True)
|
||||
@ -29,7 +32,7 @@ class PyTorch(PythonPackage, CudaPackage):
|
||||
version('1.5.1', tag='v1.5.1', submodules=True)
|
||||
version('1.5.0', tag='v1.5.0', submodules=True)
|
||||
version('1.4.1', tag='v1.4.1', submodules=True)
|
||||
version('1.4.0', tag='v1.4.0', submodules=True,
|
||||
version('1.4.0', tag='v1.4.0', submodules=True, deprecated=True,
|
||||
submodules_delete=['third_party/fbgemm'])
|
||||
version('1.3.1', tag='v1.3.1', submodules=True)
|
||||
version('1.3.0', tag='v1.3.0', submodules=True)
|
||||
@ -37,142 +40,180 @@ class PyTorch(PythonPackage, CudaPackage):
|
||||
version('1.1.0', tag='v1.1.0', submodules=True)
|
||||
version('1.0.1', tag='v1.0.1', submodules=True)
|
||||
version('1.0.0', tag='v1.0.0', submodules=True)
|
||||
version('0.4.1', tag='v0.4.1', submodules=True,
|
||||
version('0.4.1', tag='v0.4.1', submodules=True, deprecated=True,
|
||||
submodules_delete=['third_party/nervanagpu'])
|
||||
version('0.4.0', tag='v0.4.0', submodules=True)
|
||||
version('0.3.1', tag='v0.3.1', submodules=True)
|
||||
version('0.4.0', tag='v0.4.0', submodules=True, deprecated=True)
|
||||
version('0.3.1', tag='v0.3.1', submodules=True, deprecated=True)
|
||||
|
||||
variant('cuda', default=True, description='Build with CUDA')
|
||||
variant('cudnn', default=True, description='Enables the cuDNN build')
|
||||
variant('rocm', default=False, description='Build with ROCm build')
|
||||
variant('magma', default=False, description='Enables the MAGMA build')
|
||||
variant('fbgemm', default=False, description='Enables the FBGEMM build')
|
||||
variant('test', default=False, description='Enables the test build')
|
||||
variant('miopen', default=False, description='Enables the MIOpen build')
|
||||
variant('mkldnn', default=True, description='Enables use of MKLDNN')
|
||||
variant('nnpack', default=False, description='Enables NNPACK build')
|
||||
variant('qnnpack', default=False, description='Enables QNNPACK build (quantized 8-bit operators)')
|
||||
variant('xnnpack', default=False, description='Enables XNNPACK build')
|
||||
variant('distributed', default=False, description='Enables distributed (c10d, gloo, mpi, etc.) build')
|
||||
variant('nccl', default=True, description='Use Spack-installed NCCL')
|
||||
variant('caffe2', default=False, description='Enables Caffe2 operators build')
|
||||
variant('gloo', default=False, description='Enables features related to distributed support')
|
||||
variant('opencv', default=False, description='Enables use of OpenCV for additional operators')
|
||||
variant('openmp', default=True, description='Enables use of OpenMP for parallelization')
|
||||
variant('ffmpeg', default=False, description='Enables use of ffmpeg for additional operators')
|
||||
variant('leveldb', default=False, description='Enables use of LevelDB for storage')
|
||||
variant('lmdb', default=False, description='Enables use of LMDB for storage')
|
||||
variant('binary', default=False, description='Enables the additional binaries/ build')
|
||||
variant('redis', default=False, description='Use Redis for distributed workflows')
|
||||
variant('zstd', default=False, description='Enables use of ZSTD')
|
||||
variant('tbb', default=False, description='Enables TBB support')
|
||||
is_darwin = sys.platform == 'darwin'
|
||||
|
||||
# All options are defined in CMakeLists.txt.
|
||||
# Some are listed in setup.py, but not all.
|
||||
variant('caffe2', default=True, description='Build Caffe2')
|
||||
variant('cuda', default=not is_darwin, description='Use CUDA')
|
||||
variant('rocm', default=False, description='Use ROCm')
|
||||
variant('cudnn', default=not is_darwin, description='Use cuDNN')
|
||||
variant('fbgemm', default=True, description='Use FBGEMM (quantized 8-bit server operators)')
|
||||
variant('kineto', default=True, description='Use Kineto profiling library')
|
||||
variant('magma', default=not is_darwin, description='Use MAGMA')
|
||||
variant('metal', default=is_darwin, description='Use Metal for Caffe2 iOS build')
|
||||
variant('nccl', default=not is_darwin, description='Use NCCL')
|
||||
variant('nnpack', default=True, description='Use NNPACK')
|
||||
variant('numa', default=not is_darwin, description='Use NUMA')
|
||||
variant('numpy', default=True, description='Use NumPy')
|
||||
variant('openmp', default=True, description='Use OpenMP for parallel code')
|
||||
variant('qnnpack', default=True, description='Use QNNPACK (quantized 8-bit operators)')
|
||||
variant('valgrind', default=not is_darwin, description='Use Valgrind')
|
||||
variant('xnnpack', default=True, description='Use XNNPACK')
|
||||
variant('mkldnn', default=True, description='Use MKLDNN')
|
||||
variant('distributed', default=not is_darwin, description='Use distributed')
|
||||
variant('mpi', default=not is_darwin, description='Use MPI for Caffe2')
|
||||
variant('gloo', default=not is_darwin, description='Use Gloo')
|
||||
variant('tensorpipe', default=not is_darwin, description='Use TensorPipe')
|
||||
variant('onnx_ml', default=True, description='Enable traditional ONNX ML API')
|
||||
|
||||
conflicts('+cuda', when='+rocm')
|
||||
conflicts('+cudnn', when='~cuda')
|
||||
conflicts('+magma', when='~cuda')
|
||||
conflicts('+fbgemm', when='@:0.4')
|
||||
conflicts('+miopen', when='@:0.4')
|
||||
conflicts('+mkldnn', when='@:0.3')
|
||||
conflicts('+qnnpack', when='@:0.4')
|
||||
conflicts('+xnnpack', when='@:1.4')
|
||||
conflicts('+nccl', when='~cuda')
|
||||
conflicts('+opencv', when='@:0.4')
|
||||
conflicts('+ffmpeg', when='@:0.4')
|
||||
conflicts('+leveldb', when='@:0.4')
|
||||
conflicts('+lmdb', when='@:0.4')
|
||||
conflicts('+binary', when='@:0.4')
|
||||
conflicts('+redis', when='@:1.0')
|
||||
conflicts('+zstd', when='@:1.0')
|
||||
conflicts('+tbb', when='@:1.1')
|
||||
# https://github.com/pytorch/pytorch/issues/35149
|
||||
conflicts('+fbgemm', when='@1.4.0')
|
||||
conflicts('+nccl', when='~cuda~rocm')
|
||||
conflicts('+nccl', when='platform=darwin')
|
||||
conflicts('+numa', when='platform=darwin', msg='Only available on Linux')
|
||||
conflicts('+valgrind', when='platform=darwin', msg='Only available on Linux')
|
||||
conflicts('+mpi', when='~distributed')
|
||||
conflicts('+gloo', when='~distributed')
|
||||
conflicts('+tensorpipe', when='~distributed')
|
||||
conflicts('+kineto', when='@:1.7.999')
|
||||
conflicts('+valgrind', when='@:1.7.999')
|
||||
conflicts('~caffe2', when='@0.4.0:1.6.999') # no way to disable caffe2?
|
||||
conflicts('+caffe2', when='@:0.3.1') # caffe2 did not yet exist?
|
||||
conflicts('+tensorpipe', when='@:1.5.999')
|
||||
conflicts('+xnnpack', when='@:1.4.999')
|
||||
conflicts('~onnx_ml', when='@:1.4.999') # no way to disable ONNX?
|
||||
conflicts('+rocm', when='@:0.4.999')
|
||||
conflicts('+cudnn', when='@:0.4.999')
|
||||
conflicts('+fbgemm', when='@:0.4.999,1.4.0')
|
||||
conflicts('+qnnpack', when='@:0.4.999')
|
||||
conflicts('+mkldnn', when='@:0.4.999')
|
||||
|
||||
conflicts('cuda_arch=none', when='+cuda',
|
||||
msg='Must specify CUDA compute capabilities of your GPU, see '
|
||||
'https://developer.nvidia.com/cuda-gpus')
|
||||
conflicts('+rocm', when='+cuda')
|
||||
|
||||
# Required dependencies
|
||||
depends_on('cmake@3.5:', type='build')
|
||||
# Use Ninja generator to speed up build times
|
||||
# Automatically used if found
|
||||
depends_on('ninja@1.5:', type='build')
|
||||
# Use Ninja generator to speed up build times, automatically used if found
|
||||
depends_on('ninja@1.5:', when='@1.1.0:', type='build')
|
||||
# See python_min_version in setup.py
|
||||
depends_on('python@3.6.2:', when='@1.7.1:', type=('build', 'link', 'run'))
|
||||
depends_on('python@3.6.1:', when='@1.6:', type=('build', 'link', 'run'))
|
||||
depends_on('python@3.5:', when='@1.5:', type=('build', 'link', 'run'))
|
||||
depends_on('python@2.7:2.8,3.5:', type=('build', 'link', 'run'))
|
||||
depends_on('python@3.6.1:', when='@1.6.0:1.7.0', type=('build', 'link', 'run'))
|
||||
depends_on('python@3.5:', when='@1.5.0:1.5.999', type=('build', 'link', 'run'))
|
||||
depends_on('python@2.7:2.8,3.5:', when='@1.4.0:1.4.999', type=('build', 'link', 'run'))
|
||||
depends_on('python@2.7:2.8,3.5:3.7.999', when='@:1.3.999', type=('build', 'link', 'run'))
|
||||
depends_on('py-setuptools', type=('build', 'run'))
|
||||
depends_on('py-numpy', type=('build', 'run'))
|
||||
depends_on('py-future', when='@1.5:', type=('build', 'run'))
|
||||
depends_on('py-future', when='@1.1: ^python@:2', type=('build', 'run'))
|
||||
depends_on('py-pyyaml', type=('build', 'run'))
|
||||
depends_on('py-typing', when='@0.4: ^python@:3.4', type=('build', 'run'))
|
||||
depends_on('py-typing-extensions', when='@1.7:', type=('build', 'run'))
|
||||
depends_on('py-pybind11', when='@0.4:', type=('build', 'link', 'run'))
|
||||
depends_on('py-pybind11@master', when='@master', type=('build', 'link', 'run'))
|
||||
depends_on('py-pybind11@2.6.2', when='@1.8.0:1.9.999', type=('build', 'link', 'run'))
|
||||
depends_on('py-pybind11@2.3.0', when='@1.1.0:1.7.999', type=('build', 'link', 'run'))
|
||||
depends_on('py-pybind11@2.2.4', when='@1.0.0:1.0.999', type=('build', 'link', 'run'))
|
||||
depends_on('py-pybind11@2.2.2', when='@0.4.0:0.4.999', type=('build', 'link', 'run'))
|
||||
depends_on('py-dataclasses', when='@1.7: ^python@3.6.0:3.6.999', type=('build', 'run'))
|
||||
depends_on('py-tqdm', type='run')
|
||||
depends_on('py-protobuf', when='@0.4:', type=('build', 'run'))
|
||||
depends_on('protobuf', when='@0.4:')
|
||||
depends_on('blas')
|
||||
depends_on('lapack')
|
||||
depends_on('protobuf', when='@0.4:')
|
||||
depends_on('py-protobuf', when='@0.4:', type=('build', 'run'))
|
||||
depends_on('eigen', when='@0.4:')
|
||||
# TODO: replace all third_party packages with Spack packages
|
||||
# https://github.com/pytorch/pytorch/issues/60329
|
||||
# depends_on('cpuinfo@master', when='@master')
|
||||
# depends_on('cpuinfo@2020-12-17', when='@1.8.0:1.9.999')
|
||||
# depends_on('cpuinfo@2020-06-11', when='@1.6.0:1.7.999')
|
||||
depends_on('sleef@master', when='@master')
|
||||
depends_on('sleef@3.5.1_2020-12-22', when='@1.8.0:1.9.999')
|
||||
# https://github.com/pytorch/pytorch/issues/60334
|
||||
# depends_on('sleef@3.4.0_2019-07-30', when='@1.6.0:1.7.999')
|
||||
depends_on('fp16@master', when='@master')
|
||||
depends_on('fp16@2020-05-14', when='@1.6.0:1.9.999')
|
||||
depends_on('pthreadpool@master', when='@master')
|
||||
depends_on('pthreadpool@2021-04-13', when='@1.9.0:1.9.999')
|
||||
depends_on('pthreadpool@2020-10-05', when='@1.8.0:1.8.999')
|
||||
depends_on('pthreadpool@2020-06-15', when='@1.6.0:1.7.999')
|
||||
depends_on('psimd@master', when='@master')
|
||||
depends_on('psimd@2020-05-17', when='@1.6.0:1.9.999')
|
||||
depends_on('fxdiv@master', when='@master')
|
||||
depends_on('fxdiv@2020-04-17', when='@1.6.0:1.9.999')
|
||||
depends_on('benchmark', when='@1.6:')
|
||||
|
||||
# Optional dependencies
|
||||
depends_on('cuda@7.5:', when='+cuda', type=('build', 'link', 'run'))
|
||||
depends_on('cuda@9:', when='@1.1:+cuda', type=('build', 'link', 'run'))
|
||||
depends_on('cudnn@6:', when='+cudnn')
|
||||
depends_on('cudnn@7:', when='@1.1:+cudnn')
|
||||
depends_on('cudnn@6.0:7.999', when='@:1.0.999+cudnn')
|
||||
depends_on('cudnn@7.0:7.999', when='@1.1.0:1.5.999+cudnn')
|
||||
depends_on('cudnn@7.0:', when='@1.6.0:+cudnn')
|
||||
depends_on('magma', when='+magma')
|
||||
# TODO: add dependency: https://github.com/pytorch/FBGEMM
|
||||
# depends_on('fbgemm', when='+fbgemm')
|
||||
# TODO: add dependency: https://github.com/ROCmSoftwarePlatform/MIOpen
|
||||
# depends_on('miopen', when='+miopen')
|
||||
# TODO: See if there is a way to use an external mkldnn installation.
|
||||
# Currently, only older versions of py-torch use an external mkldnn
|
||||
# library.
|
||||
depends_on('onednn', when='@0.4:0.4.1+mkldnn')
|
||||
# TODO: add dependency: https://github.com/Maratyszcza/NNPACK
|
||||
# depends_on('nnpack', when='+nnpack')
|
||||
depends_on('qnnpack', when='+qnnpack')
|
||||
# TODO: add dependency: https://github.com/google/XNNPACK
|
||||
# depends_on('xnnpack', when='+xnnpack')
|
||||
depends_on('mpi', when='+distributed')
|
||||
depends_on('nccl', when='+nccl')
|
||||
depends_on('gloo', when='+gloo')
|
||||
depends_on('opencv', when='+opencv')
|
||||
depends_on('numactl', when='+numa')
|
||||
depends_on('py-numpy', when='+numpy', type=('build', 'run'))
|
||||
depends_on('llvm-openmp', when='%apple-clang +openmp')
|
||||
depends_on('ffmpeg', when='+ffmpeg')
|
||||
depends_on('leveldb', when='+leveldb')
|
||||
depends_on('lmdb', when='+lmdb')
|
||||
depends_on('redis', when='+redis')
|
||||
depends_on('zstd', when='+zstd')
|
||||
depends_on('tbb', when='+tbb')
|
||||
depends_on('valgrind', when='+valgrind')
|
||||
# https://github.com/pytorch/pytorch/issues/60332
|
||||
# depends_on('xnnpack@master', when='@master+xnnpack')
|
||||
# depends_on('xnnpack@2021-02-22', when='@1.8.0:1.9.999+xnnpack')
|
||||
# depends_on('xnnpack@2020-03-23', when='@1.6.0:1.7.999+xnnpack')
|
||||
depends_on('mpi', when='+mpi')
|
||||
# https://github.com/pytorch/pytorch/issues/60270
|
||||
# depends_on('gloo@master', when='@master+gloo')
|
||||
# depends_on('gloo@2021-05-04', when='@1.9.0:1.9.999+gloo')
|
||||
# depends_on('gloo@2020-09-18', when='@1.7.0:1.8.999+gloo')
|
||||
# depends_on('gloo@2020-03-17', when='@1.6.0:1.6.999+gloo')
|
||||
# https://github.com/pytorch/pytorch/issues/60331
|
||||
# depends_on('onnx@master', when='@master+onnx_ml')
|
||||
# depends_on('onnx@1.8.0_2020-11-03', when='@1.8.0:1.9.999+onnx_ml')
|
||||
# depends_on('onnx@1.7.0_2020-05-31', when='@1.6.0:1.7.999+onnx_ml')
|
||||
depends_on('mkl', when='+mkldnn')
|
||||
|
||||
# Test dependencies
|
||||
depends_on('py-hypothesis', type='test')
|
||||
depends_on('py-six', type='test')
|
||||
depends_on('py-psutil', type='test')
|
||||
|
||||
# Fix BLAS being overridden by MKL
|
||||
# https://github.com/pytorch/pytorch/issues/60328
|
||||
patch('https://patch-diff.githubusercontent.com/raw/pytorch/pytorch/pull/59220.patch',
|
||||
sha256='e37afffe45cf7594c22050109942370e49983ad772d12ebccf508377dc9dcfc9',
|
||||
when='@1.2.0:')
|
||||
|
||||
# Fixes build on older systems with glibc <2.12
|
||||
patch('https://patch-diff.githubusercontent.com/raw/pytorch/pytorch/pull/55063.patch',
|
||||
sha256='e17eaa42f5d7c18bf0d7c37d7b0910127a01ad53fdce3e226a92893356a70395',
|
||||
when='@1.1.0:')
|
||||
when='@1.1.0:1.8.1')
|
||||
|
||||
# Fixes CMake configuration error when XNNPACK is disabled
|
||||
# https://github.com/pytorch/pytorch/pull/35607
|
||||
# https://github.com/pytorch/pytorch/pull/37865
|
||||
# Fixes CMake configuration error when XNNPACK is disabled
|
||||
patch('xnnpack.patch', when='@1.5.0:1.5.999')
|
||||
|
||||
# Fixes Build error for when ROCm is enable for pytorch-1.5 release
|
||||
# Fixes build error when ROCm is enabled for pytorch-1.5 release
|
||||
patch('rocm.patch', when='@1.5.0:1.5.999+rocm')
|
||||
|
||||
# https://github.com/pytorch/pytorch/pull/37086
|
||||
# Fixes fatal error: sleef.h: No such file or directory
|
||||
# https://github.com/pytorch/pytorch/pull/35359
|
||||
# https://github.com/pytorch/pytorch/issues/26555
|
||||
patch('sleef.patch', when='@1.0.0:1.5.999')
|
||||
|
||||
# Fixes compilation with Clang 9.0.0 and Apple Clang 11.0.3
|
||||
# https://github.com/pytorch/pytorch/pull/37086
|
||||
patch('https://github.com/pytorch/pytorch/commit/e921cd222a8fbeabf5a3e74e83e0d8dfb01aa8b5.patch',
|
||||
sha256='17561b16cd2db22f10c0fe1fdcb428aecb0ac3964ba022a41343a6bb8cba7049',
|
||||
when='@1.1:1.5')
|
||||
|
||||
# Fix for 'FindOpenMP.cmake'
|
||||
# Removes duplicate definition of getCusparseErrorString
|
||||
# https://github.com/pytorch/pytorch/issues/32083
|
||||
patch('cusparseGetErrorString.patch', when='@0.4.1:1.0.999^cuda@10.1.243:')
|
||||
|
||||
# Fixes 'FindOpenMP.cmake'
|
||||
# to detect openmp settings used by Fujitsu compiler.
|
||||
patch('detect_omp_of_fujitsu_compiler.patch', when='%fj')
|
||||
|
||||
@ -182,16 +223,12 @@ class PyTorch(PythonPackage, CudaPackage):
|
||||
|
||||
@property
|
||||
def libs(self):
|
||||
root = join_path(
|
||||
self.prefix, self.spec['python'].package.site_packages_dir,
|
||||
'torch', 'lib')
|
||||
root = join_path(site_packages_dir, 'torch', 'lib')
|
||||
return find_libraries('libtorch', root)
|
||||
|
||||
@property
|
||||
def headers(self):
|
||||
root = join_path(
|
||||
self.prefix, self.spec['python'].package.site_packages_dir,
|
||||
'torch', 'include')
|
||||
root = join_path(site_packages_dir, 'torch', 'include')
|
||||
headers = find_all_headers(root)
|
||||
headers.directories = [root]
|
||||
return headers
|
||||
@ -204,6 +241,16 @@ def patch(self):
|
||||
'caffe2/CMakeLists.txt')
|
||||
|
||||
def setup_build_environment(self, env):
|
||||
"""Set environment variables used to control the build.
|
||||
|
||||
PyTorch's ``setup.py`` is a thin wrapper around ``cmake``.
|
||||
In ``tools/setup_helpers/cmake.py``, you can see that all
|
||||
environment variables that start with ``BUILD_``, ``USE_``,
|
||||
or ``CMAKE_``, plus a few more explicitly specified variable
|
||||
names, are passed directly to the ``cmake`` call. Therefore,
|
||||
most flags defined in ``CMakeLists.txt`` can be specified as
|
||||
environment variables.
|
||||
"""
|
||||
def enable_or_disable(variant, keyword='USE', var=None, newer=False):
|
||||
"""Set environment variable to enable or disable support for a
|
||||
particular variant.
|
||||
@ -236,84 +283,129 @@ def enable_or_disable(variant, keyword='USE', var=None, newer=False):
|
||||
# Spack logs have trouble handling colored output
|
||||
env.set('COLORIZE_OUTPUT', 'OFF')
|
||||
|
||||
# Don't use vendored third-party libraries
|
||||
env.set('BUILD_CUSTOM_PROTOBUF', 'OFF')
|
||||
env.set('USE_PYTORCH_QNNPACK', 'OFF')
|
||||
env.set('USE_SYSTEM_EIGEN_INSTALL', 'ON')
|
||||
env.set('pybind11_DIR', self.spec['py-pybind11'].prefix)
|
||||
env.set('pybind11_INCLUDE_DIR',
|
||||
self.spec['py-pybind11'].prefix.include)
|
||||
if self.spec.satisfies('@1.7:'):
|
||||
enable_or_disable('caffe2', keyword='BUILD')
|
||||
|
||||
enable_or_disable('cuda')
|
||||
if '+cuda' in self.spec:
|
||||
env.set('CUDA_HOME', self.spec['cuda'].prefix)
|
||||
# cmake/public/cuda.cmake
|
||||
# cmake/Modules_CUDA_fix/upstream/FindCUDA.cmake
|
||||
env.unset('CUDA_ROOT')
|
||||
torch_cuda_arch = ';'.join('{0:.1f}'.format(float(i) / 10.0) for i
|
||||
in
|
||||
self.spec.variants['cuda_arch'].value)
|
||||
env.set('TORCH_CUDA_ARCH_LIST', torch_cuda_arch)
|
||||
|
||||
enable_or_disable('rocm')
|
||||
|
||||
enable_or_disable('cudnn')
|
||||
if '+cudnn' in self.spec:
|
||||
env.set('CUDNN_LIB_DIR', self.spec['cudnn'].libs.directories[0])
|
||||
# cmake/Modules_CUDA_fix/FindCUDNN.cmake
|
||||
env.set('CUDNN_INCLUDE_DIR', self.spec['cudnn'].prefix.include)
|
||||
env.set('CUDNN_LIBRARY', self.spec['cudnn'].libs[0])
|
||||
|
||||
enable_or_disable('fbgemm')
|
||||
enable_or_disable('test', keyword='BUILD')
|
||||
enable_or_disable('rocm')
|
||||
if '+rocm' in self.spec:
|
||||
env.set('USE_MKLDNN', 0)
|
||||
if '+miopen' in self.spec:
|
||||
env.set('MIOPEN_LIB_DIR', self.spec['miopen'].libs.directories[0])
|
||||
env.set('MIOPEN_INCLUDE_DIR', self.spec['miopen'].prefix.include)
|
||||
env.set('MIOPEN_LIBRARY', self.spec['miopen'].libs[0])
|
||||
|
||||
enable_or_disable('mkldnn')
|
||||
if '@0.4:0.4.1+mkldnn' in self.spec:
|
||||
env.set('MKLDNN_HOME', self.spec['onednn'].prefix)
|
||||
|
||||
enable_or_disable('nnpack')
|
||||
enable_or_disable('qnnpack')
|
||||
enable_or_disable('xnnpack')
|
||||
enable_or_disable('distributed')
|
||||
if self.spec.satisfies('@1.8:'):
|
||||
enable_or_disable('kineto')
|
||||
enable_or_disable('magma')
|
||||
enable_or_disable('metal')
|
||||
|
||||
enable_or_disable('nccl')
|
||||
enable_or_disable('nccl', var='SYSTEM_NCCL')
|
||||
if '+nccl' in self.spec:
|
||||
env.set('NCCL_ROOT', self.spec['nccl'].prefix)
|
||||
env.set('NCCL_LIB_DIR', self.spec['nccl'].libs.directories[0])
|
||||
env.set('NCCL_INCLUDE_DIR', self.spec['nccl'].prefix.include)
|
||||
|
||||
enable_or_disable('caffe2', keyword='BUILD', var='CAFFE2_OPS')
|
||||
enable_or_disable('gloo', newer=True)
|
||||
enable_or_disable('opencv', newer=True)
|
||||
# cmake/External/nnpack.cmake
|
||||
enable_or_disable('nnpack')
|
||||
|
||||
enable_or_disable('numa')
|
||||
if '+numa' in self.spec:
|
||||
# cmake/Modules/FindNuma.cmake
|
||||
env.set('NUMA_ROOT_DIR', self.spec['numactl'].prefix)
|
||||
|
||||
# cmake/Modules/FindNumPy.cmake
|
||||
enable_or_disable('numpy')
|
||||
# cmake/Modules/FindOpenMP.cmake
|
||||
enable_or_disable('openmp', newer=True)
|
||||
enable_or_disable('ffmpeg', newer=True)
|
||||
enable_or_disable('leveldb', newer=True)
|
||||
enable_or_disable('lmdb', newer=True)
|
||||
enable_or_disable('binary', keyword='BUILD', newer=True)
|
||||
enable_or_disable('qnnpack')
|
||||
if self.spec.satisfies('@1.3:'):
|
||||
enable_or_disable('qnnpack', var='PYTORCH_QNNPACK')
|
||||
if self.spec.satisfies('@1.8:'):
|
||||
enable_or_disable('valgrind')
|
||||
if self.spec.satisfies('@1.5:'):
|
||||
enable_or_disable('xnnpack')
|
||||
enable_or_disable('mkldnn')
|
||||
enable_or_disable('distributed')
|
||||
enable_or_disable('mpi')
|
||||
# cmake/Modules/FindGloo.cmake
|
||||
enable_or_disable('gloo', newer=True)
|
||||
if self.spec.satisfies('@1.6:'):
|
||||
enable_or_disable('tensorpipe')
|
||||
|
||||
if '+onnx_ml' in self.spec:
|
||||
env.set('ONNX_ML', 'ON')
|
||||
else:
|
||||
env.set('ONNX_ML', 'OFF')
|
||||
|
||||
if not self.spec.satisfies('@master'):
|
||||
env.set('PYTORCH_BUILD_VERSION', self.version)
|
||||
env.set('PYTORCH_BUILD_NUMBER', 0)
|
||||
|
||||
# BLAS to be used by Caffe2
|
||||
if '^mkl' in self.spec:
|
||||
env.set('BLAS', 'MKL')
|
||||
elif '^atlas' in self.spec:
|
||||
# Options defined in cmake/Dependencies.cmake and cmake/Modules/FindBLAS.cmake
|
||||
if self.spec['blas'].name == 'atlas':
|
||||
env.set('BLAS', 'ATLAS')
|
||||
elif '^openblas' in self.spec:
|
||||
env.set('BLAS', 'OpenBLAS')
|
||||
elif '^veclibfort' in self.spec:
|
||||
env.set('BLAS', 'vecLib')
|
||||
elif '^libflame' in self.spec:
|
||||
env.set('BLAS', 'FLAME')
|
||||
elif '^eigen' in self.spec:
|
||||
env.set('WITH_BLAS', 'atlas')
|
||||
elif self.spec['blas'].name in ['blis', 'amdblis']:
|
||||
env.set('BLAS', 'BLIS')
|
||||
env.set('WITH_BLAS', 'blis')
|
||||
elif self.spec['blas'].name == 'eigen':
|
||||
env.set('BLAS', 'Eigen')
|
||||
elif self.spec['lapack'].name in ['libflame', 'amdlibflame']:
|
||||
env.set('BLAS', 'FLAME')
|
||||
env.set('WITH_BLAS', 'FLAME')
|
||||
elif self.spec['blas'].name in [
|
||||
'intel-mkl', 'intel-parallel-studio', 'intel-oneapi-mkl']:
|
||||
env.set('BLAS', 'MKL')
|
||||
env.set('WITH_BLAS', 'mkl')
|
||||
elif self.spec['blas'].name == 'openblas':
|
||||
env.set('BLAS', 'OpenBLAS')
|
||||
env.set('WITH_BLAS', 'open')
|
||||
elif self.spec['blas'].name == 'veclibfort':
|
||||
env.set('BLAS', 'vecLib')
|
||||
env.set('WITH_BLAS', 'veclib')
|
||||
else:
|
||||
env.set('BLAS', 'Generic')
|
||||
env.set('WITH_BLAS', 'generic')
|
||||
|
||||
enable_or_disable('redis', newer=True)
|
||||
enable_or_disable('zstd', newer=True)
|
||||
enable_or_disable('tbb', newer=True)
|
||||
# Don't use vendored third-party libraries when possible
|
||||
env.set('BUILD_CUSTOM_PROTOBUF', 'OFF')
|
||||
env.set('USE_SYSTEM_NCCL', 'ON')
|
||||
env.set('USE_SYSTEM_EIGEN_INSTALL', 'ON')
|
||||
if self.spec.satisfies('@0.4:'):
|
||||
env.set('pybind11_DIR', self.spec['py-pybind11'].prefix)
|
||||
env.set('pybind11_INCLUDE_DIR',
|
||||
self.spec['py-pybind11'].prefix.include)
|
||||
if self.spec.satisfies('@1.10:'):
|
||||
env.set('USE_SYSTEM_PYBIND11', 'ON')
|
||||
# https://github.com/pytorch/pytorch/issues/60334
|
||||
if self.spec.satisfies('@1.8:'):
|
||||
env.set('USE_SYSTEM_SLEEF', 'ON')
|
||||
if self.spec.satisfies('@1.6:'):
|
||||
# env.set('USE_SYSTEM_LIBS', 'ON')
|
||||
# https://github.com/pytorch/pytorch/issues/60329
|
||||
# env.set('USE_SYSTEM_CPUINFO', 'ON')
|
||||
# https://github.com/pytorch/pytorch/issues/60270
|
||||
# env.set('USE_SYSTEM_GLOO', 'ON')
|
||||
env.set('USE_SYSTEM_FP16', 'ON')
|
||||
env.set('USE_SYSTEM_PTHREADPOOL', 'ON')
|
||||
env.set('USE_SYSTEM_PSIMD', 'ON')
|
||||
env.set('USE_SYSTEM_FXDIV', 'ON')
|
||||
env.set('USE_SYSTEM_BENCHMARK', 'ON')
|
||||
# https://github.com/pytorch/pytorch/issues/60331
|
||||
# env.set('USE_SYSTEM_ONNX', 'ON')
|
||||
# https://github.com/pytorch/pytorch/issues/60332
|
||||
# env.set('USE_SYSTEM_XNNPACK', 'ON')
|
||||
|
||||
@run_before('install')
|
||||
def build_amd(self):
|
||||
|
12
var/spack/repos/builtin/packages/py-torch/sleef.patch
Normal file
12
var/spack/repos/builtin/packages/py-torch/sleef.patch
Normal file
@ -0,0 +1,12 @@
|
||||
diff --git a/caffe2/CMakeLists.txt b/caffe2/CMakeLists.txt
|
||||
index 8025a7de3c..2e5cdbb5c9 100644
|
||||
--- a/caffe2/CMakeLists.txt
|
||||
+++ b/caffe2/CMakeLists.txt
|
||||
@@ -1232,6 +1232,7 @@ if (BUILD_TEST)
|
||||
add_executable(${test_name} "${test_src}")
|
||||
target_link_libraries(${test_name} ${Caffe2_MAIN_LIBS} gtest_main)
|
||||
target_include_directories(${test_name} PRIVATE $<INSTALL_INTERFACE:include>)
|
||||
+ target_include_directories(${test_name} PRIVATE $<BUILD_INTERFACE:${CMAKE_BINARY_DIR}/include>)
|
||||
target_include_directories(${test_name} PRIVATE ${Caffe2_CPU_INCLUDE})
|
||||
add_test(NAME ${test_name} COMMAND $<TARGET_FILE:${test_name}>)
|
||||
if (INSTALL_TEST)
|
Loading…
Reference in New Issue
Block a user