
Many noqa's in the code are no longer necessary now that the column limit is 99 characters. Others can easily be eliminated, and still more can just be made more specific if they do not have to do with line length. The only E501's still in the code are in the tests for `spack.util.path` and the tests for `spack style`.
916 lines
45 KiB
Python
916 lines
45 KiB
Python
# Copyright 2013-2022 Lawrence Livermore National Security, LLC and other
|
|
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
|
#
|
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
|
|
|
import sys
|
|
import tempfile
|
|
|
|
from spack.package import *
|
|
|
|
|
|
class PyTensorflow(Package, CudaPackage):
|
|
"""TensorFlow is an Open Source Software Library for Machine Intelligence
|
|
"""
|
|
|
|
homepage = "https://www.tensorflow.org"
|
|
url = "https://github.com/tensorflow/tensorflow/archive/v2.3.1.tar.gz"
|
|
|
|
maintainers = ['adamjstewart', 'aweits']
|
|
import_modules = ['tensorflow']
|
|
|
|
version('2.7.0', sha256='bb124905c7fdacd81e7c842b287c169bbf377d29c74c9dacc04f96c9793747bb')
|
|
version('2.6.2', sha256='e68c1d346fc3d529653530ca346b2c62f5b31bd4fcca7ffc9c65bb39ab2f6ed3')
|
|
version('2.6.1', sha256='8e457f617bc2eb43de2a51900e7922b60a8107e2524b2576438f1acccee1d043')
|
|
version('2.6.0', sha256='41b32eeaddcbc02b0583660bcf508469550e4cd0f86b22d2abe72dfebeacde0f')
|
|
version('2.5.0', sha256='233875ea27fc357f6b714b2a0de5f6ff124b50c1ee9b3b41f9e726e9e677b86c')
|
|
version('2.4.1', sha256='f681331f8fc0800883761c7709d13cda11942d4ad5ff9f44ad855e9dc78387e0')
|
|
version('2.4.0', sha256='26c833b7e1873936379e810a39d14700281125257ddda8cd822c89111db6f6ae')
|
|
version('2.3.2', sha256='21a703d2e68cd0677f6f9ce329198c24fd8203125599d791af9f1de61aadf31f')
|
|
version('2.3.1', sha256='ee534dd31a811f7a759453567257d1e643f216d8d55a25c32d2fbfff8153a1ac')
|
|
version('2.3.0', sha256='2595a5c401521f20a2734c4e5d54120996f8391f00bb62a57267d930bce95350')
|
|
version('2.2.2', sha256='fb4b5d26c5b983350f7ce8297b71176a86a69e91faf66e6ebb1e58538ad3bb51')
|
|
version('2.2.1', sha256='e6a28e64236d729e598dbeaa02152219e67d0ac94d6ed22438606026a02e0f88')
|
|
version('2.2.0', sha256='69cd836f87b8c53506c4f706f655d423270f5a563b76dc1cfa60fbc3184185a3')
|
|
version('2.1.3', sha256='cfa66cce372f486e95a42beb1aacfefdaf0092c5efaaaa92459b381fde931fb8')
|
|
version('2.1.2', sha256='3f941cda0ed12dfef5472e46f1d0238ea85da7583d73f1132d2ef050fda6e8ad')
|
|
version('2.1.1', sha256='a200bc16e4b630db3ac7225bcb6f239a76841967b0aec1d7d7bbe44dc5661318')
|
|
version('2.1.0', sha256='638e541a4981f52c69da4a311815f1e7989bf1d67a41d204511966e1daed14f7')
|
|
version('2.0.4', sha256='6ca3ce1255da8d655080a89db10da03f72c361d7faecc9a35e6af26ff12c06e6')
|
|
version('2.0.3', sha256='6314299a723441bd9892e5c2af182c2be7d2256e20e71026e1cb1264cb497f33')
|
|
version('2.0.2', sha256='a548742bbafd302eec51e2794d7687674a64f6b10ce1414073858cb83c0cefc2')
|
|
version('2.0.1', sha256='29197d30923b9670992ee4b9c6161f50c7452e9a4158c720746e846080ac245a')
|
|
version('2.0.0', sha256='49b5f0495cd681cbcb5296a4476853d4aea19a43bdd9f179c928a977308a0617')
|
|
version('1.15.5', sha256='4c4d23e311093ded2d2e287b18d7c45b07b5984ab88a1d2f91f8f13c886123db')
|
|
version('1.15.4', sha256='e18c55e771ad136f9bf3a70ea8f0e2d36662b2ba7c890f9eaf7950554557c7fa')
|
|
version('1.15.3', sha256='9ab1d92e58eb813922b040acc7622b32d73c2d8d971fe6491a06f9df4c778151')
|
|
version('1.15.2', sha256='d95d75d26a298211b5e802842e87fda5b8b14f6ad83719377b391e5fb71b8746')
|
|
version('1.15.1', sha256='19b6e72bc8675937f618cede364d7228a71c2eeaffc42801bcefd98dda7ca056')
|
|
version('1.15.0', sha256='a5d49c00a175a61da7431a9b289747d62339be9cf37600330ad63b611f7f5dc9')
|
|
version('1.14.0', sha256='aa2a6a1daafa3af66807cfe0bc77bfe1144a9a53df9a96bab52e3e575b3047ed')
|
|
version('1.13.2', sha256='abe3bf0c47845a628b7df4c57646f41a10ee70f914f1b018a5c761be75e1f1a9')
|
|
version('1.13.1', sha256='7cd19978e6bc7edc2c847bce19f95515a742b34ea5e28e4389dade35348f58ed')
|
|
version('1.12.3', sha256='b9e5488e84f4a133ed20b18605f0cd6301f11d356bd959712db4e7b9301d0462')
|
|
version('1.12.2', sha256='90ffc7cf1df5e4b8385c9108db18d5d5034ec423547c0e167d44f5746a20d06b')
|
|
version('1.12.1', sha256='7b559a3ae56322b7a7e4307f45f9fce96022c533a98b32c18bfdff8c5838271d')
|
|
version('1.12.0', sha256='3c87b81e37d4ed7f3da6200474fa5e656ffd20d8811068572f43610cae97ca92')
|
|
version('1.11.0', sha256='f49ce3f1d04cee854bc9f74fa9696991140b34a2e2447f35f01391b72c8bfa9f')
|
|
version('1.10.1', sha256='83092d709800e2d93d4d4b1bcacaeb74f2f328962ed764cb35bbee20402879c6')
|
|
version('1.10.0', sha256='ee9cb98d9e0d8106f2f4ed52a38fe89399324af303e1401567e5b64a9f86744b')
|
|
version('1.9.0', sha256='ffc3151b06823d57b4a408261ba8efe53601563dfe93af0866751d4f6ca5068c')
|
|
version('1.8.0', sha256='47646952590fd213b747247e6870d89bb4a368a95ae3561513d6c76e44f92a75')
|
|
version('1.7.1', sha256='3147f8c60d1f30da23a831bcf732e74b935dcee7c62e4b8b85f0f093030b52c8')
|
|
version('1.7.0', sha256='c676a96fc8700722816b2b98c85578b2f99fac7a7b2484c9c7f0641484f8d50d')
|
|
version('1.6.0', sha256='03cf1423446abbead6bd8c3cf6e6affa7d99746cd119691b012aac9a1795f4fb')
|
|
version('1.5.1', sha256='cab2157783905e12a7a3baae3264edfb739dd92d5658019a131fff4b14190240')
|
|
version('1.5.0', sha256='0642781c3a3a8c2c4834b91b86aec385f0b2ada7d721571458079478cc5b29c8')
|
|
version('1.4.1', sha256='1f75e463318419a1b3ae076d5a92697c1d3a85e8377c946a5510b651ff5c0d60')
|
|
version('1.4.0', sha256='8a0ad8d61f8f6c0282c548661994a5ab83ac531bac496c3041dedc1ab021107b')
|
|
version('1.3.1', sha256='ded509c209f8a1d390df8a2f44be5b5c29963172b0e0f095304efb59765d0523')
|
|
version('1.3.0', sha256='e1af1bb767b57c3416de0d43a5f74d174c42b85231dffd36f3630173534d4307')
|
|
version('1.2.1', sha256='f2baf09b1a9a0500907b4d5cb5473070b3ecede06ed6e8d1096873c91922fb9e')
|
|
version('1.2.0', sha256='03dbf7548d1fc1c11ed58da5fa68616f795c819f868f43478cbcaa26abed374f')
|
|
version('1.1.0', sha256='aad4470f52fa59f54de7b9a2da727429e6755d91d756f245f952698c42a60027')
|
|
version('1.0.1', sha256='deea3c65e0703da96d9c3f1162e464c51d37659dd129396af134e9e8f1ea8c05')
|
|
version('1.0.0', sha256='db8b3b8f4134b7c9c1b4165492ad5d5bb78889fcd99ffdffc325e97da3e8c677')
|
|
version('0.12.0', sha256='13a1d4e98c82eae7e26fe75384de1517d6126f63ba5d302392ec02ac3ae4b1b9')
|
|
version('0.11.0', sha256='24242ff696234bb1e58d09d45169b148525ccb706f980a4a92ddd3b82c7546dc')
|
|
version('0.10.0', sha256='f32df04e8f7186aaf6723fc5396733b2f6c2fd6fe4a53a54a68b80f3ec855680')
|
|
version('0.9.0', sha256='3128c396af19518c642d3e590212291e1d93c5b047472a10cf3245b53adac9c9')
|
|
version('0.8.0', sha256='f201ba7fb7609a6416968d4e1920d87d67be693b5bc7d34b6b4a79860a9a8a4e')
|
|
version('0.7.1', sha256='ef34121432f7a522cf9f99a56cdd86e370cc5fa3ee31255ca7cb17f36b8dfc0d')
|
|
version('0.7.0', sha256='43dd3051f947aa66e6fc09dac2f86a2efe2e019736bbd091c138544b86d717ce')
|
|
version('0.6.0', sha256='f86ace45e99053b09749cd55ab79c57274d8c7460ae763c5e808d81ffbc3b657')
|
|
|
|
variant('mkl', default=False, description='Build with MKL support')
|
|
variant('jemalloc', default=False, description='Build with jemalloc as malloc support')
|
|
variant('gcp', default=False, description='Build with Google Cloud Platform support')
|
|
variant('hdfs', default=False, description='Build with Hadoop File System support')
|
|
variant('aws', default=False, description='Build with Amazon AWS Platform support')
|
|
variant('kafka', default=False, description='Build with Apache Kafka Platform support')
|
|
variant('ignite', default=False, description='Build with Apache Ignite support')
|
|
variant('xla', default=False, description='Build with XLA JIT support')
|
|
variant('gdr', default=False, description='Build with GDR support')
|
|
variant('verbs', default=False, description='Build with libverbs support')
|
|
variant('ngraph', default=False, description='Build with Intel nGraph support')
|
|
variant('opencl', default=False, description='Build with OpenCL SYCL support')
|
|
variant('computecpp', default=False, description='Build with ComputeCPP support')
|
|
variant('rocm', default=False, description='Build with ROCm support')
|
|
variant('tensorrt', default=False, description='Build with TensorRT support')
|
|
variant('cuda', default=sys.platform != 'darwin', description='Build with CUDA support')
|
|
variant('nccl', default=sys.platform.startswith('linux'), description='Enable NVIDIA NCCL support')
|
|
variant('mpi', default=False, description='Build with MPI support')
|
|
variant('android', default=False, description='Configure for Android builds')
|
|
variant('ios', default=False, description='Build with iOS support (macOS only)')
|
|
variant('monolithic', default=False, description='Static monolithic build')
|
|
variant('numa', default=False, description='Build with NUMA support')
|
|
variant('dynamic_kernels', default=False, description='Build kernels into separate shared objects')
|
|
|
|
extends('python')
|
|
depends_on('python@3:', type=('build', 'run'), when='@2.1:')
|
|
# python 3.8 support in tensorflow 2.2
|
|
# see tensorflow issue #33374
|
|
depends_on('python@:3.7', type=('build', 'run'), when='@:2.2')
|
|
|
|
# TODO: Older versions of TensorFlow don't list the viable version range,
|
|
# just the minimum version of bazel that will work. The latest version of
|
|
# bazel doesn't seem to work, so for now we force them to use min version.
|
|
# Need to investigate further.
|
|
|
|
# See _TF_MIN_BAZEL_VERSION and _TF_MAX_BAZEL_VERSION in configure.py
|
|
depends_on('bazel@3.7.2:4.99.0', type='build', when='@2.7:')
|
|
depends_on('bazel@3.1.0:3.99.0', type='build', when='@2.3:2.6')
|
|
depends_on('bazel@2.0.0', type='build', when='@2.2.0:2.2')
|
|
depends_on('bazel@0.27.1:0.29.1', type='build', when='@2.1.0:2.1')
|
|
depends_on('bazel@0.24.1:0.26.1', type='build', when='@1.15:2.0')
|
|
# See call to check_bazel_version in configure.py
|
|
depends_on('bazel@0.24.1:0.25.2', type='build', when='@1.14.0')
|
|
depends_on('bazel@0.19.0:0.21.0', type='build', when='@1.13.0:1.13.2')
|
|
depends_on('bazel@0.24.1:0.25.0', type='build', when='@1.12.1')
|
|
depends_on('bazel@0.15.0', type='build', when='@1.10:1.12.0,1.12.2:1.12.3')
|
|
depends_on('bazel@0.10.0', type='build', when='@1.7:1.9')
|
|
# See call to check_version in tensorflow/workspace.bzl
|
|
depends_on('bazel@0.5.4', type='build', when='@1.4:1.6')
|
|
# See MIN_BAZEL_VERSION in configure
|
|
depends_on('bazel@0.4.5', type='build', when='@1.2:1.3')
|
|
# See call to check_version in WORKSPACE
|
|
depends_on('bazel@0.4.2', type='build', when='@1.0:1.1')
|
|
depends_on('bazel@0.3.2', type='build', when='@0.12.0:0.12.1')
|
|
depends_on('bazel@0.3.0', type='build', when='@0.11.0')
|
|
depends_on('bazel@0.2.0', type='build', when='@0.9:0.10')
|
|
depends_on('bazel@0.1.4', type='build', when='@0.7:0.8')
|
|
depends_on('bazel@0.1.1', type='build', when='@0.5:0.6')
|
|
|
|
depends_on('swig', type='build')
|
|
depends_on('py-pip', type='build')
|
|
depends_on('py-wheel', type='build')
|
|
depends_on('py-setuptools', type='build')
|
|
depends_on('py-future', type='build', when='^python@:2')
|
|
|
|
# Listed under REQUIRED_PACKAGES in tensorflow/tools/pip_package/setup.py
|
|
depends_on('py-absl-py@0.4:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-absl-py@0.10:0', type=('build', 'run'), when='@2.4.0:2.6')
|
|
depends_on('py-absl-py@0.7.0:', type=('build', 'run'), when='@1.12.1,1.14:2.3')
|
|
depends_on('py-absl-py@0.1.6:', type=('build', 'run'), when='@1.5:1.11')
|
|
depends_on('py-astunparse@1.6.0:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-astunparse@1.6.3:1.6', type=('build', 'run'), when='@2.4.0:2.6')
|
|
depends_on('py-astunparse@1.6.3', type=('build', 'run'), when='@2.2:2.3')
|
|
depends_on('py-astor@0.6.0:', type=('build', 'run'), when='@1.6:2.1')
|
|
depends_on('py-backports-weakref@1.0:', type=('build', 'run'), when='@1.3: ^python@:3.3')
|
|
depends_on('py-backports-weakref@1.0rc1', type=('build', 'run'), when='@1.2.0:1.2.1')
|
|
depends_on('py-libclang@9.0.1:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-enum34@1.1.6:', type=('build', 'run'), when='@1.5: ^python@:3.3')
|
|
depends_on('py-enum34@1.1.6:', type=('build', 'run'), when='@1.4.0:1.4.1')
|
|
depends_on('py-gast@0.2.1:0.4', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-gast@0.4.0', type=('build', 'run'), when='@2.5:2.6')
|
|
depends_on('py-gast@0.3.3', type=('build', 'run'), when='@2.2:2.4')
|
|
depends_on('py-gast@0.2.2', type=('build', 'run'), when='@1.15:2.1')
|
|
depends_on('py-gast@0.2.0:', type=('build', 'run'), when='@1.6:1.14')
|
|
depends_on('py-google-pasta@0.1.1:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-google-pasta@0.2:0', type=('build', 'run'), when='@2.4.0:2.6')
|
|
depends_on('py-google-pasta@0.1.8:', type=('build', 'run'), when='@2.1:2.3')
|
|
depends_on('py-google-pasta@0.1.6:', type=('build', 'run'), when='@1.14:2.0')
|
|
depends_on('py-google-pasta@0.1.2:', type=('build', 'run'), when='@1.12.1')
|
|
# propagate the mpi variant setting for h5py/hdf5 to avoid unexpected crashes
|
|
depends_on('py-h5py@2.9:+mpi', type=('build', 'run'), when='@2.7:+mpi')
|
|
depends_on('py-h5py@2.9:~mpi', type=('build', 'run'), when='@2.7:~mpi')
|
|
depends_on('py-h5py@3.1+mpi', type=('build', 'run'), when='@2.5:2.6+mpi')
|
|
depends_on('py-h5py@3.1~mpi', type=('build', 'run'), when='@2.5:2.6~mpi')
|
|
depends_on('py-h5py@2.10+mpi', type=('build', 'run'), when='@2.2:2.4+mpi')
|
|
depends_on('py-h5py@2.10~mpi', type=('build', 'run'), when='@2.2:2.4~mpi')
|
|
depends_on('hdf5+mpi', type='build', when='@2.2:+mpi')
|
|
depends_on('hdf5~mpi', type='build', when='@2.2:~mpi')
|
|
depends_on('py-keras-applications@1.0.8:', type=('build', 'run'), when='@1.15:2.1')
|
|
depends_on('py-keras-applications@1.0.6:', type=('build', 'run'), when='@1.12:1.14')
|
|
depends_on('py-keras-applications@1.0.5:', type=('build', 'run'), when='@1.11.0:1.11')
|
|
depends_on('py-keras-preprocessing@1.1.1:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-keras-preprocessing@1.1.2:1.1', type=('build', 'run'), when='@2.4:2.6')
|
|
depends_on('py-keras-preprocessing@1.1.1:1', type=('build', 'run'), when='@2.3.0:2.3')
|
|
depends_on('py-keras-preprocessing@1.1.0:', type=('build', 'run'), when='@2.1:2.2')
|
|
depends_on('py-keras-preprocessing@1.0.5:', type=('build', 'run'), when='@1.12:2.0')
|
|
depends_on('py-keras-preprocessing@1.0.3:', type=('build', 'run'), when='@1.11.0:1.11')
|
|
# https://github.com/tensorflow/tensorflow/issues/40688
|
|
depends_on('py-numpy@1.14.5:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-numpy@1.19.2:1.19', type=('build', 'run'), when='@2.4:2.6')
|
|
depends_on('py-numpy@1.16.0:1.18', type=('build', 'run'), when='@1.13.2,1.15:2.3')
|
|
depends_on('py-numpy@1.14.5:1.18', type=('build', 'run'), when='@1.12.1,1.14.0')
|
|
depends_on('py-numpy@1.13.3:1.14.5', type=('build', 'run'), when='@1.10.0:1.10.1')
|
|
depends_on('py-numpy@1.13.3:', type=('build', 'run'), when='@1.6:1.9')
|
|
depends_on('py-numpy@1.12.1:', type=('build', 'run'), when='@1.4:1.5')
|
|
depends_on('py-numpy@1.11.0:', type=('build', 'run'), when='@0.11:1.3')
|
|
depends_on('py-numpy@1.10.1:', type=('build', 'run'), when='@0.7.1:0.7 platform=darwin')
|
|
depends_on('py-numpy@1.8.2:', type=('build', 'run'), when='@0.6:0.10')
|
|
depends_on('py-numpy@1.9.2:', type=('build', 'run'), when='@0.5.0')
|
|
depends_on('py-opt-einsum@3.3.0:3.3', type=('build', 'run'), when='@2.4.0:2.6')
|
|
depends_on('py-opt-einsum@2.3.2:', type=('build', 'run'), when='@1.15:2.3,2.7:')
|
|
depends_on('py-protobuf@3.9.2:', type=('build', 'run'), when='@2.3:')
|
|
depends_on('py-protobuf@3.8.0:', type=('build', 'run'), when='@2.1:2.2')
|
|
depends_on('py-protobuf@3.6.1:', type=('build', 'run'), when='@1.12:2.0')
|
|
depends_on('py-protobuf@3.6.0:', type=('build', 'run'), when='@1.10:1.11')
|
|
depends_on('py-protobuf@3.4.0:', type=('build', 'run'), when='@1.5:1.9')
|
|
depends_on('py-protobuf@3.3.0:', type=('build', 'run'), when='@1.3:1.4')
|
|
depends_on('py-protobuf@3.2.0:', type=('build', 'run'), when='@1.1:1.2')
|
|
depends_on('py-protobuf@3.1.0:', type=('build', 'run'), when='@0.12.1:1.0')
|
|
depends_on('py-protobuf@3.1.0', type=('build', 'run'), when='@0.12.0')
|
|
depends_on('py-protobuf@3.0.0', type=('build', 'run'), when='@0.11.0')
|
|
depends_on('py-protobuf@3.0.0b2', type=('build', 'run'), when='@0.7.1:0.10')
|
|
depends_on('py-protobuf@3.0.0a3', type=('build', 'run'), when='@0.6:0.7.0')
|
|
depends_on('protobuf@:3.12', when='@:2.4')
|
|
depends_on('protobuf@:3.17')
|
|
# https://github.com/protocolbuffers/protobuf/issues/10051
|
|
# https://github.com/tensorflow/tensorflow/issues/56266
|
|
depends_on('py-protobuf@:3', when='@:2.7.2', type=('build', 'run'))
|
|
depends_on('protobuf@:3', when='@:2.7.2', type=('build', 'run'))
|
|
depends_on('flatbuffers+python@1.12:2', type=('build', 'run'), when='@2.7:')
|
|
depends_on('flatbuffers+python@1.12', type=('build', 'run'), when='@2.4:2.6')
|
|
|
|
# tensorboard
|
|
depends_on('py-tensorboard@2.5.0:2.5', type=('build', 'run'), when='@2.5')
|
|
depends_on('py-tensorboard@2.6.0:2.6', type=('build', 'run'), when='@2.6')
|
|
depends_on('py-tensorboard@2.7.0:2.7', type=('build', 'run'), when='@2.7')
|
|
|
|
depends_on('py-termcolor@1.1.0:1.1', type=('build', 'run'), when='@2.4:2.6')
|
|
depends_on('py-termcolor@1.1.0:', type=('build', 'run'), when='@1.6:2.3,2.7:')
|
|
depends_on('py-wrapt@1.11.0:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-wrapt@1.12.1:1.12', type=('build', 'run'), when='@2.4:2.6')
|
|
depends_on('py-wrapt@1.11.1:', type=('build', 'run'), when='@1.12.1,1.14:2.3')
|
|
depends_on('py-wheel@0.32:0', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-wheel@0.35:0', type=('build', 'run'), when='@2.4:2.6 ^python@3:')
|
|
depends_on('py-wheel', type=('build', 'run'), when='@0.6:2.3')
|
|
depends_on('py-wheel@0.26:', type=('build', 'run'), when='@0.6:2.3 ^python@3:')
|
|
depends_on('py-mock@2.0.0:', type=('build', 'run'), when='@0.10: ^python@:2')
|
|
depends_on('py-functools32@3.2.3:', type=('build', 'run'), when='@1.15: ^python@:2')
|
|
depends_on('py-six@1.15.0:1.15', type=('build', 'run'), when='@2.4:2.6')
|
|
depends_on('py-six@1.12.0:', type=('build', 'run'), when='@2.1:2.3,2.7:')
|
|
depends_on('py-six@1.10.0:', type=('build', 'run'), when='@:2.0')
|
|
depends_on('py-scipy@1.2.2', type=('build', 'run'), when='@2.1.0:2.1.1,2.2.0,2.3.0 ^python@:2')
|
|
depends_on('py-scipy@1.4.1', type=('build', 'run'), when='@2.1.0:2.1.1,2.2.0,2.3.0 ^python@3:')
|
|
depends_on('py-grpcio@1.8.6:', type=('build', 'run'), when='@1.6:1.7')
|
|
depends_on('py-typing-extensions@3.6.6:', type=('build', 'run'), when='@2.7:')
|
|
depends_on('py-typing-extensions@3.7.4:3.7', type=('build', 'run'), when='@2.4.0:2.6')
|
|
|
|
if sys.byteorder == 'little':
|
|
# Only builds correctly on little-endian machines
|
|
depends_on('py-grpcio@1.8.6:', type=('build', 'run'), when='@1.8:2.3')
|
|
depends_on('py-grpcio@1.32.0:1.32', type=('build', 'run'), when='@2.4.0:2.4')
|
|
depends_on('py-grpcio@1.34.0:1.34', type=('build', 'run'), when='@2.5.0:2.5')
|
|
depends_on('py-grpcio@1.37.0:1', type=('build', 'run'), when='@2.6.0:2.6')
|
|
depends_on('py-grpcio@1.24.3:1', type=('build', 'run'), when='@2.7.0:')
|
|
|
|
# TODO: add packages for some of these dependencies
|
|
depends_on('mkl', when='+mkl')
|
|
depends_on('curl', when='+gcp')
|
|
# depends_on('computecpp', when='+opencl+computecpp')
|
|
# depends_on('trisycl', when='+opencl~computepp')
|
|
depends_on('cuda@:10.2', when='+cuda @:2.3')
|
|
depends_on('cuda@:11.4', when='+cuda @2.4:')
|
|
depends_on('cudnn', when='+cuda')
|
|
depends_on('cudnn@:6', when='@0.5:0.6 +cuda')
|
|
depends_on('cudnn@:7', when='@0.7:2.2 +cuda')
|
|
# depends_on('tensorrt', when='+tensorrt')
|
|
depends_on('nccl', when='+nccl')
|
|
depends_on('mpi', when='+mpi')
|
|
# depends_on('android-ndk@10:18', when='+android')
|
|
# depends_on('android-sdk', when='+android')
|
|
|
|
# TODO: add support for tensorflow-io-gcs-filesystem
|
|
# depends_on('tensorflow-io-gcs-filesystem
|
|
|
|
# Check configure and configure.py to see when these variants are supported
|
|
conflicts('+mkl', when='@:1.0')
|
|
conflicts('+mkl', when='platform=darwin', msg='Darwin is not yet supported')
|
|
conflicts('+jemalloc', when='@:0')
|
|
conflicts('+jemalloc', when='platform=darwin', msg='Currently jemalloc is only support on Linux platform')
|
|
conflicts('+jemalloc', when='platform=cray', msg='Currently jemalloc is only support on Linux platform')
|
|
conflicts('+gcp', when='@:0.8')
|
|
conflicts('+hdfs', when='@:0.10')
|
|
conflicts('+aws', when='@:1.3')
|
|
conflicts('+kafka', when='@:1.5,2.1:')
|
|
conflicts('+ignite', when='@:1.11,2.1:')
|
|
conflicts('+xla', when='@:0')
|
|
conflicts('+gdr', when='@:1.3')
|
|
conflicts('+verbs', when='@:1.1')
|
|
conflicts('+ngraph', when='@:1.10')
|
|
conflicts('+opencl', when='@:0.11')
|
|
conflicts('+computecpp', when='@:0.11')
|
|
conflicts('+computecpp', when='~opencl')
|
|
conflicts('+rocm', when='@:1.11')
|
|
conflicts('+cuda', when='platform=darwin', msg='There is no GPU support for macOS')
|
|
conflicts('cuda_arch=none', when='+cuda', msg='Must specify CUDA compute capabilities of your GPU, see https://developer.nvidia.com/cuda-gpus')
|
|
conflicts('cuda_arch=20', when='@1.12.1,1.14:', msg='TensorFlow only supports compute capabilities >= 3.5')
|
|
conflicts('cuda_arch=30', when='@1.12.1,1.14:', msg='TensorFlow only supports compute capabilities >= 3.5')
|
|
conflicts('cuda_arch=32', when='@1.12.1,1.14:', msg='TensorFlow only supports compute capabilities >= 3.5')
|
|
conflicts('cuda_arch=20', when='@1.4:1.12.0,1.12.2:1.12.3', msg='Only compute capabilities 3.0 or higher are supported')
|
|
conflicts('+tensorrt', when='@:1.5')
|
|
conflicts('+tensorrt', when='~cuda')
|
|
conflicts('+tensorrt', when='platform=darwin', msg='Currently TensorRT is only supported on Linux platform')
|
|
conflicts('+tensorrt', when='platform=cray', msg='Currently TensorRT is only supported on Linux platform')
|
|
conflicts('+nccl', when='@:1.7')
|
|
conflicts('+nccl', when='~cuda')
|
|
conflicts('+nccl', when='platform=darwin', msg='Currently NCCL is only supported on Linux platform')
|
|
conflicts('+nccl', when='platform=cray', msg='Currently NCCL is only supported on Linux platform')
|
|
conflicts('+mpi', when='@:1.2')
|
|
conflicts('+android', when='@:1.4')
|
|
conflicts('+ios', when='@:1.12.0,1.12.2:1.13')
|
|
conflicts('+ios', when='platform=linux', msg='iOS support only available on macOS')
|
|
conflicts('+ios', when='platform=cray', msg='iOS support only available on macOS')
|
|
conflicts('+monolithic', when='@:1.3')
|
|
conflicts('+numa', when='@:1.12.0,1.12.2:1.13')
|
|
conflicts('+dynamic_kernels', when='@:1.12.0,1.12.2:1.12.3')
|
|
|
|
# TODO: why is this needed?
|
|
patch('url-zlib.patch', when='@0.10.0')
|
|
# TODO: why is this needed?
|
|
patch('crosstool.patch', when='@0.10.0+cuda')
|
|
# Avoid build error: "no such package '@io_bazel_rules_docker..."
|
|
patch('io_bazel_rules_docker2.patch', when='@1.15:2.0')
|
|
# Avoide build error: "name 'new_http_archive' is not defined"
|
|
patch('http_archive.patch', when='@1.12.3')
|
|
# Backport of 837c8b6b upstream
|
|
# "Remove contrib cloud bigtable and storage ops/kernels."
|
|
# Allows 2.0.* releases to build with '--config=nogcp'
|
|
patch('0001-Remove-contrib-cloud-bigtable-and-storage-ops-kernel.patch',
|
|
when='@2.0.0:2.0')
|
|
|
|
# for fcc
|
|
patch('1-1_fcc_tf_patch.patch', when='@2.1.0:2.1%fj')
|
|
|
|
# do not import contrib.cloud if not available
|
|
patch('https://github.com/tensorflow/tensorflow/commit/ed62ac8203999513dfae03498e871ea35eb60cc4.patch?full_index=1',
|
|
sha256='ff02e249532a5661b123108734a39534992d81da90f0c8187bf4e151a865effc', when='@1.14.0')
|
|
# import_contrib_cloud patch for older versions
|
|
patch('contrib_cloud_1.10.patch', when='@1.10:1.13')
|
|
patch('contrib_cloud_1.9.patch', when='@1.9')
|
|
patch('contrib_cloud_1.4.patch', when='@1.4:1.8')
|
|
patch('contrib_cloud_1.1.patch', when='@1.1:1.3')
|
|
|
|
# needed for protobuf-3.16 and greater
|
|
patch('example_parsing.patch', when='^protobuf@3.16:')
|
|
|
|
# allow linker to be found in PATH
|
|
# https://github.com/tensorflow/tensorflow/issues/39263
|
|
patch('null_linker_bin_path.patch', when='@2.5:')
|
|
|
|
# Reset import order to that of 2.4. Part of
|
|
# https://bugs.gentoo.org/800824#c3 From the patch:
|
|
# When tensorflow and python protobuf use the same instance of libprotobuf,
|
|
# pywrap_tensorflow must be imported before anything else that would import
|
|
# protobuf definitions.
|
|
patch('0008-Fix-protobuf-errors-when-using-system-protobuf.patch', when='@2.5:2.6')
|
|
|
|
phases = ['configure', 'build', 'install']
|
|
|
|
# https://www.tensorflow.org/install/source
|
|
def setup_build_environment(self, env):
|
|
spec = self.spec
|
|
|
|
# Please specify the location of python
|
|
env.set('PYTHON_BIN_PATH', spec['python'].command.path)
|
|
|
|
# Please input the desired Python library path to use
|
|
env.set('PYTHON_LIB_PATH', python_platlib)
|
|
|
|
# Ensure swig is in PATH or set SWIG_PATH
|
|
env.set('SWIG_PATH', spec['swig'].prefix.bin.swig)
|
|
|
|
# Do you wish to build TensorFlow with MKL support?
|
|
if '+mkl' in spec:
|
|
env.set('TF_NEED_MKL', '1')
|
|
|
|
# Do you wish to download MKL LIB from the web?
|
|
env.set('TF_DOWNLOAD_MKL', '0')
|
|
|
|
# Please specify the location where MKL is installed
|
|
env.set('MKL_INSTALL_PATH', spec['mkl'].prefix)
|
|
else:
|
|
env.set('TF_NEED_MKL', '0')
|
|
|
|
# Do you wish to build TensorFlow with jemalloc as malloc support?
|
|
if '+jemalloc' in spec:
|
|
env.set('TF_NEED_JEMALLOC', '1')
|
|
else:
|
|
env.set('TF_NEED_JEMALLOC', '0')
|
|
|
|
# Do you wish to build TensorFlow with Google Cloud Platform support?
|
|
if '+gcp' in spec:
|
|
env.set('TF_NEED_GCP', '1')
|
|
else:
|
|
env.set('TF_NEED_GCP', '0')
|
|
|
|
# Do you wish to build TensorFlow with Hadoop File System support?
|
|
if '+hdfs' in spec:
|
|
env.set('TF_NEED_HDFS', '1')
|
|
else:
|
|
env.set('TF_NEED_HDFS', '0')
|
|
|
|
# Do you wish to build TensorFlow with Amazon AWS Platform support?
|
|
if '+aws' in spec:
|
|
env.set('TF_NEED_AWS', '1')
|
|
env.set('TF_NEED_S3', '1')
|
|
else:
|
|
env.set('TF_NEED_AWS', '0')
|
|
env.set('TF_NEED_S3', '0')
|
|
|
|
# Do you wish to build TensorFlow with Apache Kafka Platform support?
|
|
if '+kafka' in spec:
|
|
env.set('TF_NEED_KAFKA', '1')
|
|
else:
|
|
env.set('TF_NEED_KAFKA', '0')
|
|
|
|
# Do you wish to build TensorFlow with Apache Ignite support?
|
|
if '+ignite' in spec:
|
|
env.set('TF_NEED_IGNITE', '1')
|
|
else:
|
|
env.set('TF_NEED_IGNITE', '0')
|
|
|
|
# Do you wish to build TensorFlow with XLA JIT support?
|
|
if '+xla' in spec:
|
|
env.set('TF_ENABLE_XLA', '1')
|
|
else:
|
|
env.set('TF_ENABLE_XLA', '0')
|
|
|
|
# Do you wish to build TensorFlow with GDR support?
|
|
if '+gdr' in spec:
|
|
env.set('TF_NEED_GDR', '1')
|
|
else:
|
|
env.set('TF_NEED_GDR', '0')
|
|
|
|
# Do you wish to build TensorFlow with VERBS support?
|
|
if '+verbs' in spec:
|
|
env.set('TF_NEED_VERBS', '1')
|
|
else:
|
|
env.set('TF_NEED_VERBS', '0')
|
|
|
|
# Do you wish to build TensorFlow with nGraph support?
|
|
if '+ngraph' in spec:
|
|
env.set('TF_NEED_NGRAPH', '1')
|
|
else:
|
|
env.set('TF_NEED_NGRAPH', '0')
|
|
|
|
# Do you wish to build TensorFlow with OpenCL SYCL support?
|
|
if '+opencl' in spec:
|
|
env.set('TF_NEED_OPENCL_SYCL', '1')
|
|
env.set('TF_NEED_OPENCL', '1')
|
|
|
|
# Please specify which C++ compiler should be used as the host
|
|
# C++ compiler
|
|
env.set('HOST_CXX_COMPILER', spack_cxx)
|
|
|
|
# Please specify which C compiler should be used as the host
|
|
# C compiler
|
|
env.set('HOST_C_COMPILER', spack_cc)
|
|
|
|
# Do you wish to build TensorFlow with ComputeCPP support?
|
|
if '+computecpp' in spec:
|
|
env.set('TF_NEED_COMPUTECPP', '1')
|
|
|
|
# Please specify the location where ComputeCpp is installed
|
|
env.set('COMPUTECPP_TOOLKIT_PATH', spec['computecpp'].prefix)
|
|
else:
|
|
env.set('TF_NEED_COMPUTECPP', '0')
|
|
|
|
# Please specify the location of the triSYCL include directory
|
|
env.set('TRISYCL_INCLUDE_DIR', spec['trisycl'].prefix.include)
|
|
else:
|
|
env.set('TF_NEED_OPENCL_SYCL', '0')
|
|
env.set('TF_NEED_OPENCL', '0')
|
|
|
|
# Do you wish to build TensorFlow with ROCm support?
|
|
if '+rocm' in spec:
|
|
env.set('TF_NEED_ROCM', '1')
|
|
else:
|
|
env.set('TF_NEED_ROCM', '0')
|
|
|
|
# Do you wish to build TensorFlow with CUDA support?
|
|
if '+cuda' in spec:
|
|
env.set('TF_NEED_CUDA', '1')
|
|
|
|
# Do you want to use clang as CUDA compiler?
|
|
env.set('TF_CUDA_CLANG', '0')
|
|
|
|
# Please specify which gcc nvcc should use as the host compiler
|
|
env.set('GCC_HOST_COMPILER_PATH', spack_cc)
|
|
|
|
cuda_paths = [
|
|
spec['cuda'].prefix,
|
|
spec['cudnn'].prefix,
|
|
]
|
|
|
|
# Do you wish to build TensorFlow with TensorRT support?
|
|
if '+tensorrt' in spec:
|
|
env.set('TF_NEED_TENSORRT', '1')
|
|
|
|
cuda_paths.append(spec['tensorrt'].prefix)
|
|
|
|
# Please specify the TensorRT version you want to use
|
|
env.set('TF_TENSORRT_VERSION',
|
|
spec['tensorrt'].version.up_to(1))
|
|
|
|
# Please specify the location where TensorRT is installed
|
|
env.set('TENSORRT_INSTALL_PATH', spec['tensorrt'].prefix)
|
|
else:
|
|
env.set('TF_NEED_TENSORRT', '0')
|
|
env.unset('TF_TENSORRT_VERSION')
|
|
|
|
# Please specify the CUDA SDK version you want to use
|
|
env.set('TF_CUDA_VERSION', spec['cuda'].version.up_to(2))
|
|
|
|
# Please specify the cuDNN version you want to use
|
|
env.set('TF_CUDNN_VERSION', spec['cudnn'].version.up_to(1))
|
|
|
|
if '+nccl' in spec:
|
|
cuda_paths.append(spec['nccl'].prefix)
|
|
|
|
# Please specify the locally installed NCCL version to use
|
|
env.set('TF_NCCL_VERSION', spec['nccl'].version.up_to(1))
|
|
|
|
# Please specify the location where NCCL is installed
|
|
env.set('NCCL_INSTALL_PATH', spec['nccl'].prefix)
|
|
env.set('NCCL_HDR_PATH', spec['nccl'].prefix.include)
|
|
else:
|
|
env.unset('TF_NCCL_VERSION')
|
|
|
|
# Please specify the comma-separated list of base paths to
|
|
# look for CUDA libraries and headers
|
|
env.set('TF_CUDA_PATHS', ','.join(cuda_paths))
|
|
|
|
# Please specify the location where CUDA toolkit is installed
|
|
env.set('CUDA_TOOLKIT_PATH', spec['cuda'].prefix)
|
|
|
|
# Please specify the location where CUDNN library is installed
|
|
env.set('CUDNN_INSTALL_PATH', spec['cudnn'].prefix)
|
|
|
|
# Please specify a list of comma-separated CUDA compute
|
|
# capabilities you want to build with. You can find the compute
|
|
# capability of your device at:
|
|
# https://developer.nvidia.com/cuda-gpus.
|
|
# Please note that each additional compute capability significantly
|
|
# increases your build time and binary size, and that TensorFlow
|
|
# only supports compute capabilities >= 3.5
|
|
capabilities = ','.join('{0:.1f}'.format(
|
|
float(i) / 10.0) for i in spec.variants['cuda_arch'].value)
|
|
env.set('TF_CUDA_COMPUTE_CAPABILITIES', capabilities)
|
|
else:
|
|
env.set('TF_NEED_CUDA', '0')
|
|
|
|
# Do you wish to download a fresh release of clang? (Experimental)
|
|
env.set('TF_DOWNLOAD_CLANG', '0')
|
|
|
|
# Do you wish to build TensorFlow with MPI support?
|
|
if '+mpi' in spec:
|
|
env.set('TF_NEED_MPI', '1')
|
|
|
|
# Please specify the MPI toolkit folder
|
|
env.set('MPI_HOME', spec['mpi'].prefix)
|
|
else:
|
|
env.set('TF_NEED_MPI', '0')
|
|
env.unset('MPI_HOME')
|
|
|
|
# Please specify optimization flags to use during compilation when
|
|
# bazel option '--config=opt' is specified
|
|
env.set('CC_OPT_FLAGS', spec.target.optimization_flags(
|
|
spec.compiler.name, spec.compiler.version))
|
|
|
|
# Would you like to interactively configure ./WORKSPACE for
|
|
# Android builds?
|
|
if '+android' in spec:
|
|
env.set('TF_SET_ANDROID_WORKSPACE', '1')
|
|
|
|
# Please specify the home path of the Android NDK to use
|
|
env.set('ANDROID_NDK_HOME', spec['android-ndk'].prefix)
|
|
env.set('ANDROID_NDK_API_LEVEL', spec['android-ndk'].version)
|
|
|
|
# Please specify the home path of the Android SDK to use
|
|
env.set('ANDROID_SDK_HOME', spec['android-sdk'].prefix)
|
|
env.set('ANDROID_SDK_API_LEVEL', spec['android-sdk'].version)
|
|
|
|
# Please specify the Android SDK API level to use
|
|
env.set('ANDROID_API_LEVEL', spec['android-sdk'].version)
|
|
|
|
# Please specify an Android build tools version to use
|
|
env.set('ANDROID_BUILD_TOOLS_VERSION', spec['android-sdk'].version)
|
|
else:
|
|
env.set('TF_SET_ANDROID_WORKSPACE', '0')
|
|
|
|
# Do you wish to build TensorFlow with iOS support?
|
|
if '+ios' in spec:
|
|
env.set('TF_CONFIGURE_IOS', '1')
|
|
else:
|
|
env.set('TF_CONFIGURE_IOS', '0')
|
|
|
|
# set tmpdir to a non-NFS filesystem
|
|
# (because bazel uses ~/.cache/bazel)
|
|
# TODO: This should be checked for non-nfsy filesystem, but the current
|
|
# best idea for it is to check
|
|
# subprocess.call([
|
|
# 'stat', '--file-system', '--format=%T', tmp_path
|
|
# ])
|
|
# to not be nfs. This is only valid for Linux and we'd like to
|
|
# stay at least also OSX compatible
|
|
tmp_path = tempfile.mkdtemp(prefix='spack')
|
|
env.set('TEST_TMPDIR', tmp_path)
|
|
|
|
env.set('TF_SYSTEM_LIBS', 'com_google_protobuf')
|
|
if spec.satisfies('@:2.3'):
|
|
# NOTE: INCLUDEDIR is not just relevant to protobuf
|
|
# see third_party/systemlibs/jsoncpp.BUILD
|
|
env.set('INCLUDEDIR', spec['protobuf'].prefix.include)
|
|
|
|
def patch(self):
|
|
if self.spec.satisfies('@2.3.0:'):
|
|
filter_file('deps = protodeps + well_known_proto_libs(),',
|
|
'deps = protodeps,',
|
|
'tensorflow/core/platform/default/build_config.bzl',
|
|
string=True)
|
|
if self.spec.satisfies('@2.4.0:2.5'):
|
|
text = '''
|
|
def protobuf_deps():
|
|
pass
|
|
'''
|
|
with open('third_party/systemlibs/protobuf_deps.bzl', 'w') as f:
|
|
f.write(text)
|
|
|
|
if self.spec.satisfies('@2.5.0'):
|
|
file_to_patch = 'tensorflow/workspace2.bzl'
|
|
else:
|
|
file_to_patch = 'tensorflow/workspace.bzl'
|
|
|
|
filter_file(
|
|
'"//third_party/systemlibs:protobuf.bzl": "protobuf.bzl",',
|
|
'"//third_party/systemlibs:protobuf.bzl": "protobuf.bzl",\n'
|
|
'"//third_party/systemlibs:protobuf_deps.bzl": "protobuf_deps.bzl",',
|
|
file_to_patch,
|
|
string=True)
|
|
|
|
# Set protobuf path
|
|
filter_file(r'(^build:linux --define=PROTOBUF_INCLUDE_PATH=).*',
|
|
r'\1{0}'.format(self.spec['protobuf'].prefix.include),
|
|
'.bazelrc')
|
|
|
|
def configure(self, spec, prefix):
|
|
# NOTE: configure script is interactive. If you set the appropriate
|
|
# environment variables, this interactivity is skipped. If you don't,
|
|
# Spack hangs during the configure phase. Use `spack build-env` to
|
|
# determine which environment variables must be set for a particular
|
|
# version.
|
|
configure()
|
|
|
|
@run_after('configure')
|
|
def post_configure_fixes(self):
|
|
spec = self.spec
|
|
|
|
# make sure xla is actually turned off
|
|
if spec.satisfies('~xla'):
|
|
filter_file(
|
|
r'--define with_xla_support=true',
|
|
r'--define with_xla_support=false',
|
|
'.tf_configure.bazelrc')
|
|
|
|
if spec.satisfies('@1.5.0: ~android'):
|
|
# env variable is somehow ignored -> brute force
|
|
# TODO: find a better solution
|
|
filter_file(
|
|
r'if workspace_has_any_android_rule\(\)',
|
|
r'if True',
|
|
'configure.py')
|
|
|
|
# version dependent fixes
|
|
if spec.satisfies('@1.3.0:1.5.0'):
|
|
# checksum for protobuf that bazel downloads (@github) changed
|
|
filter_file(
|
|
r'sha256 = "6d43b9d223ce09e5d4ce8b0060cb8a7513577a35a64c7e3dad10f0703bf3ad93"',
|
|
r'sha256 = "e5fdeee6b28cf6c38d61243adff06628baa434a22b5ebb7432d2a7fbabbdb13d"',
|
|
'tensorflow/workspace.bzl')
|
|
|
|
# starting with tensorflow 1.3, tensorboard becomes a dependency
|
|
# -> remove from list of required packages
|
|
filter_file(
|
|
r"'tensorflow-tensorboard",
|
|
r"#'tensorflow-tensorboard",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@1.5.0: ~gcp'):
|
|
# google cloud support seems to be installed on default, leading
|
|
# to boringssl error manually set the flag to false to avoid
|
|
# installing gcp support
|
|
# https://github.com/tensorflow/tensorflow/issues/20677#issuecomment-404634519
|
|
filter_file(
|
|
r'--define with_gcp_support=true',
|
|
r'--define with_gcp_support=false',
|
|
'.tf_configure.bazelrc')
|
|
|
|
if spec.satisfies('@1.6.0:2.1'):
|
|
# tensorboard name changed
|
|
# there are no corresponding versions of these in spack
|
|
filter_file(
|
|
r"(^\s*)'tensorboard (>=|~=)",
|
|
r"\1#'tensorboard \2",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@1.8.0: ~opencl'):
|
|
# 1.8.0 and 1.9.0 aborts with numpy import error during python_api
|
|
# generation somehow the wrong PYTHONPATH is used...
|
|
# set --distinct_host_configuration=false as a workaround
|
|
# https://github.com/tensorflow/tensorflow/issues/22395#issuecomment-431229451
|
|
with open('.tf_configure.bazelrc', mode='a') as f:
|
|
f.write('build --distinct_host_configuration=false\n')
|
|
f.write('build --action_env PYTHONPATH="{0}"\n'.format(
|
|
env['PYTHONPATH']))
|
|
|
|
if spec.satisfies('@1.13.1:'):
|
|
# tensorflow_estimator is an API for tensorflow
|
|
# tensorflow-estimator imports tensorflow during build, so
|
|
# tensorflow has to be set up first
|
|
filter_file(
|
|
r"(^\s*)'tensorflow_estimator (>=|~=)",
|
|
r"\1#'tensorflow_estimator \2",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@2.5'):
|
|
filter_file(
|
|
r"(^\s*)'keras-nightly (>=|~=)",
|
|
r"\1#'keras-nightly \2",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@2.6:'):
|
|
filter_file(
|
|
r"(^\s*)'keras (>=|~=)",
|
|
r"\1#'keras \2",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@2.6'):
|
|
filter_file(
|
|
r"(^\s*)'clang (>=|~=)",
|
|
r"\1#'clang \2",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
# TODO: add support for tensorflow-io-gcs-filesystem
|
|
if spec.satisfies('@2.7:'):
|
|
filter_file(
|
|
r"(^\s*)'tensorflow-io-gcs-filesystem (>=|~=)",
|
|
r"\1#'tensorflow-io-gcs-filesystem \2",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@2.0.0:'):
|
|
# now it depends on the nightly versions...
|
|
filter_file(
|
|
r"REQUIRED_PACKAGES\[i\] = 'tb-nightly (>=|~=)",
|
|
r"pass #REQUIRED_PACKAGES[i] = 'tb-nightly \1",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
filter_file(
|
|
r"REQUIRED_PACKAGES\[i\] = 'tensorflow-estimator-2.0-preview",
|
|
r"pass #REQUIRED_PACKAGES[i] = 'tensorflow-estimator-2.0-preview",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
filter_file(
|
|
r"REQUIRED_PACKAGES\[i\] = 'tf-estimator-nightly (>=|~=)",
|
|
r"pass #REQUIRED_PACKAGES[i] = 'tf-estimator-nightly \1",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
filter_file(
|
|
r"REQUIRED_PACKAGES\[i\] = 'keras-nightly (>=|~=)",
|
|
r"pass #REQUIRED_PACKAGES[i] = 'keras-nightly \1",
|
|
'tensorflow/tools/pip_package/setup.py')
|
|
|
|
if spec.satisfies('@1.13.1 +nccl'):
|
|
filter_file(
|
|
r'^build --action_env NCCL_INSTALL_PATH=.*',
|
|
r'build --action_env NCCL_INSTALL_PATH="' +
|
|
spec['nccl'].libs.directories[0] + '"',
|
|
'.tf_configure.bazelrc')
|
|
filter_file(
|
|
r'^build --action_env NCCL_HDR_PATH=.*',
|
|
r'build --action_env NCCL_HDR_PATH="' +
|
|
spec['nccl'].prefix.include + '"',
|
|
'.tf_configure.bazelrc')
|
|
|
|
# see tensorflow issue #31187 on github
|
|
if spec.satisfies('@2.0.0:2.0'):
|
|
filter_file(
|
|
r'\#define RUY_DONOTUSEDIRECTLY_AVX512 1',
|
|
'#define RUY_DONOTUSEDIRECTLY_AVX512 0',
|
|
'tensorflow/lite/experimental/ruy/platform.h')
|
|
|
|
if spec.satisfies('+cuda'):
|
|
libs = spec['cuda'].libs.directories
|
|
libs.extend(spec['cudnn'].libs.directories)
|
|
if '+nccl' in spec:
|
|
libs.extend(spec['nccl'].libs.directories)
|
|
|
|
if '+tensorrt' in spec:
|
|
libs.extend(spec['tensorrt'].libs.directories)
|
|
slibs = ':'.join(libs)
|
|
|
|
with open('.tf_configure.bazelrc', mode='a') as f:
|
|
f.write('build --action_env LD_LIBRARY_PATH="' + slibs + '"')
|
|
|
|
filter_file('build:opt --copt=-march=native', '',
|
|
'.tf_configure.bazelrc')
|
|
filter_file('build:opt --host_copt=-march=native', '',
|
|
'.tf_configure.bazelrc')
|
|
|
|
def build(self, spec, prefix):
|
|
tmp_path = env['TEST_TMPDIR']
|
|
|
|
# https://docs.bazel.build/versions/master/command-line-reference.html
|
|
args = [
|
|
# Don't allow user or system .bazelrc to override build settings
|
|
'--nohome_rc',
|
|
'--nosystem_rc',
|
|
# Bazel does not work properly on NFS, switch to /tmp
|
|
'--output_user_root=' + tmp_path,
|
|
'build',
|
|
# Spack logs don't handle colored output well
|
|
'--color=no',
|
|
'--jobs={0}'.format(make_jobs),
|
|
'--config=opt',
|
|
# Enable verbose output for failures
|
|
'--verbose_failures',
|
|
# Show (formatted) subcommands being executed
|
|
'--subcommands=pretty_print',
|
|
# Ask bazel to explain what it's up to
|
|
# Needs a filename as argument
|
|
'--explain=explainlogfile.txt',
|
|
# Increase verbosity of explanation,
|
|
'--verbose_explanations',
|
|
]
|
|
|
|
if spec.satisfies('^bazel@:3.5'):
|
|
# removed in bazel 3.6
|
|
args.append('--incompatible_no_support_tools_in_action_inputs=false')
|
|
|
|
# See .bazelrc for when each config flag is supported
|
|
if spec.satisfies('@1.12.1:'):
|
|
if '+mkl' in spec:
|
|
args.append('--config=mkl')
|
|
|
|
if '+monolithic' in spec:
|
|
args.append('--config=monolithic')
|
|
|
|
if '+gdr' in spec:
|
|
args.append('--config=gdr')
|
|
|
|
if '+verbs' in spec:
|
|
args.append('--config=verbs')
|
|
|
|
if '+ngraph' in spec:
|
|
args.append('--config=ngraph')
|
|
|
|
if '+dynamic_kernels' in spec:
|
|
args.append('--config=dynamic_kernels')
|
|
|
|
if '+cuda' in spec:
|
|
args.append('--config=cuda')
|
|
|
|
if '~aws' in spec:
|
|
args.append('--config=noaws')
|
|
|
|
if '~gcp' in spec:
|
|
args.append('--config=nogcp')
|
|
|
|
if '~hdfs' in spec:
|
|
args.append('--config=nohdfs')
|
|
|
|
if '~nccl' in spec:
|
|
args.append('--config=nonccl')
|
|
|
|
if spec.satisfies('@1.12.1:2.0'):
|
|
if '~ignite' in spec:
|
|
args.append('--config=noignite')
|
|
|
|
if '~kafka' in spec:
|
|
args.append('--config=nokafka')
|
|
|
|
if spec.satisfies('@1.12.1,1.14:'):
|
|
if '+numa' in spec:
|
|
args.append('--config=numa')
|
|
|
|
if spec.satisfies('@2:'):
|
|
args.append('--config=v2')
|
|
|
|
args.append('//tensorflow/tools/pip_package:build_pip_package')
|
|
|
|
bazel(*args)
|
|
|
|
build_pip_package = Executable(
|
|
'bazel-bin/tensorflow/tools/pip_package/build_pip_package')
|
|
buildpath = join_path(self.stage.source_path, 'spack-build')
|
|
build_pip_package('--src', buildpath)
|
|
|
|
def install(self, spec, prefix):
|
|
tmp_path = env['TEST_TMPDIR']
|
|
buildpath = join_path(self.stage.source_path, 'spack-build')
|
|
with working_dir(buildpath):
|
|
args = std_pip_args + ['--prefix=' + prefix, '.']
|
|
pip(*args)
|
|
remove_linked_tree(tmp_path)
|
|
|
|
def test(self):
|
|
"""Attempts to import modules of the installed package."""
|
|
|
|
# Make sure we are importing the installed modules,
|
|
# not the ones in the source directory
|
|
for module in self.import_modules:
|
|
self.run_test(self.spec['python'].command.path,
|
|
['-c', 'import {0}'.format(module)],
|
|
purpose='checking import of {0}'.format(module),
|
|
work_dir='spack-test')
|