spack/var/spack/repos/builtin/packages/py-tensorflow/package.py
2023-09-23 20:04:10 +02:00

1010 lines
49 KiB
Python

# Copyright 2013-2023 Lawrence Livermore National Security, LLC and other
# Spack Project Developers. See the top-level COPYRIGHT file for details.
#
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
import sys
import tempfile
from spack.package import *
class PyTensorflow(Package, CudaPackage, ROCmPackage, PythonExtension):
"""An Open Source Machine Learning Framework for Everyone.
TensorFlow is an end-to-end open source platform for machine learning. It has a
comprehensive, flexible ecosystem of tools, libraries, and community resources that
lets researchers push the state-of-the-art in ML and developers easily build and
deploy ML-powered applications.
TensorFlow was originally developed by researchers and engineers working on the
Google Brain team within Google's Machine Intelligence Research organization to
conduct machine learning and deep neural networks research. The system is general
enough to be applicable in a wide variety of other domains, as well.
"""
homepage = "https://www.tensorflow.org"
url = "https://github.com/tensorflow/tensorflow/archive/v2.3.1.tar.gz"
maintainers("adamjstewart", "aweits")
import_modules = ["tensorflow"]
version("2.13.0", sha256="e58c939079588623e6fa1d054aec2f90f95018266e0a970fd353a5244f5173dc")
version("2.12.1", sha256="6bc4600cc0b88e9e40f1800096f5bddbbd3b6e5527a030dea631b87f2ae46b5b")
version("2.12.0", sha256="c030cb1905bff1d2446615992aad8d8d85cbe90c4fb625cee458c63bf466bc8e")
version("2.11.1", sha256="624ed1cc170cdcc19e8a15d8cdde989a9a1c6b0534c90b38a6b2f06fb2963e5f")
version("2.11.0", sha256="99c732b92b1b37fc243a559e02f9aef5671771e272758aa4aec7f34dc92dac48")
version("2.10.1", sha256="622a92e22e6f3f4300ea43b3025a0b6122f1cc0e2d9233235e4c628c331a94a3")
version("2.10.0", sha256="b5a1bb04c84b6fe1538377e5a1f649bb5d5f0b2e3625a3c526ff3a8af88633e8")
version("2.9.3", sha256="59d09bd00eef6f07477eea2f50778582edd4b7b2850a396f1fd0c646b357a573")
version("2.9.2", sha256="8cd7ed82b096dc349764c3369331751e870d39c86e73bbb5374e1664a59dcdf7")
version("2.9.1", sha256="6eaf86ead73e23988fe192da1db68f4d3828bcdd0f3a9dc195935e339c95dbdc")
version("2.9.0", sha256="8087cb0c529f04a4bfe480e49925cd64a904ad16d8ec66b98e2aacdfd53c80ff")
version("2.8.4", sha256="c08a222792bdbff9da299c7885561ee27b95d414d1111c426efac4ccdce92cde")
version("2.8.3", sha256="4b7ecbe50b36887e1615bc2a582cb86df1250004d8bb540e18336d539803b5a7")
version("2.8.2", sha256="b3f860c02c22a30e9787e2548ca252ab289a76b7778af6e9fa763d4aafd904c7")
version("2.8.1", sha256="4b487a63d6f0c1ca46a2ac37ba4687eabdc3a260c222616fa414f6df73228cec")
version("2.8.0", sha256="66b953ae7fba61fd78969a2e24e350b26ec116cf2e6a7eb93d02c63939c6f9f7")
version(
"2.7.4-rocm-enhanced",
sha256="45b79c125edfdc008274f1b150d8b5a53b3ff4713fd1ad1ff4738f515aad8191",
url="https://github.com/ROCmSoftwarePlatform/tensorflow-upstream/archive/refs/tags/v2.7.4-rocm-enhanced.tar.gz",
)
version("2.7.4", sha256="75b2e40a9623df32da16d8e97528f5e02e4a958e23b1f2ee9637be8eec5d021b")
version("2.7.3", sha256="b576c2e124cd6d4d04cbfe985430a0d955614e882172b2258217f0ec9b61f39b")
version("2.7.2", sha256="b3c8577f3b7cc82368ff7f9315821d506abd2f716ea6692977d255b7d8bc54c0")
version("2.7.1", sha256="abebe2cf5ca379e18071693ca5f45b88ade941b16258a21cc1f12d77d5387a21")
version("2.7.0", sha256="bb124905c7fdacd81e7c842b287c169bbf377d29c74c9dacc04f96c9793747bb")
version("2.6.5", sha256="305da42845ac584a42494e521c92a88ce92ee47d93022d4c0bb45180b5c19a8c")
version("2.6.4", sha256="6a9e54f46039ef0a6f0a1adf19befa510044d3203d1e124dba8318ec4b1e0210")
version("2.6.3", sha256="7a71dde0987677b9512b202eb6ae119e0e308b1ea15b66dcfce001a44873997b")
version("2.6.2", sha256="e68c1d346fc3d529653530ca346b2c62f5b31bd4fcca7ffc9c65bb39ab2f6ed3")
version("2.6.1", sha256="8e457f617bc2eb43de2a51900e7922b60a8107e2524b2576438f1acccee1d043")
version("2.6.0", sha256="41b32eeaddcbc02b0583660bcf508469550e4cd0f86b22d2abe72dfebeacde0f")
version("2.5.3", sha256="58d69b7163f7624debc243750976d27fa7dddbc6fb7c5215aec94732bcc670e1")
version("2.5.2", sha256="bcccc6ba0b8ac1d10d3302f766eed71911acecc0bc43d0bd27d97a1e7ce275a8")
version("2.5.1", sha256="8d2728e155a3aa6befd9cb3d0980fabd25e2142d124f8f6b6c78cdf17ff79da5")
version("2.5.0", sha256="233875ea27fc357f6b714b2a0de5f6ff124b50c1ee9b3b41f9e726e9e677b86c")
version("2.4.4", sha256="f1abc3ed92c3ce955db2a7db5ec422a3a98f015331183194f97b99fe77a09bb4")
version("2.4.3", sha256="cafd520c753f8755a9eb1262932f685dc722d8658f08373f8ec88d8acd58d7d4")
version("2.4.2", sha256="edc88da97277906513d53eeee57997a2036fa32ac1f1937730301764fa06cdc0")
version("2.4.1", sha256="f681331f8fc0800883761c7709d13cda11942d4ad5ff9f44ad855e9dc78387e0")
version("2.4.0", sha256="26c833b7e1873936379e810a39d14700281125257ddda8cd822c89111db6f6ae")
version("2.3.4", sha256="195947838b0918c15d79bc6ed85ff714b24d6d564b4d07ba3de0b745a2f9b656")
version("2.3.3", sha256="b91e5bcd373b942c4a62c6bcb7ff6f968b1448152b82f54a95dfb0d8fb9c6093")
version("2.3.2", sha256="21a703d2e68cd0677f6f9ce329198c24fd8203125599d791af9f1de61aadf31f")
version("2.3.2", sha256="21a703d2e68cd0677f6f9ce329198c24fd8203125599d791af9f1de61aadf31f")
version("2.3.1", sha256="ee534dd31a811f7a759453567257d1e643f216d8d55a25c32d2fbfff8153a1ac")
version("2.3.0", sha256="2595a5c401521f20a2734c4e5d54120996f8391f00bb62a57267d930bce95350")
version("2.2.3", sha256="5e6c779ca8392864d436d88893461dcce783c3a8d46dcb2b2f2ee8ece3cc4538")
version("2.2.2", sha256="fb4b5d26c5b983350f7ce8297b71176a86a69e91faf66e6ebb1e58538ad3bb51")
version("2.2.1", sha256="e6a28e64236d729e598dbeaa02152219e67d0ac94d6ed22438606026a02e0f88")
version("2.2.0", sha256="69cd836f87b8c53506c4f706f655d423270f5a563b76dc1cfa60fbc3184185a3")
version("2.1.4", sha256="f5bd53802b616cc5b4fe5e57a5d3b0f090500a87790020d5fccb0773be7c4b47")
version("2.1.3", sha256="cfa66cce372f486e95a42beb1aacfefdaf0092c5efaaaa92459b381fde931fb8")
version("2.1.2", sha256="3f941cda0ed12dfef5472e46f1d0238ea85da7583d73f1132d2ef050fda6e8ad")
version("2.1.1", sha256="a200bc16e4b630db3ac7225bcb6f239a76841967b0aec1d7d7bbe44dc5661318")
version("2.1.0", sha256="638e541a4981f52c69da4a311815f1e7989bf1d67a41d204511966e1daed14f7")
version("2.0.4", sha256="6ca3ce1255da8d655080a89db10da03f72c361d7faecc9a35e6af26ff12c06e6")
version("2.0.3", sha256="6314299a723441bd9892e5c2af182c2be7d2256e20e71026e1cb1264cb497f33")
version("2.0.2", sha256="a548742bbafd302eec51e2794d7687674a64f6b10ce1414073858cb83c0cefc2")
version("2.0.1", sha256="29197d30923b9670992ee4b9c6161f50c7452e9a4158c720746e846080ac245a")
version("2.0.0", sha256="49b5f0495cd681cbcb5296a4476853d4aea19a43bdd9f179c928a977308a0617")
version("1.15.5", sha256="4c4d23e311093ded2d2e287b18d7c45b07b5984ab88a1d2f91f8f13c886123db")
version("1.15.4", sha256="e18c55e771ad136f9bf3a70ea8f0e2d36662b2ba7c890f9eaf7950554557c7fa")
version("1.15.3", sha256="9ab1d92e58eb813922b040acc7622b32d73c2d8d971fe6491a06f9df4c778151")
version("1.15.2", sha256="d95d75d26a298211b5e802842e87fda5b8b14f6ad83719377b391e5fb71b8746")
version("1.15.1", sha256="19b6e72bc8675937f618cede364d7228a71c2eeaffc42801bcefd98dda7ca056")
version("1.15.0", sha256="a5d49c00a175a61da7431a9b289747d62339be9cf37600330ad63b611f7f5dc9")
version("1.14.0", sha256="aa2a6a1daafa3af66807cfe0bc77bfe1144a9a53df9a96bab52e3e575b3047ed")
version("1.13.2", sha256="abe3bf0c47845a628b7df4c57646f41a10ee70f914f1b018a5c761be75e1f1a9")
version("1.13.1", sha256="7cd19978e6bc7edc2c847bce19f95515a742b34ea5e28e4389dade35348f58ed")
version("1.12.3", sha256="b9e5488e84f4a133ed20b18605f0cd6301f11d356bd959712db4e7b9301d0462")
version("1.12.2", sha256="90ffc7cf1df5e4b8385c9108db18d5d5034ec423547c0e167d44f5746a20d06b")
version("1.12.1", sha256="7b559a3ae56322b7a7e4307f45f9fce96022c533a98b32c18bfdff8c5838271d")
version("1.12.0", sha256="3c87b81e37d4ed7f3da6200474fa5e656ffd20d8811068572f43610cae97ca92")
version("1.11.0", sha256="f49ce3f1d04cee854bc9f74fa9696991140b34a2e2447f35f01391b72c8bfa9f")
version("1.10.1", sha256="83092d709800e2d93d4d4b1bcacaeb74f2f328962ed764cb35bbee20402879c6")
version("1.10.0", sha256="ee9cb98d9e0d8106f2f4ed52a38fe89399324af303e1401567e5b64a9f86744b")
version("1.9.0", sha256="ffc3151b06823d57b4a408261ba8efe53601563dfe93af0866751d4f6ca5068c")
version("1.8.0", sha256="47646952590fd213b747247e6870d89bb4a368a95ae3561513d6c76e44f92a75")
version("1.7.1", sha256="3147f8c60d1f30da23a831bcf732e74b935dcee7c62e4b8b85f0f093030b52c8")
version("1.7.0", sha256="c676a96fc8700722816b2b98c85578b2f99fac7a7b2484c9c7f0641484f8d50d")
version("1.6.0", sha256="03cf1423446abbead6bd8c3cf6e6affa7d99746cd119691b012aac9a1795f4fb")
version("1.5.1", sha256="cab2157783905e12a7a3baae3264edfb739dd92d5658019a131fff4b14190240")
version("1.5.0", sha256="0642781c3a3a8c2c4834b91b86aec385f0b2ada7d721571458079478cc5b29c8")
version("1.4.1", sha256="1f75e463318419a1b3ae076d5a92697c1d3a85e8377c946a5510b651ff5c0d60")
version("1.4.0", sha256="8a0ad8d61f8f6c0282c548661994a5ab83ac531bac496c3041dedc1ab021107b")
version("1.3.1", sha256="ded509c209f8a1d390df8a2f44be5b5c29963172b0e0f095304efb59765d0523")
version("1.3.0", sha256="e1af1bb767b57c3416de0d43a5f74d174c42b85231dffd36f3630173534d4307")
version("1.2.1", sha256="f2baf09b1a9a0500907b4d5cb5473070b3ecede06ed6e8d1096873c91922fb9e")
version("1.2.0", sha256="03dbf7548d1fc1c11ed58da5fa68616f795c819f868f43478cbcaa26abed374f")
version("1.1.0", sha256="aad4470f52fa59f54de7b9a2da727429e6755d91d756f245f952698c42a60027")
version("1.0.1", sha256="deea3c65e0703da96d9c3f1162e464c51d37659dd129396af134e9e8f1ea8c05")
version("1.0.0", sha256="db8b3b8f4134b7c9c1b4165492ad5d5bb78889fcd99ffdffc325e97da3e8c677")
variant("mkl", default=False, description="Build with MKL support")
variant("jemalloc", default=False, description="Build with jemalloc as malloc support")
variant("gcp", default=False, description="Build with Google Cloud Platform support")
variant("hdfs", default=False, description="Build with Hadoop File System support")
variant("aws", default=False, description="Build with Amazon AWS Platform support")
variant("kafka", default=False, description="Build with Apache Kafka Platform support")
variant("ignite", default=False, description="Build with Apache Ignite support")
variant("xla", default=sys.platform != "darwin", description="Build with XLA JIT support")
variant("gdr", default=False, description="Build with GDR support")
variant("verbs", default=False, description="Build with libverbs support")
variant("ngraph", default=False, description="Build with Intel nGraph support")
variant("opencl", default=False, description="Build with OpenCL SYCL support")
variant("computecpp", default=False, description="Build with ComputeCPP support")
variant(
"tensorrt", default=False, description="Build with TensorRT support"
) # TODO: enable when TensorRT in Spack
variant("cuda", default=sys.platform != "darwin", description="Build with CUDA support")
variant(
"nccl", default=sys.platform.startswith("linux"), description="Enable NVIDIA NCCL support"
)
variant("mpi", default=False, description="Build with MPI support")
variant("android", default=False, description="Configure for Android builds")
variant("ios", default=False, description="Build with iOS support (macOS only)")
variant("monolithic", default=False, description="Static monolithic build")
variant("numa", default=False, description="Build with NUMA support")
variant(
"dynamic_kernels",
default=sys.platform.startswith("linux"),
description="Build kernels into separate shared objects",
)
extends("python")
# Python support based on wheel availability
depends_on("python@3.8:3.11", when="@2.12:", type=("build", "run"))
depends_on("python@3.7:3.10", when="@2.8:2.11", type=("build", "run"))
depends_on("python@3.7:3.9", when="@2.7", type=("build", "run"))
depends_on("python@3.6:3.9", when="@2.5:2.6", type=("build", "run"))
depends_on("python@3.6:3.8", when="@2.4", type=("build", "run"))
depends_on("python@3.5:3.8", when="@2.2:2.3", type=("build", "run"))
depends_on("python@2.7,3.5:3.7", when="@:2.1", type=("build", "run"))
# See .bazelversion
depends_on("bazel@5.3.0", type="build", when="@2.11:")
depends_on("bazel@5.1.1", type="build", when="@2.10")
# See _TF_MIN_BAZEL_VERSION and _TF_MAX_BAZEL_VERSION in configure.py
depends_on("bazel@4.2.2:5.99.0", type="build", when="@2.9")
depends_on("bazel@4.2.1:4.99.0", type="build", when="@2.8")
depends_on("bazel@3.7.2:4.99.0", type="build", when="@2.7")
depends_on("bazel@3.7.2:3.99.0", type="build", when="@2.5:2.6")
depends_on("bazel@3.1.0:3.99.0", type="build", when="@2.3:2.4")
depends_on("bazel@2.0.0", type="build", when="@2.2")
depends_on("bazel@0.27.1:0.29.1", type="build", when="@2.1")
depends_on("bazel@0.24.1:0.26.1", type="build", when="@1.15:2.0")
# See call to check_bazel_version in configure.py
depends_on("bazel@0.24.1:0.25.2", type="build", when="@1.14")
depends_on("bazel@0.19.0:0.21.0", type="build", when="@1.13")
depends_on("bazel@0.24.1:0.25.0", type="build", when="@1.12.1")
depends_on("bazel@0.15.0", type="build", when="@1.10:1.12.0,1.12.2:1.12.3")
depends_on("bazel@0.10.0", type="build", when="@1.7:1.9")
# See call to check_version in tensorflow/workspace.bzl
depends_on("bazel@0.5.4", type="build", when="@1.4:1.6")
# See MIN_BAZEL_VERSION in configure
depends_on("bazel@0.4.5", type="build", when="@1.2:1.3")
# See call to check_version in WORKSPACE
depends_on("bazel@0.4.2", type="build", when="@1.0:1.1")
depends_on("swig", type="build")
depends_on("py-pip", type="build")
depends_on("py-wheel", type="build")
# Listed under REQUIRED_PACKAGES in tensorflow/tools/pip_package/setup.py
depends_on("py-absl-py@1:", type=("build", "run"), when="@2.9:")
depends_on("py-absl-py@0.4:", type=("build", "run"), when="@2.7:2.8")
depends_on("py-absl-py@0.10:0", type=("build", "run"), when="@2.4:2.6")
depends_on("py-absl-py@0.7:", type=("build", "run"), when="@1.12.1,1.14:2.3")
depends_on("py-absl-py@0.1.6:", type=("build", "run"), when="@1.5:1.12.0,1.12.2:1.13.2")
depends_on("py-astunparse@1.6:", type=("build", "run"), when="@2.7:")
depends_on("py-astunparse@1.6.3:1.6", type=("build", "run"), when="@2.4:2.6")
depends_on("py-astunparse@1.6.3", type=("build", "run"), when="@2.2:2.3")
depends_on("py-flatbuffers@23.1.21:", type=("build", "run"), when="@2.13:")
depends_on("py-flatbuffers@2:", type=("build", "run"), when="@2.10:2.12")
depends_on("py-flatbuffers@1.12:1", type=("build", "run"), when="@2.9")
depends_on("py-flatbuffers@1.12:", type=("build", "run"), when="@2.8")
depends_on("py-flatbuffers@1.12:2", type=("build", "run"), when="@2.7")
depends_on("py-flatbuffers@1.12", type=("build", "run"), when="@2.4:2.6")
depends_on("py-gast@0.2.1:0.4.0", type=("build", "run"), when="@2.9:")
depends_on("py-gast@0.2.1:", type=("build", "run"), when="@2.8")
depends_on("py-gast@0.2.1:0.4", type=("build", "run"), when="@2.7")
depends_on("py-gast@0.4.0", type=("build", "run"), when="@2.5:2.6")
depends_on("py-gast@0.3.3", type=("build", "run"), when="@2.2:2.4")
depends_on("py-gast@0.2.2", type=("build", "run"), when="@1.15:2.1")
depends_on("py-gast@0.2:", type=("build", "run"), when="@1.6:1.14")
depends_on("py-google-pasta@0.1.1:", type=("build", "run"), when="@2.7:")
depends_on("py-google-pasta@0.2:0", type=("build", "run"), when="@2.4:2.6")
depends_on("py-google-pasta@0.1.8:", type=("build", "run"), when="@2.2:2.3")
depends_on("py-google-pasta@0.1.6:", type=("build", "run"), when="@1.14:2.1")
depends_on("py-google-pasta@0.1.2:", type=("build", "run"), when="@1.12.1")
depends_on("py-h5py@2.9:", type=("build", "run"), when="@2.7:")
depends_on("py-h5py@3.1", type=("build", "run"), when="@2.5:2.6")
depends_on("py-h5py@2.10", type=("build", "run"), when="@2.2:2.4")
depends_on("py-h5py@:2.10.0", type=("build", "run"), when="@1.15.5,2.0.4,2.1.3:2.1")
# propagate the mpi variant setting for h5py/hdf5 to avoid unexpected crashes
depends_on("py-h5py+mpi", type=("build", "run"), when="@1.15.5,2.0.4,2.1.3:+mpi")
depends_on("py-h5py~mpi", type=("build", "run"), when="@1.15.5,2.0.4,2.1.3:~mpi")
depends_on("hdf5+mpi", type="build", when="@1.15.5,2.0.4,2.1.3:+mpi")
depends_on("hdf5~mpi", type="build", when="@1.15.5,2.0.4,2.1.3:~mpi")
depends_on("py-libclang@13:", type=("build", "run"), when="@2.9:")
depends_on("py-libclang@9.0.1:", type=("build", "run"), when="@2.7:2.8")
depends_on("py-numpy@1.22:1.24.3", type=("build", "run"), when="@2.13:")
depends_on("py-numpy@1.22:1.23", type=("build", "run"), when="@2.12")
depends_on("py-numpy@1.20:", type=("build", "run"), when="@2.8:2.11")
depends_on("py-numpy@1.14.5:", type=("build", "run"), when="@2.7")
depends_on("py-numpy@1.19.2:1.19", type=("build", "run"), when="@2.4:2.6")
# https://github.com/tensorflow/tensorflow/issues/40688
depends_on("py-numpy@1.16.0:1.18", type=("build", "run"), when="@1.13.2,1.15:2.3")
depends_on("py-numpy@1.14.5:1.18", type=("build", "run"), when="@1.12.1,1.14")
depends_on(
"py-numpy@1.13.3:1.18", type=("build", "run"), when="@1.6:1.9,1.11:1.12.0,1.12.2:1.13.1"
)
depends_on("py-numpy@1.13.3:1.14.5", type=("build", "run"), when="@1.10")
depends_on("py-numpy@1.12.1:1.14.5", type=("build", "run"), when="@1.4:1.5")
depends_on("py-numpy@1.11.0:1.14.5", type=("build", "run"), when="@:1.3")
depends_on("py-opt-einsum@2.3.2:", type=("build", "run"), when="@1.15:2.3,2.7:")
depends_on("py-opt-einsum@3.3", type=("build", "run"), when="@2.4:2.6")
depends_on("py-packaging", type=("build", "run"), when="@2.9:")
depends_on("py-protobuf@3.20.3:4.20,4.21.6:4", type=("build", "run"), when="@2.12:")
depends_on("py-protobuf@3.9.2:", type=("build", "run"), when="@2.3:2.11")
depends_on("py-protobuf@3.8.0:", type=("build", "run"), when="@2.1:2.2")
depends_on("py-protobuf@3.6.1:", type=("build", "run"), when="@1.12:2.0")
depends_on("py-protobuf@3.6.0:", type=("build", "run"), when="@1.10:1.11")
depends_on("py-protobuf@3.4.0:", type=("build", "run"), when="@1.5:1.9")
depends_on("py-protobuf@3.3.0:", type=("build", "run"), when="@1.3:1.4")
depends_on("py-protobuf@3.2.0:", type=("build", "run"), when="@1.1:1.2")
depends_on("py-protobuf@3.1.0:", type=("build", "run"), when="@:1.0")
# https://github.com/protocolbuffers/protobuf/issues/10051
# https://github.com/tensorflow/tensorflow/issues/56266
depends_on("py-protobuf@:3.19", type=("build", "run"), when="@:2.11")
# Must be matching versions of py-protobuf and protobuf
conflicts("^py-protobuf~cpp")
depends_on("py-setuptools", type=("build", "run"))
depends_on("py-six@1.12:", type=("build", "run"), when="@2.1:2.3,2.7:")
depends_on("py-six@1.15", type=("build", "run"), when="@2.4:2.6")
depends_on("py-six@1.10:", type=("build", "run"), when="@:2.0")
depends_on("py-termcolor@1.1:", type=("build", "run"), when="@1.6:2.3,2.7:")
depends_on("py-termcolor@1.1", type=("build", "run"), when="@2.4:2.6")
depends_on("py-typing-extensions@3.6.6:4.5", type=("build", "run"), when="@2.13:")
depends_on("py-typing-extensions@3.6.6:", type=("build", "run"), when="@2.7:2.12")
depends_on("py-typing-extensions@3.7.4:3.7", type=("build", "run"), when="@2.4:2.6")
depends_on("py-wrapt@1.11:", type=("build", "run"), when="@2.13:")
depends_on("py-wrapt@1.11:1.14", type=("build", "run"), when="@2.12")
depends_on("py-wrapt@1.11:", type=("build", "run"), when="@2.7:2.11")
depends_on("py-wrapt@1.12.1:1.12", type=("build", "run"), when="@2.4:2.6")
depends_on("py-wrapt@1.11.1:", type=("build", "run"), when="@1.12.1,1.14:2.3")
# TODO: add packages for these dependencies
# depends_on('py-tensorflow-io-gcs-filesystem@0.23.1:', type=('build', 'run'), when='@2.8:')
# depends_on('py-tensorflow-io-gcs-filesystem@0.21:', type=('build', 'run'), when='@2.7')
with when("+rocm"):
depends_on("hip")
depends_on("rocrand")
depends_on("rocblas")
depends_on("rocfft")
depends_on("hipfft")
depends_on("rccl", when="+nccl")
depends_on("hipsparse")
depends_on("hipcub")
depends_on("rocsolver")
depends_on("rocprim")
depends_on("miopen-hip")
depends_on("llvm-amdgpu")
depends_on("hsa-rocr-dev")
depends_on("rocminfo")
if sys.byteorder == "little":
# Only builds correctly on little-endian machines
depends_on("py-grpcio@1.24.3:1", type=("build", "run"), when="@2.7:")
depends_on("py-grpcio@1.37.0:1", type=("build", "run"), when="@2.6")
depends_on("py-grpcio@1.34", type=("build", "run"), when="@2.5")
depends_on("py-grpcio@1.32", type=("build", "run"), when="@2.4")
depends_on("py-grpcio@1.8.6:", type=("build", "run"), when="@1.6:2.3")
for minor_ver in range(5, 14):
depends_on(
"py-tensorboard@2.{}".format(minor_ver),
type=("build", "run"),
when="@2.{}".format(minor_ver),
)
# TODO: is this still true? We now install tensorboard from wheel for all versions
# depends_on('py-tensorboard', when='@:2.4') # circular dep
# depends_on('py-tensorflow-estimator') # circular dep
# depends_on('py-keras') # circular dep
# Undocumented dependencies
depends_on("py-requests", type=("build", "run"))
# https://github.com/tensorflow/tensorflow/issues/60179#issuecomment-1491238631
depends_on("coreutils", when="@2.13: platform=darwin", type="build")
# No longer a dependency in latest versions
depends_on("py-astor@0.6:", type=("build", "run"), when="@1.6:2.1")
depends_on("py-backports-weakref@1.0rc1", type=("build", "run"), when="@1.2")
depends_on("py-jax@0.3.15:", type=("build", "run"), when="@2.12")
depends_on("py-keras-applications@1.0.8:", type=("build", "run"), when="@1.15:2.1")
depends_on("py-keras-applications@1.0.6:", type=("build", "run"), when="@1.12:1.14")
depends_on("py-keras-applications@1.0.5:", type=("build", "run"), when="@1.11")
depends_on("py-keras-preprocessing@1.1.1:", type=("build", "run"), when="@2.7:2.10")
depends_on("py-keras-preprocessing@1.1.2:1.1", type=("build", "run"), when="@2.4:2.6")
depends_on("py-keras-preprocessing@1.1.1:1.1", type=("build", "run"), when="@2.3")
depends_on("py-keras-preprocessing@1.1:", type=("build", "run"), when="@2.1.0,2.2")
depends_on("py-keras-preprocessing@1.1.0", type=("build", "run"), when="@2.1.1:2.1")
depends_on("py-keras-preprocessing@1.0.5:", type=("build", "run"), when="@1.12:2.0")
depends_on("py-keras-preprocessing@1.0.3:", type=("build", "run"), when="@1.11")
depends_on("py-scipy@1.4.1", type=("build", "run"), when="@2.1.0:2.1.1,2.2.0,2.3.0")
depends_on("py-wheel@0.32:0", type=("build", "run"), when="@2.7")
depends_on("py-wheel@0.35:0", type=("build", "run"), when="@2.4:2.6")
depends_on("py-wheel@0.26:", type=("build", "run"), when="@:2.3")
# TODO: add packages for some of these dependencies
depends_on("mkl", when="+mkl")
depends_on("curl", when="+gcp")
# depends_on('computecpp', when='+opencl+computecpp')
# depends_on('trisycl', when='+opencl~computepp')
depends_on("cuda@:10.2", when="+cuda @:2.3")
depends_on("cuda@:11.4", when="+cuda @2.4:2.7")
# avoid problem fixed by commit a76f797b9cd4b9b15bec4c503b16236a804f676f
depends_on("cuda@:11.7.0", when="+cuda @:2.9")
depends_on("cudnn", when="+cuda")
depends_on("cudnn@:7", when="@:2.2 +cuda")
# depends_on('tensorrt', when='+tensorrt')
depends_on("nccl", when="+nccl+cuda")
depends_on("mpi", when="+mpi")
# depends_on('android-ndk@10:18', when='+android')
# depends_on('android-sdk', when='+android')
# Check configure and configure.py to see when these variants are supported
conflicts("+mkl", when="@:1.0")
conflicts("+mkl", when="platform=darwin", msg="Darwin is not yet supported")
conflicts(
"+jemalloc",
when="platform=darwin",
msg="Currently jemalloc is only support on Linux platform",
)
conflicts(
"+jemalloc",
when="platform=cray",
msg="Currently jemalloc is only support on Linux platform",
)
conflicts("+aws", when="@:1.3")
conflicts("+kafka", when="@:1.5,2.1:")
conflicts("+ignite", when="@:1.11,2.1:")
conflicts("+gdr", when="@:1.3")
conflicts("+verbs", when="@:1.1")
conflicts("+ngraph", when="@:1.10")
conflicts("+computecpp", when="~opencl")
conflicts("+cuda", when="platform=darwin", msg="There is no GPU support for macOS")
conflicts(
"cuda_arch=none",
when="+cuda",
msg="Must specify CUDA compute capabilities of your GPU, see https://developer.nvidia.com/cuda-gpus",
)
conflicts(
"cuda_arch=20",
when="@1.12.1,1.14:",
msg="TensorFlow only supports compute capabilities >= 3.5",
)
conflicts(
"cuda_arch=30",
when="@1.12.1,1.14:",
msg="TensorFlow only supports compute capabilities >= 3.5",
)
conflicts(
"cuda_arch=32",
when="@1.12.1,1.14:",
msg="TensorFlow only supports compute capabilities >= 3.5",
)
conflicts(
"cuda_arch=20",
when="@1.4:1.12.0,1.12.2:1.12.3",
msg="Only compute capabilities 3.0 or higher are supported",
)
conflicts("+tensorrt", when="@:1.5")
conflicts("+tensorrt", when="~cuda")
conflicts(
"+tensorrt",
when="platform=darwin",
msg="Currently TensorRT is only supported on Linux platform",
)
conflicts(
"+tensorrt",
when="platform=cray",
msg="Currently TensorRT is only supported on Linux platform",
)
conflicts("+nccl", when="@:1.7")
conflicts("+nccl", when="~cuda~rocm")
conflicts(
"+nccl", when="platform=darwin", msg="Currently NCCL is only supported on Linux platform"
)
conflicts(
"+nccl", when="platform=cray", msg="Currently NCCL is only supported on Linux platform"
)
conflicts("+mpi", when="@:1.2")
conflicts("+android", when="@:1.4")
conflicts("+ios", when="@:1.12.0,1.12.2:1.13")
conflicts("+ios", when="platform=linux", msg="iOS support only available on macOS")
conflicts("+ios", when="platform=cray", msg="iOS support only available on macOS")
conflicts("+monolithic", when="@:1.3")
conflicts("+numa", when="@:1.12.0,1.12.2:1.13")
conflicts("+dynamic_kernels", when="@:1.12.0,1.12.2:1.12.3")
# https://github.com/tensorflow/tensorflow/pull/45404
conflicts("platform=darwin target=aarch64:", when="@:2.4")
# https://github.com/tensorflow/tensorflow/pull/39225
conflicts("target=aarch64:", when="@:2.2")
conflicts("~rocm", when="@2.7.4-rocm-enhanced")
conflicts("+rocm", when="@:2.7.4-a,2.7.4.0:")
# https://www.tensorflow.org/install/source#tested_build_configurations
conflicts("%gcc@:9.3.0", when="@2.9:")
conflicts("%gcc@:7.3.0", when="@1.15:")
conflicts("%gcc@:4.7")
# zlib is vendored and downloaded directly from zlib.org (or mirrors), but
# old downloads are removed from that site immediately after a new release.
# If the tf mirrors don't work, make sure the fallback is to something existing.
patch(
"https://github.com/tensorflow/tensorflow/commit/76b9fa22857148a562f3d9b5af6843402a93c15b.patch?full_index=1",
sha256="f9e26c544da729cfd376dbd3b096030e3777d3592459add1f3c78b1b9828d493",
when="@2.9:2.10.0",
)
# Version 2.10 produces an error related to cuBLAS:
# E tensorflow/stream_executor/cuda/cuda_blas.cc:2981] Unable to register
# cuBLAS factory: Attempting to register factory for plugin cuBLAS when one
# has already been registered
# See https://github.com/tensorflow/tensorflow/issues/57663
# This is fixed for 2.11 but 2.10 needs the following patch.
patch(
"https://github.com/tensorflow/tensorflow/pull/56691.patch?full_index=1",
sha256="d635ea6d6c1571505871d0caba3e2cd939ea0f4aff972095d552913a8109def3",
when="@2.10",
)
# Avoid build error: "no such package '@io_bazel_rules_docker..."
patch("io_bazel_rules_docker2.patch", when="@1.15:2.0")
# Avoide build error: "name 'new_http_archive' is not defined"
patch("http_archive.patch", when="@1.12.3")
# Backport of 837c8b6b upstream
# "Remove contrib cloud bigtable and storage ops/kernels."
# Allows 2.0.* releases to build with '--config=nogcp'
patch("0001-Remove-contrib-cloud-bigtable-and-storage-ops-kernel.patch", when="@2.0.0:2.0")
# for fcc
patch("1-1_fcc_tf_patch.patch", when="@2.1.0:2.1%fj")
# do not import contrib.cloud if not available
patch(
"https://github.com/tensorflow/tensorflow/commit/ed62ac8203999513dfae03498e871ea35eb60cc4.patch?full_index=1",
sha256="ff02e249532a5661b123108734a39534992d81da90f0c8187bf4e151a865effc",
when="@1.14.0",
)
# import_contrib_cloud patch for older versions
patch("contrib_cloud_1.10.patch", when="@1.10:1.13")
patch("contrib_cloud_1.9.patch", when="@1.9")
patch("contrib_cloud_1.4.patch", when="@1.4:1.8")
patch("contrib_cloud_1.1.patch", when="@1.1:1.3")
# needed for protobuf 3.16+
patch("example_parsing.patch", when="@:2.7 ^protobuf@3.16:")
# allow linker to be found in PATH
# https://github.com/tensorflow/tensorflow/issues/39263
patch("null_linker_bin_path.patch", when="@2.5:")
# Reset import order to that of 2.4. Part of
# https://bugs.gentoo.org/800824#c3 From the patch:
# When tensorflow and python protobuf use the same instance of libprotobuf,
# pywrap_tensorflow must be imported before anything else that would import
# protobuf definitions.
patch("0008-Fix-protobuf-errors-when-using-system-protobuf.patch", when="@2.5:2.6")
phases = ["configure", "build", "install"]
# https://www.tensorflow.org/install/source
def setup_build_environment(self, env):
spec = self.spec
# Please specify the location of python
env.set("PYTHON_BIN_PATH", spec["python"].command.path)
# Please input the desired Python library path to use
env.set("PYTHON_LIB_PATH", python_platlib)
# Ensure swig is in PATH or set SWIG_PATH
env.set("SWIG_PATH", spec["swig"].prefix.bin.swig)
# Do you wish to build TensorFlow with MKL support?
if "+mkl" in spec:
env.set("TF_NEED_MKL", "1")
# Do you wish to download MKL LIB from the web?
env.set("TF_DOWNLOAD_MKL", "0")
# Please specify the location where MKL is installed
env.set("MKL_INSTALL_PATH", spec["mkl"].prefix)
else:
env.set("TF_NEED_MKL", "0")
# Do you wish to build TensorFlow with jemalloc as malloc support?
if "+jemalloc" in spec:
env.set("TF_NEED_JEMALLOC", "1")
else:
env.set("TF_NEED_JEMALLOC", "0")
# Do you wish to build TensorFlow with Google Cloud Platform support?
if "+gcp" in spec:
env.set("TF_NEED_GCP", "1")
else:
env.set("TF_NEED_GCP", "0")
# Do you wish to build TensorFlow with Hadoop File System support?
if "+hdfs" in spec:
env.set("TF_NEED_HDFS", "1")
else:
env.set("TF_NEED_HDFS", "0")
# Do you wish to build TensorFlow with Amazon AWS Platform support?
if "+aws" in spec:
env.set("TF_NEED_AWS", "1")
env.set("TF_NEED_S3", "1")
else:
env.set("TF_NEED_AWS", "0")
env.set("TF_NEED_S3", "0")
# Do you wish to build TensorFlow with Apache Kafka Platform support?
if "+kafka" in spec:
env.set("TF_NEED_KAFKA", "1")
else:
env.set("TF_NEED_KAFKA", "0")
# Do you wish to build TensorFlow with Apache Ignite support?
if "+ignite" in spec:
env.set("TF_NEED_IGNITE", "1")
else:
env.set("TF_NEED_IGNITE", "0")
# Do you wish to build TensorFlow with XLA JIT support?
if "+xla" in spec:
env.set("TF_ENABLE_XLA", "1")
else:
env.set("TF_ENABLE_XLA", "0")
# Do you wish to build TensorFlow with GDR support?
if "+gdr" in spec:
env.set("TF_NEED_GDR", "1")
else:
env.set("TF_NEED_GDR", "0")
# Do you wish to build TensorFlow with VERBS support?
if "+verbs" in spec:
env.set("TF_NEED_VERBS", "1")
else:
env.set("TF_NEED_VERBS", "0")
# Do you wish to build TensorFlow with nGraph support?
if "+ngraph" in spec:
env.set("TF_NEED_NGRAPH", "1")
else:
env.set("TF_NEED_NGRAPH", "0")
# Do you wish to build TensorFlow with OpenCL SYCL support?
if "+opencl" in spec:
env.set("TF_NEED_OPENCL_SYCL", "1")
env.set("TF_NEED_OPENCL", "1")
# Please specify which C++ compiler should be used as the host
# C++ compiler
env.set("HOST_CXX_COMPILER", spack_cxx)
# Please specify which C compiler should be used as the host
# C compiler
env.set("HOST_C_COMPILER", spack_cc)
# Do you wish to build TensorFlow with ComputeCPP support?
if "+computecpp" in spec:
env.set("TF_NEED_COMPUTECPP", "1")
# Please specify the location where ComputeCpp is installed
env.set("COMPUTECPP_TOOLKIT_PATH", spec["computecpp"].prefix)
else:
env.set("TF_NEED_COMPUTECPP", "0")
# Please specify the location of the triSYCL include directory
env.set("TRISYCL_INCLUDE_DIR", spec["trisycl"].prefix.include)
else:
env.set("TF_NEED_OPENCL_SYCL", "0")
env.set("TF_NEED_OPENCL", "0")
# Do you wish to build TensorFlow with ROCm support?
if "+rocm" in spec:
env.set("TF_NEED_ROCM", "1")
else:
env.set("TF_NEED_ROCM", "0")
# Do you wish to build TensorFlow with CUDA support?
if "+cuda" in spec:
env.set("TF_NEED_CUDA", "1")
# Do you want to use clang as CUDA compiler?
env.set("TF_CUDA_CLANG", "0")
# Please specify which gcc nvcc should use as the host compiler
env.set("GCC_HOST_COMPILER_PATH", spack_cc)
cuda_paths = [spec["cuda"].prefix, spec["cudnn"].prefix]
# Do you wish to build TensorFlow with TensorRT support?
if "+tensorrt" in spec:
env.set("TF_NEED_TENSORRT", "1")
cuda_paths.append(spec["tensorrt"].prefix)
# Please specify the TensorRT version you want to use
env.set("TF_TENSORRT_VERSION", spec["tensorrt"].version.up_to(1))
# Please specify the location where TensorRT is installed
env.set("TENSORRT_INSTALL_PATH", spec["tensorrt"].prefix)
else:
env.set("TF_NEED_TENSORRT", "0")
env.unset("TF_TENSORRT_VERSION")
# Please specify the CUDA SDK version you want to use
env.set("TF_CUDA_VERSION", spec["cuda"].version.up_to(2))
# Please specify the cuDNN version you want to use
env.set("TF_CUDNN_VERSION", spec["cudnn"].version.up_to(1))
if "+nccl" in spec:
cuda_paths.append(spec["nccl"].prefix)
# Please specify the locally installed NCCL version to use
env.set("TF_NCCL_VERSION", spec["nccl"].version.up_to(1))
# Please specify the location where NCCL is installed
env.set("NCCL_INSTALL_PATH", spec["nccl"].prefix)
env.set("NCCL_HDR_PATH", spec["nccl"].prefix.include)
else:
env.unset("TF_NCCL_VERSION")
# Please specify the comma-separated list of base paths to
# look for CUDA libraries and headers
env.set("TF_CUDA_PATHS", ",".join(cuda_paths))
# Please specify the location where CUDA toolkit is installed
env.set("CUDA_TOOLKIT_PATH", spec["cuda"].prefix)
# Please specify the location where CUDNN library is installed
env.set("CUDNN_INSTALL_PATH", spec["cudnn"].prefix)
# Please specify a list of comma-separated CUDA compute
# capabilities you want to build with. You can find the compute
# capability of your device at:
# https://developer.nvidia.com/cuda-gpus.
# Please note that each additional compute capability significantly
# increases your build time and binary size, and that TensorFlow
# only supports compute capabilities >= 3.5
capabilities = ",".join(
"{0:.1f}".format(float(i) / 10.0) for i in spec.variants["cuda_arch"].value
)
env.set("TF_CUDA_COMPUTE_CAPABILITIES", capabilities)
else:
env.set("TF_NEED_CUDA", "0")
# Do you wish to download a fresh release of clang? (Experimental)
env.set("TF_DOWNLOAD_CLANG", "0")
# Do you wish to build TensorFlow with MPI support?
if "+mpi" in spec:
env.set("TF_NEED_MPI", "1")
# Please specify the MPI toolkit folder
env.set("MPI_HOME", spec["mpi"].prefix)
else:
env.set("TF_NEED_MPI", "0")
env.unset("MPI_HOME")
# Please specify optimization flags to use during compilation when
# bazel option '--config=opt' is specified
env.set(
"CC_OPT_FLAGS",
spec.target.optimization_flags(spec.compiler.name, spec.compiler.version),
)
# Would you like to interactively configure ./WORKSPACE for
# Android builds?
if "+android" in spec:
env.set("TF_SET_ANDROID_WORKSPACE", "1")
# Please specify the home path of the Android NDK to use
env.set("ANDROID_NDK_HOME", spec["android-ndk"].prefix)
env.set("ANDROID_NDK_API_LEVEL", spec["android-ndk"].version)
# Please specify the home path of the Android SDK to use
env.set("ANDROID_SDK_HOME", spec["android-sdk"].prefix)
env.set("ANDROID_SDK_API_LEVEL", spec["android-sdk"].version)
# Please specify the Android SDK API level to use
env.set("ANDROID_API_LEVEL", spec["android-sdk"].version)
# Please specify an Android build tools version to use
env.set("ANDROID_BUILD_TOOLS_VERSION", spec["android-sdk"].version)
else:
env.set("TF_SET_ANDROID_WORKSPACE", "0")
# Do you wish to build TensorFlow with iOS support?
if "+ios" in spec:
env.set("TF_CONFIGURE_IOS", "1")
else:
env.set("TF_CONFIGURE_IOS", "0")
# set tmpdir to a non-NFS filesystem
# (because bazel uses ~/.cache/bazel)
# TODO: This should be checked for non-nfsy filesystem, but the current
# best idea for it is to check
# subprocess.call([
# 'stat', '--file-system', '--format=%T', tmp_path
# ])
# to not be nfs. This is only valid for Linux and we'd like to
# stay at least also OSX compatible
tmp_path = tempfile.mkdtemp(prefix="spack")
env.set("TEST_TMPDIR", tmp_path)
def configure(self, spec, prefix):
# NOTE: configure script is interactive. If you set the appropriate
# environment variables, this interactivity is skipped. If you don't,
# Spack hangs during the configure phase. Use `spack build-env` to
# determine which environment variables must be set for a particular
# version.
configure()
@run_after("configure")
def post_configure_fixes(self):
spec = self.spec
# make sure xla is actually turned off
if spec.satisfies("~xla"):
filter_file(
r"--define with_xla_support=true",
r"--define with_xla_support=false",
".tf_configure.bazelrc",
)
if spec.satisfies("@1.5.0: ~android"):
# env variable is somehow ignored -> brute force
# TODO: find a better solution
filter_file(r"if workspace_has_any_android_rule\(\)", r"if True", "configure.py")
# version dependent fixes
if spec.satisfies("@1.3.0:1.5.0"):
# checksum for protobuf that bazel downloads (@github) changed
filter_file(
r'sha256 = "6d43b9d223ce09e5d4ce8b0060cb8a7513577a35a64c7e3dad10f0703bf3ad93"',
r'sha256 = "e5fdeee6b28cf6c38d61243adff06628baa434a22b5ebb7432d2a7fbabbdb13d"',
"tensorflow/workspace.bzl",
)
# starting with tensorflow 1.3, tensorboard becomes a dependency
# -> remove from list of required packages
filter_file(
r"'tensorflow-tensorboard",
r"#'tensorflow-tensorboard",
"tensorflow/tools/pip_package/setup.py",
)
if spec.satisfies("@1.5.0: ~gcp"):
# google cloud support seems to be installed on default, leading
# to boringssl error manually set the flag to false to avoid
# installing gcp support
# https://github.com/tensorflow/tensorflow/issues/20677#issuecomment-404634519
filter_file(
r"--define with_gcp_support=true",
r"--define with_gcp_support=false",
".tf_configure.bazelrc",
)
if spec.satisfies("@1.6.0:2.1"):
# tensorboard name changed
# there are no corresponding versions of these in spack
filter_file(
r"(^\s*)'tensorboard (>=|~=)",
r"\1#'tensorboard \2",
"tensorflow/tools/pip_package/setup.py",
)
if spec.satisfies("@1.8.0: ~opencl"):
# 1.8.0 and 1.9.0 aborts with numpy import error during python_api
# generation somehow the wrong PYTHONPATH is used...
# set --distinct_host_configuration=false as a workaround
# https://github.com/tensorflow/tensorflow/issues/22395#issuecomment-431229451
with open(".tf_configure.bazelrc", mode="a") as f:
f.write("build --distinct_host_configuration=false\n")
f.write('build --action_env PYTHONPATH="{0}"\n'.format(env["PYTHONPATH"]))
if spec.satisfies("@1.13.1:"):
# tensorflow_estimator is an API for tensorflow
# tensorflow-estimator imports tensorflow during build, so
# tensorflow has to be set up first
filter_file(
r"(^\s*)'tensorflow_estimator (>=|~=)",
r"\1#'tensorflow_estimator \2",
"tensorflow/tools/pip_package/setup.py",
)
if spec.satisfies("@2.5"):
filter_file(
r"(^\s*)'keras-nightly (>=|~=)",
r"\1#'keras-nightly \2",
"tensorflow/tools/pip_package/setup.py",
)
if spec.satisfies("@2.6:"):
filter_file(
r"(^\s*)'keras (>=|~=)", r"\1#'keras \2", "tensorflow/tools/pip_package/setup.py"
)
if spec.satisfies("@2.6"):
filter_file(
r"(^\s*)'clang (>=|~=)", r"\1#'clang \2", "tensorflow/tools/pip_package/setup.py"
)
# TODO: add support for tensorflow-io-gcs-filesystem
if spec.satisfies("@2.7:"):
filter_file(
r"(^\s*)'tensorflow-io-gcs-filesystem (>=|~=)",
r"\1#'tensorflow-io-gcs-filesystem \2",
"tensorflow/tools/pip_package/setup.py",
)
if spec.satisfies("@2.0.0:"):
# now it depends on the nightly versions...
filter_file(
r"REQUIRED_PACKAGES\[i\] = 'tb-nightly (>=|~=)",
r"pass #REQUIRED_PACKAGES[i] = 'tb-nightly \1",
"tensorflow/tools/pip_package/setup.py",
)
filter_file(
r"REQUIRED_PACKAGES\[i\] = 'tensorflow-estimator-2.0-preview",
r"pass #REQUIRED_PACKAGES[i] = 'tensorflow-estimator-2.0-preview",
"tensorflow/tools/pip_package/setup.py",
)
filter_file(
r"REQUIRED_PACKAGES\[i\] = 'tf-estimator-nightly (>=|~=)",
r"pass #REQUIRED_PACKAGES[i] = 'tf-estimator-nightly \1",
"tensorflow/tools/pip_package/setup.py",
)
filter_file(
r"REQUIRED_PACKAGES\[i\] = 'keras-nightly (>=|~=)",
r"pass #REQUIRED_PACKAGES[i] = 'keras-nightly \1",
"tensorflow/tools/pip_package/setup.py",
)
if spec.satisfies("@1.13.1 +nccl"):
filter_file(
r"^build --action_env NCCL_INSTALL_PATH=.*",
r'build --action_env NCCL_INSTALL_PATH="' + spec["nccl"].libs.directories[0] + '"',
".tf_configure.bazelrc",
)
filter_file(
r"^build --action_env NCCL_HDR_PATH=.*",
r'build --action_env NCCL_HDR_PATH="' + spec["nccl"].prefix.include + '"',
".tf_configure.bazelrc",
)
# see tensorflow issue #31187 on github
if spec.satisfies("@2.0.0:2.0"):
filter_file(
r"\#define RUY_DONOTUSEDIRECTLY_AVX512 1",
"#define RUY_DONOTUSEDIRECTLY_AVX512 0",
"tensorflow/lite/experimental/ruy/platform.h",
)
if spec.satisfies("+cuda"):
libs = spec["cuda"].libs.directories
libs.extend(spec["cudnn"].libs.directories)
if "+nccl" in spec:
libs.extend(spec["nccl"].libs.directories)
if "+tensorrt" in spec:
libs.extend(spec["tensorrt"].libs.directories)
slibs = ":".join(libs)
with open(".tf_configure.bazelrc", mode="a") as f:
f.write('build --action_env LD_LIBRARY_PATH="' + slibs + '"')
filter_file("build:opt --copt=-march=native", "", ".tf_configure.bazelrc")
filter_file("build:opt --host_copt=-march=native", "", ".tf_configure.bazelrc")
def build(self, spec, prefix):
tmp_path = env["TEST_TMPDIR"]
# https://docs.bazel.build/versions/master/command-line-reference.html
args = [
# Don't allow user or system .bazelrc to override build settings
"--nohome_rc",
"--nosystem_rc",
# Bazel does not work properly on NFS, switch to /tmp
"--output_user_root=" + tmp_path,
"build",
# Spack logs don't handle colored output well
"--color=no",
"--jobs={0}".format(make_jobs),
"--config=opt",
# Enable verbose output for failures
"--verbose_failures",
]
if spec.satisfies("^bazel@:3.5"):
# removed in bazel 3.6
args.append("--incompatible_no_support_tools_in_action_inputs=false")
# See .bazelrc for when each config flag is supported
if spec.satisfies("@1.12.1:"):
if "+mkl" in spec:
args.append("--config=mkl")
if "+monolithic" in spec:
args.append("--config=monolithic")
if "+gdr" in spec:
args.append("--config=gdr")
if "+verbs" in spec:
args.append("--config=verbs")
if "+ngraph" in spec:
args.append("--config=ngraph")
if "+dynamic_kernels" in spec:
args.append("--config=dynamic_kernels")
if "+cuda" in spec:
args.append("--config=cuda")
if "+rocm" in spec:
args.append("--config=rocm")
if "~aws" in spec:
args.append("--config=noaws")
if "~gcp" in spec:
args.append("--config=nogcp")
if "~hdfs" in spec:
args.append("--config=nohdfs")
if "~nccl" in spec:
args.append("--config=nonccl")
if spec.satisfies("@1.12.1:2.0"):
if "~ignite" in spec:
args.append("--config=noignite")
if "~kafka" in spec:
args.append("--config=nokafka")
if spec.satisfies("@1.12.1,1.14:"):
if "+numa" in spec:
args.append("--config=numa")
if spec.satisfies("@2:"):
args.append("--config=v2")
args.append("//tensorflow/tools/pip_package:build_pip_package")
bazel(*args)
build_pip_package = Executable("bazel-bin/tensorflow/tools/pip_package/build_pip_package")
buildpath = join_path(self.stage.source_path, "spack-build")
build_pip_package("--src", buildpath)
def install(self, spec, prefix):
tmp_path = env["TEST_TMPDIR"]
buildpath = join_path(self.stage.source_path, "spack-build")
with working_dir(buildpath):
args = std_pip_args + ["--prefix=" + prefix, "."]
pip(*args)
remove_linked_tree(tmp_path)