Re enable aws pcluster buildcache stack (#38931)
* Changes to re-enable aws-pcluster pipelines - Use compilers from pre-installed spack store such that compiler path relocation works when downloading from buildcache. - Install gcc from hash so there is no risk of building gcc from source in pipleine. - `packages.yam` files are now part of the pipelines. - No more eternal `postinstall.sh`. The necessary steps are in `setup=pcluster.sh` and will be version controlled within this repo. - Re-enable pipelines. * Add and * Debugging output & mv skylake -> skylake_avx512 * Explicilty check for packages * Handle case with no intel compiler * compatibility when using setup-pcluster.sh on a pre-installed cluster. * Disable palace as parser cannot read require clause at the moment * ifort cannot build superlu in buildcache `ifort` is unable to handle such long file names as used when cmake compiles test programs inside build cache. * Fix spack commit for intel compiler installation * Need to fetch other commits before using them * fix style * Add TODO * Update packages.yaml to not use 'compiler:', 'target:' or 'provider:' Synchronize with changes in https://github.com/spack/spack-configs/blob/main/AWS/parallelcluster/ * Use Intel compiler from later version (orig commit no longer found) * Use envsubst to deal with quoted newlines This is cleaner than the `eval` command used. * Need to fetch tags for checkout on version number * Intel compiler needs to be from version that has compatible DB * Install intel compiler with commit that has DB ver 7 * Decouple the intel compiler installation from current commit - Use a completely different spack installation such that this current pipeline commit remains untouched. - Make the script suceed even if the compiler installation fails (e.g. because the Database version has been updated) - Make the install targets fall back to gcc in case the compiler did not install correctly. * Use generic target for x86_64_vX There is no way to provision a skylake/icelake/zen runner. They are all in the same pools under x86_64_v3 and x86_64_v4. * Find the intel compiler in the current spack installation * Remove SPACK_TARGET_ARCH * Fix virtual package index & use package.yaml for intel compiler * Use only one stack & pipeline per generic architecture * Fix yaml format * Cleanup typos * Include fix for ifx.cfg to get the right gcc toolchain when linking * [removeme] Adding timeout to debug hang in make (palace) * Revert "[removeme] Adding timeout to debug hang in make (palace)" This reverts commit fee8a01580489a4ea364368459e9353b46d0d7e2. * palace x86_64_v4 gets stuck when compiling try newer oneapi * Update comment * Use the latest container image * Update gcc_hashes to match new container * Use only one tag providing tags per extends call Also removed an unnecessary tag. * Move generic setup script out of individual stack * Cleanup from last commit * Enable checking signature for packages available on the container * Remove commented packages / Add comment for palace * Enable openmpi@5 which needs pmix>3 * don't look for intel compiler on aarch64
This commit is contained in:
parent
09810a5e7c
commit
aa39465188
@ -39,20 +39,10 @@ default:
|
||||
SPACK_TARGET_PLATFORM: "linux"
|
||||
SPACK_TARGET_ARCH: "x86_64_v3"
|
||||
|
||||
.linux_skylake:
|
||||
.linux_x86_64_v4:
|
||||
variables:
|
||||
SPACK_TARGET_PLATFORM: "linux"
|
||||
SPACK_TARGET_ARCH: "skylake_avx512"
|
||||
|
||||
.linux_icelake:
|
||||
variables:
|
||||
SPACK_TARGET_PLATFORM: "linux"
|
||||
SPACK_TARGET_ARCH: "icelake"
|
||||
|
||||
.linux_neoverse_n1:
|
||||
variables:
|
||||
SPACK_TARGET_PLATFORM: "linux"
|
||||
SPACK_TARGET_ARCH: "neoverse_n1"
|
||||
SPACK_TARGET_ARCH: "x86_64_v4"
|
||||
|
||||
.linux_neoverse_v1:
|
||||
variables:
|
||||
@ -815,104 +805,58 @@ deprecated-ci-build:
|
||||
job: deprecated-ci-generate
|
||||
|
||||
########################################
|
||||
# AWS PCLUSTER
|
||||
# AWS ParallelCluster
|
||||
########################################
|
||||
|
||||
.aws-pcluster-generate-image:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2023-05-25", "entrypoint": [""] }
|
||||
|
||||
.aws-pcluster-generate:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2024-01-29", "entrypoint": [""] }
|
||||
before_script:
|
||||
# Use gcc from local container buildcache
|
||||
# Use gcc from pre-installed spack store
|
||||
- - . "./share/spack/setup-env.sh"
|
||||
- . /etc/profile.d/modules.sh
|
||||
- spack buildcache rebuild-index /bootstrap/local-cache/
|
||||
- spack mirror add local-cache /bootstrap/local-cache
|
||||
- spack gpg trust /bootstrap/public-key
|
||||
- cd "${CI_PROJECT_DIR}" && curl -sOL https://raw.githubusercontent.com/spack/spack-configs/main/AWS/parallelcluster/postinstall.sh
|
||||
- sed -i -e "s/spack arch -t/echo ${SPACK_TARGET_ARCH}/g" postinstall.sh
|
||||
- sed -i.bkp s/"spack install gcc"/"spack install --cache-only --reuse gcc"/ postinstall.sh
|
||||
- diff postinstall.sh postinstall.sh.bkp || echo Done
|
||||
- /bin/bash postinstall.sh -fg
|
||||
- spack config --scope site add "packages:all:target:[${SPACK_TARGET_ARCH}]"
|
||||
after_script:
|
||||
- - mv "${CI_PROJECT_DIR}/postinstall.sh" "${CI_PROJECT_DIR}/jobs_scratch_dir/"
|
||||
- . "/etc/profile.d/modules.sh"
|
||||
- diff -q "/bootstrap/cloud_pipelines-config.yaml" "share/spack/gitlab/cloud_pipelines/configs/config.yaml" || echo "WARNING Install tree might have changed. You need to rebuild the pcluster-amazonlinux-2 container in spack/gitlab-runners."
|
||||
- cp "share/spack/gitlab/cloud_pipelines/configs/config.yaml" "etc/spack/"
|
||||
- /bin/bash "${SPACK_ROOT}/share/spack/gitlab/cloud_pipelines/scripts/pcluster/setup-pcluster.sh"
|
||||
- rm "etc/spack/config.yaml"
|
||||
|
||||
# Icelake (one pipeline per target)
|
||||
.aws-pcluster-icelake:
|
||||
# X86_64_V4 (one pipeline per target)
|
||||
.aws-pcluster-x86_64_v4:
|
||||
variables:
|
||||
SPACK_CI_STACK_NAME: aws-pcluster-icelake
|
||||
SPACK_CI_STACK_NAME: aws-pcluster-x86_64_v4
|
||||
|
||||
# aws-pcluster-generate-icelake:
|
||||
# extends: [ ".linux_icelake", ".aws-pcluster-icelake", ".generate-x86_64", ".tags-x86_64_v4", ".aws-pcluster-generate", ".aws-pcluster-generate-image" ]
|
||||
aws-pcluster-generate-x86_64_v4:
|
||||
extends: [ ".linux_x86_64_v4", ".aws-pcluster-x86_64_v4", ".generate-base", ".tags-x86_64_v4", ".aws-pcluster-generate"]
|
||||
|
||||
# aws-pcluster-build-icelake:
|
||||
# extends: [ ".linux_icelake", ".aws-pcluster-icelake", ".build" ]
|
||||
# trigger:
|
||||
# include:
|
||||
# - artifact: jobs_scratch_dir/cloud-ci-pipeline.yml
|
||||
# job: aws-pcluster-generate-icelake
|
||||
# strategy: depend
|
||||
# needs:
|
||||
# - artifacts: True
|
||||
# job: aws-pcluster-generate-icelake
|
||||
|
||||
# Skylake_avx512 (one pipeline per target)
|
||||
.aws-pcluster-skylake:
|
||||
variables:
|
||||
SPACK_CI_STACK_NAME: aws-pcluster-skylake
|
||||
|
||||
# aws-pcluster-generate-skylake:
|
||||
# extends: [ ".linux_skylake", ".aws-pcluster-skylake", ".generate-x86_64", ".tags-x86_64_v4", ".aws-pcluster-generate", ".aws-pcluster-generate-image" ]
|
||||
|
||||
# aws-pcluster-build-skylake:
|
||||
# extends: [ ".linux_skylake", ".aws-pcluster-skylake", ".build" ]
|
||||
# trigger:
|
||||
# include:
|
||||
# - artifact: jobs_scratch_dir/cloud-ci-pipeline.yml
|
||||
# job: aws-pcluster-generate-skylake
|
||||
# strategy: depend
|
||||
# needs:
|
||||
# - artifacts: True
|
||||
# job: aws-pcluster-generate-skylake
|
||||
|
||||
# Neoverse_n1 (one pipeline per target)
|
||||
.aws-pcluster-neoverse_n1:
|
||||
variables:
|
||||
SPACK_CI_STACK_NAME: aws-pcluster-neoverse_n1
|
||||
|
||||
# aws-pcluster-generate-neoverse_n1:
|
||||
# extends: [ ".linux_neoverse_n1", ".aws-pcluster-neoverse_n1", ".generate-aarch64", ".aws-pcluster-generate", ".aws-pcluster-generate-image" ]
|
||||
|
||||
# aws-pcluster-build-neoverse_n1:
|
||||
# extends: [ ".linux_neoverse_n1", ".aws-pcluster-neoverse_n1", ".build" ]
|
||||
# trigger:
|
||||
# include:
|
||||
# - artifact: jobs_scratch_dir/cloud-ci-pipeline.yml
|
||||
# job: aws-pcluster-generate-neoverse_n1
|
||||
# strategy: depend
|
||||
# needs:
|
||||
# - artifacts: True
|
||||
# job: aws-pcluster-generate-neoverse_n1
|
||||
aws-pcluster-build-x86_64_v4:
|
||||
extends: [ ".linux_x86_64_v4", ".aws-pcluster-x86_64_v4", ".build" ]
|
||||
trigger:
|
||||
include:
|
||||
- artifact: jobs_scratch_dir/cloud-ci-pipeline.yml
|
||||
job: aws-pcluster-generate-x86_64_v4
|
||||
strategy: depend
|
||||
needs:
|
||||
- artifacts: True
|
||||
job: aws-pcluster-generate-x86_64_v4
|
||||
|
||||
# Neoverse_v1 (one pipeline per target)
|
||||
.aws-pcluster-neoverse_v1:
|
||||
variables:
|
||||
SPACK_CI_STACK_NAME: aws-pcluster-neoverse_v1
|
||||
|
||||
# aws-pcluster-generate-neoverse_v1:
|
||||
# extends: [ ".linux_neoverse_v1", ".aws-pcluster-neoverse_v1", ".generate-aarch64", ".aws-pcluster-generate", ".aws-pcluster-generate-image" ]
|
||||
aws-pcluster-generate-neoverse_v1:
|
||||
# TODO: Use updated runner tags: https://github.com/spack/spack-infrastructure/pull/694/files
|
||||
extends: [ ".linux_neoverse_v1", ".aws-pcluster-neoverse_v1", ".generate-neoverse_v1", ".aws-pcluster-generate"]
|
||||
|
||||
# aws-pcluster-build-neoverse_v1:
|
||||
# extends: [ ".linux_neoverse_v1", ".aws-pcluster-neoverse_v1", ".build" ]
|
||||
# trigger:
|
||||
# include:
|
||||
# - artifact: jobs_scratch_dir/cloud-ci-pipeline.yml
|
||||
# job: aws-pcluster-generate-neoverse_v1
|
||||
# strategy: depend
|
||||
# needs:
|
||||
# - artifacts: True
|
||||
# job: aws-pcluster-generate-neoverse_v1
|
||||
aws-pcluster-build-neoverse_v1:
|
||||
extends: [ ".linux_neoverse_v1", ".aws-pcluster-neoverse_v1", ".build" ]
|
||||
trigger:
|
||||
include:
|
||||
- artifact: jobs_scratch_dir/cloud-ci-pipeline.yml
|
||||
job: aws-pcluster-generate-neoverse_v1
|
||||
strategy: depend
|
||||
needs:
|
||||
- artifacts: True
|
||||
job: aws-pcluster-generate-neoverse_v1
|
||||
|
||||
# Cray definitions
|
||||
.generate-cray:
|
||||
|
@ -1,7 +0,0 @@
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- any-job:
|
||||
variables:
|
||||
SPACK_TARGET_ARCH: neoverse_n1
|
||||
- build-job:
|
||||
tags: ["aarch64", "graviton2"]
|
@ -1,7 +1,4 @@
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- any-job:
|
||||
variables:
|
||||
SPACK_TARGET_ARCH: neoverse_v1
|
||||
- build-job:
|
||||
tags: ["aarch64", "graviton3"]
|
||||
|
@ -1,11 +0,0 @@
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- any-job:
|
||||
variables:
|
||||
SPACK_TARGET_ARCH: skylake_avx512
|
||||
- build-job:
|
||||
before_script:
|
||||
- - curl -LfsS "https://github.com/JuliaBinaryWrappers/GNUMake_jll.jl/releases/download/GNUMake-v4.3.0+1/GNUMake.v4.3.0.x86_64-linux-gnu.tar.gz" -o gmake.tar.gz
|
||||
- printf "fef1f59e56d2d11e6d700ba22d3444b6e583c663d6883fd0a4f63ab8bd280f0f gmake.tar.gz" | sha256sum --check --strict --quiet
|
||||
- tar -xzf gmake.tar.gz -C /usr bin/make 2> /dev/null
|
||||
tags: ["x86_64_v4"]
|
@ -1,8 +1,5 @@
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- any-job:
|
||||
variables:
|
||||
SPACK_TARGET_ARCH: icelake
|
||||
- build-job:
|
||||
before_script:
|
||||
- - curl -LfsS "https://github.com/JuliaBinaryWrappers/GNUMake_jll.jl/releases/download/GNUMake-v4.3.0+1/GNUMake.v4.3.0.x86_64-linux-gnu.tar.gz" -o gmake.tar.gz
|
130
share/spack/gitlab/cloud_pipelines/scripts/pcluster/setup-pcluster.sh
Executable file
130
share/spack/gitlab/cloud_pipelines/scripts/pcluster/setup-pcluster.sh
Executable file
@ -0,0 +1,130 @@
|
||||
#!/usr/bin/env bash
|
||||
#
|
||||
# Copyright 2013-2023 Lawrence Livermore National Security, LLC and other
|
||||
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
||||
#
|
||||
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
||||
set -e
|
||||
|
||||
# Intel compiler needs to be installed from a specific spack git commit.
|
||||
# The best solution would be to have the compilers hash (or packages contents) be part of the
|
||||
# individual packages hashes. I don't see this at the moment.
|
||||
# Set to the latest tag including a recent oneapi compiler.
|
||||
spack_intel_compiler_commit="develop-2023-08-06"
|
||||
|
||||
set_pcluster_defaults() {
|
||||
# Set versions of pre-installed software in packages.yaml
|
||||
[ -z "${SLURM_VERSION}" ] && SLURM_VERSION=$(strings /opt/slurm/lib/libslurm.so | grep -e '^VERSION' | awk '{print $2}' | sed -e 's?"??g')
|
||||
[ -z "${LIBFABRIC_VERSION}" ] && LIBFABRIC_VERSION=$(awk '/Version:/{print $2}' "$(find /opt/amazon/efa/ -name libfabric.pc | head -n1)" | sed -e 's?~??g' -e 's?amzn.*??g')
|
||||
export SLURM_VERSION LIBFABRIC_VERSION
|
||||
|
||||
envsubst < "${SPACK_ROOT}/share/spack/gitlab/cloud_pipelines/stacks/${SPACK_CI_STACK_NAME}/packages.yaml" > "${SPACK_ROOT}"/etc/spack/packages.yaml
|
||||
}
|
||||
|
||||
setup_spack() {
|
||||
spack compiler add --scope site
|
||||
spack external find --scope site
|
||||
# Remove all autotools/buildtools packages. These versions need to be managed by spack or it will
|
||||
# eventually end up in a version mismatch (e.g. when compiling gmp).
|
||||
spack tags build-tools | xargs -I {} spack config --scope site rm packages:{}
|
||||
}
|
||||
|
||||
patch_compilers_yaml() {
|
||||
# Graceful exit if package not found by spack
|
||||
set -o pipefail
|
||||
compilers_yaml="${SPACK_ROOT}/etc/spack/compilers.yaml"
|
||||
[ -f "${compilers_yaml}" ] || {
|
||||
echo "Cannot find ${compilers_yaml}, compiler setup might now be optimal."
|
||||
return
|
||||
}
|
||||
|
||||
# System ld is too old for amzn linux2
|
||||
spack_gcc_version=$(spack find --format '{version}' gcc)
|
||||
binutils_path=$(spack find -p binutils | awk '/binutils/ {print $2}' | head -n1)
|
||||
if [ -d "${binutils_path}" ] && [ -n "${spack_gcc_version}" ]; then python3 <<EOF
|
||||
import yaml
|
||||
|
||||
with open("${compilers_yaml}",'r') as f:
|
||||
compilers=yaml.safe_load(f)
|
||||
|
||||
for c in compilers["compilers"]:
|
||||
if "arm" in c["compiler"]["spec"] or "intel" in c["compiler"]["spec"] or "oneapi" in c["compiler"]["spec"] \
|
||||
or "${spack_gcc_version}" in c["compiler"]["spec"]:
|
||||
compilers["compilers"][compilers["compilers"].index(c)]["compiler"]["environment"] = {"prepend_path":{"PATH":"${binutils_path}/bin"}}
|
||||
|
||||
with open("${compilers_yaml}",'w') as f:
|
||||
yaml.dump(compilers, f)
|
||||
EOF
|
||||
fi
|
||||
# Oneapi needs extra_rpath to gcc libstdc++.so.6
|
||||
if [ "x86_64" == "$(arch)" ] && oneapi_gcc_version=$(spack find --format '{compiler}' intel-oneapi-compilers | sed -e 's/=//g') && \
|
||||
[ -n "${oneapi_gcc_version}" ] && oneapi_gcc_path=$(spack find -p "${oneapi_gcc_version}" | grep "${oneapi_gcc_version}" | awk '{print $2}' | head -n1) && \
|
||||
[ -d "${oneapi_gcc_path}" ]; then python3 <<EOF
|
||||
import yaml
|
||||
|
||||
with open("${compilers_yaml}",'r') as f:
|
||||
compilers=yaml.safe_load(f)
|
||||
|
||||
for c in compilers["compilers"]:
|
||||
if "oneapi" in c["compiler"]["spec"]:
|
||||
compilers["compilers"][compilers["compilers"].index(c)]["compiler"]["extra_rpaths"] = ["${oneapi_gcc_path}/lib64"]
|
||||
|
||||
with open("${compilers_yaml}",'w') as f:
|
||||
yaml.dump(compilers, f)
|
||||
EOF
|
||||
fi
|
||||
}
|
||||
|
||||
install_compilers() {
|
||||
# We need to treat compilers as essentially external, i.e. their installation location
|
||||
# (including hash) must not change when any changes are pushed to spack. The reason is that
|
||||
# changes in the compilers are not reflected in the package hashes built in the CI. Hence, those
|
||||
# packages will reference a wrong compiler path once the path changes.
|
||||
|
||||
# `gcc@12.3.0%gcc@7.3.1` is created as part of building the pipeline containers.
|
||||
# `ghcr.io/spack/pcluster-amazonlinux-2:v2024-01-29` produced the following hashes.
|
||||
if [ "x86_64" == "$(arch)" ]; then
|
||||
gcc_hash="vxlibl3ubl5ptwzb3zydgksfa5osdea6"
|
||||
else
|
||||
gcc_hash="bikooik6f3fyrkroarulsadbii43ggz5"
|
||||
fi
|
||||
|
||||
spack install /${gcc_hash}
|
||||
(
|
||||
spack load gcc
|
||||
spack compiler add --scope site
|
||||
)
|
||||
|
||||
# Install Intel compilers through a static spack version such that the compiler's hash does not change.
|
||||
# The compilers needs to be in the same install tree as the rest of the software such that the path
|
||||
# relocation works correctly. This holds the danger that this part will fail when the current spack gets
|
||||
# incompatible with the one in $spack_intel_compiler_commit. Therefore, we make intel installations optional
|
||||
# in package.yaml files.
|
||||
if [ "x86_64" == "$(arch)" ]; then
|
||||
(
|
||||
CURRENT_SPACK_ROOT=${SPACK_ROOT}
|
||||
DIR="$(mktemp -d)"
|
||||
cd "${DIR}"
|
||||
# This needs to include commit 361a185ddb such that `ifx` picks up the correct toolchain. Otherwise
|
||||
# this leads to libstdc++.so errors during linking (e.g. slepc).
|
||||
git clone --depth=1 -b ${spack_intel_compiler_commit} https://github.com/spack/spack.git \
|
||||
&& cd spack \
|
||||
&& curl -sL https://github.com/spack/spack/pull/40557.patch | patch -p1 \
|
||||
&& curl -sL https://github.com/spack/spack/pull/40561.patch | patch -p1 \
|
||||
&& cp "${CURRENT_SPACK_ROOT}/etc/spack/config.yaml" etc/spack/ \
|
||||
&& cp "${CURRENT_SPACK_ROOT}/etc/spack/compilers.yaml" etc/spack/ \
|
||||
&& cp "${CURRENT_SPACK_ROOT}/etc/spack/packages.yaml" etc/spack/ \
|
||||
&& . share/spack/setup-env.sh \
|
||||
&& spack install intel-oneapi-compilers-classic
|
||||
rm -rf "${DIR}"
|
||||
)
|
||||
bash -c ". \"$(spack location -i intel-oneapi-compilers)\"/setvars.sh; spack compiler add --scope site" \
|
||||
|| true
|
||||
spack clean -m
|
||||
fi
|
||||
}
|
||||
|
||||
set_pcluster_defaults
|
||||
setup_spack
|
||||
install_compilers
|
||||
patch_compilers_yaml
|
@ -1,56 +0,0 @@
|
||||
spack:
|
||||
view: false
|
||||
|
||||
definitions:
|
||||
- compiler_specs:
|
||||
- gcc
|
||||
- gettext
|
||||
|
||||
- compiler_target:
|
||||
- '%gcc@7.3.1 target=x86_64_v3'
|
||||
|
||||
- optimized_configs:
|
||||
# - gromacs
|
||||
- lammps
|
||||
# - mpas-model
|
||||
- openfoam
|
||||
# - palace
|
||||
# - py-devito
|
||||
# - quantum-espresso
|
||||
# - wrf
|
||||
|
||||
- optimized_libs:
|
||||
- mpich
|
||||
- openmpi
|
||||
|
||||
specs:
|
||||
- matrix:
|
||||
- - $compiler_specs
|
||||
- - $compiler_target
|
||||
- $optimized_configs
|
||||
# - $optimized_libs
|
||||
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- build-job:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2023-05-25", "entrypoint": [""] }
|
||||
before_script:
|
||||
- - . "./share/spack/setup-env.sh"
|
||||
- . /etc/profile.d/modules.sh
|
||||
- spack --version
|
||||
- spack arch
|
||||
# Use gcc from local container buildcache
|
||||
- - spack buildcache rebuild-index /bootstrap/local-cache/
|
||||
- sed -i.bkp s/"spack install gcc"/"spack install --cache-only --reuse gcc"/ /bootstrap/postinstall.sh
|
||||
- spack mirror add local-cache /bootstrap/local-cache
|
||||
- spack gpg trust /bootstrap/public-key
|
||||
- diff /bootstrap/postinstall.sh /bootstrap/postinstall.sh.bkp || echo Done
|
||||
- - /bin/bash "${SPACK_ARTIFACTS_ROOT}/postinstall.sh" -fg
|
||||
- spack config --scope site add "packages:all:target:[${SPACK_TARGET_ARCH}]"
|
||||
- signing-job:
|
||||
before_script:
|
||||
# Do not distribute Intel & ARM binaries
|
||||
- - for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep intel-oneapi | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
- for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep armpl | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
cdash:
|
||||
build-group: AWS Packages
|
@ -1,58 +0,0 @@
|
||||
spack:
|
||||
view: false
|
||||
|
||||
definitions:
|
||||
- compiler_specs:
|
||||
- gcc
|
||||
- gettext
|
||||
|
||||
- compiler_target:
|
||||
- '%gcc@7.3.1 target=aarch64'
|
||||
|
||||
- optimized_configs:
|
||||
- gromacs
|
||||
# - lammps
|
||||
# - mpas-model
|
||||
- openfoam
|
||||
- palace
|
||||
# - py-devito
|
||||
# - quantum-espresso
|
||||
# - wrf
|
||||
|
||||
- optimized_libs:
|
||||
- mpich
|
||||
- openmpi
|
||||
|
||||
specs:
|
||||
- matrix:
|
||||
- - $compiler_specs
|
||||
- - $compiler_target
|
||||
- $optimized_configs
|
||||
- $optimized_libs
|
||||
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- build-job:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2023-05-25", "entrypoint": [""] }
|
||||
tags: ["aarch64"]
|
||||
before_script:
|
||||
- - . "./share/spack/setup-env.sh"
|
||||
- . /etc/profile.d/modules.sh
|
||||
- spack --version
|
||||
- spack arch
|
||||
# Use gcc from local container buildcache
|
||||
- - spack buildcache rebuild-index /bootstrap/local-cache/
|
||||
- sed -i.bkp s/"spack install gcc"/"spack install --cache-only --reuse gcc"/ /bootstrap/postinstall.sh
|
||||
- spack mirror add local-cache /bootstrap/local-cache
|
||||
- spack gpg trust /bootstrap/public-key
|
||||
- diff /bootstrap/postinstall.sh /bootstrap/postinstall.sh.bkp || echo Done
|
||||
- - /bin/bash "${SPACK_ARTIFACTS_ROOT}/postinstall.sh" -fg
|
||||
- spack config --scope site add "packages:all:target:[${SPACK_TARGET_ARCH}]"
|
||||
- signing-job:
|
||||
before_script:
|
||||
# Do not distribute Intel & ARM binaries
|
||||
- - for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep intel-oneapi | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
- for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep armpl | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
|
||||
cdash:
|
||||
build-group: AWS Packages
|
@ -0,0 +1,64 @@
|
||||
--- # Neoverse N1 / V1 packages
|
||||
packages:
|
||||
acfl:
|
||||
require:
|
||||
- one_of: ["%gcc target=aarch64"]
|
||||
message: "Clang based compilers need GCC libraries and they should be made available for the wide range of CPUs they actually support.
|
||||
Edit $SPACK_ROOT/etc/spack/packages.yaml to change this default."
|
||||
gromacs:
|
||||
require:
|
||||
- one_of:
|
||||
- "gromacs@2021.3 %arm ^fftw ^openmpi"
|
||||
- "gromacs@2021.3 %gcc ^armpl-gcc ^openmpi"
|
||||
libfabric:
|
||||
buildable: true
|
||||
externals:
|
||||
- prefix: /opt/amazon/efa/
|
||||
spec: libfabric@${LIBFABRIC_VERSION}
|
||||
require: ['fabrics=shm,efa']
|
||||
llvm:
|
||||
variants: ~lldb
|
||||
mpas-model:
|
||||
require: "precision=single make_target=llvm %arm ^parallelio+pnetcdf"
|
||||
mpich:
|
||||
require: "mpich pmi=pmi2 device=ch4 netmod=ofi +slurm"
|
||||
nvhpc:
|
||||
require:
|
||||
- one_of:
|
||||
- "nvhpc %gcc target=aarch64"
|
||||
message: "NVHPC should be built with GCC and should be made available for the wide range of CPUs they actually support.
|
||||
Edit $SPACK_ROOT/etc/spack/packages.yaml to change this default."
|
||||
openfoam:
|
||||
require: "openfoam %gcc ^scotch@6.0.9"
|
||||
openmpi:
|
||||
variants: ~atomics ~cuda ~cxx ~cxx_exceptions ~internal-hwloc ~java +legacylaunchers ~lustre ~memchecker +pmi +romio ~singularity +vt +wrapper-rpath fabrics=ofi schedulers=slurm
|
||||
require: '@4:'
|
||||
# Palace does not build correctly with armpl until https://github.com/awslabs/palace/pull/207 is merged into a version.
|
||||
# palace:
|
||||
# require:
|
||||
# - one_of: ["palace cxxflags=\"-include cstdint\" ^fmt@9.1.0"]
|
||||
pmix:
|
||||
require: "pmix@3:"
|
||||
quantum-espresso:
|
||||
require: "quantum-espresso@6.6 %gcc ^armpl-gcc"
|
||||
slurm:
|
||||
buildable: false
|
||||
externals:
|
||||
- prefix: /opt/slurm/
|
||||
spec: slurm@${SLURM_VERSION} +pmix
|
||||
wrf:
|
||||
require:
|
||||
- one_of:
|
||||
- "wrf%arm"
|
||||
- "wrf%gcc"
|
||||
all:
|
||||
compiler: [gcc, arm, nvhpc, clang]
|
||||
providers:
|
||||
blas: [armpl-gcc, openblas]
|
||||
fftw-api: [armpl-gcc, fftw]
|
||||
lapack: [armpl-gcc, openblas]
|
||||
mpi: [openmpi, mpich]
|
||||
scalapack: [netlib-scalapack]
|
||||
permissions:
|
||||
read: world
|
||||
write: user
|
@ -2,52 +2,27 @@ spack:
|
||||
view: false
|
||||
|
||||
definitions:
|
||||
- compiler_specs:
|
||||
- gcc
|
||||
- gettext
|
||||
|
||||
- compiler_target:
|
||||
- '%gcc@7.3.1 target=aarch64'
|
||||
|
||||
- optimized_configs:
|
||||
- gromacs
|
||||
# - lammps
|
||||
# - mpas-model
|
||||
- openfoam
|
||||
- palace
|
||||
# - py-devito
|
||||
# - quantum-espresso
|
||||
# - wrf
|
||||
|
||||
- optimized_libs:
|
||||
- mpich
|
||||
- openmpi
|
||||
- gromacs target=neoverse_v1
|
||||
- gromacs target=neoverse_n1
|
||||
|
||||
specs:
|
||||
- matrix:
|
||||
- - $compiler_specs
|
||||
- - $compiler_target
|
||||
- $optimized_configs
|
||||
- $optimized_libs
|
||||
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- build-job:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2023-05-25", "entrypoint": [""] }
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2024-01-29", "entrypoint": [""] }
|
||||
tags: ["aarch64"]
|
||||
before_script:
|
||||
- - . "./share/spack/setup-env.sh"
|
||||
- . /etc/profile.d/modules.sh
|
||||
- spack --version
|
||||
- spack arch
|
||||
# Use gcc from local container buildcache
|
||||
- - spack buildcache rebuild-index /bootstrap/local-cache/
|
||||
- sed -i.bkp s/"spack install gcc"/"spack install --cache-only --reuse gcc"/ /bootstrap/postinstall.sh
|
||||
- spack mirror add local-cache /bootstrap/local-cache
|
||||
- spack gpg trust /bootstrap/public-key
|
||||
- diff /bootstrap/postinstall.sh /bootstrap/postinstall.sh.bkp || echo Done
|
||||
- - /bin/bash "${SPACK_ARTIFACTS_ROOT}/postinstall.sh" -fg
|
||||
- spack config --scope site add "packages:all:target:[${SPACK_TARGET_ARCH}]"
|
||||
# Use gcc from pre-installed spack store
|
||||
- - cp share/spack/gitlab/cloud_pipelines/configs/config.yaml etc/spack/
|
||||
- /bin/bash "${SPACK_ROOT}/share/spack/gitlab/cloud_pipelines/scripts/pcluster/setup-pcluster.sh"
|
||||
- rm etc/spack/config.yaml
|
||||
- signing-job:
|
||||
before_script:
|
||||
# Do not distribute Intel & ARM binaries
|
||||
|
@ -1,56 +0,0 @@
|
||||
spack:
|
||||
view: false
|
||||
|
||||
definitions:
|
||||
- compiler_specs:
|
||||
- gcc
|
||||
- gettext
|
||||
|
||||
- compiler_target:
|
||||
- '%gcc@7.3.1 target=x86_64_v3'
|
||||
|
||||
- optimized_configs:
|
||||
# - gromacs
|
||||
- lammps
|
||||
# - mpas-model
|
||||
- openfoam
|
||||
# - palace
|
||||
# - py-devito
|
||||
# - quantum-espresso
|
||||
# - wrf
|
||||
|
||||
- optimized_libs:
|
||||
- mpich
|
||||
- openmpi
|
||||
|
||||
specs:
|
||||
- matrix:
|
||||
- - $compiler_specs
|
||||
- - $compiler_target
|
||||
- $optimized_configs
|
||||
# - $optimized_libs
|
||||
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- build-job:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2023-05-25", "entrypoint": [""] }
|
||||
before_script:
|
||||
- - . "./share/spack/setup-env.sh"
|
||||
- . /etc/profile.d/modules.sh
|
||||
- spack --version
|
||||
- spack arch
|
||||
# Use gcc from local container buildcache
|
||||
- - spack buildcache rebuild-index /bootstrap/local-cache/
|
||||
- sed -i.bkp s/"spack install gcc"/"spack install --cache-only --reuse gcc"/ /bootstrap/postinstall.sh
|
||||
- spack mirror add local-cache /bootstrap/local-cache
|
||||
- spack gpg trust /bootstrap/public-key
|
||||
- diff /bootstrap/postinstall.sh /bootstrap/postinstall.sh.bkp || echo Done
|
||||
- - /bin/bash "${SPACK_ARTIFACTS_ROOT}/postinstall.sh" -fg
|
||||
- spack config --scope site add "packages:all:target:[${SPACK_TARGET_ARCH}]"
|
||||
- signing-job:
|
||||
before_script:
|
||||
# Do not distribute Intel & ARM binaries
|
||||
- - for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep intel-oneapi | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
- for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep armpl | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
cdash:
|
||||
build-group: AWS Packages
|
@ -0,0 +1,110 @@
|
||||
--- # x86_64_v4 packages (ice/skylake) & x86_64_v3 packages (zen2/3)
|
||||
packages:
|
||||
cpio:
|
||||
require:
|
||||
- one_of:
|
||||
- "cflags=-std=c18 target=x86_64_v4"
|
||||
- "cflags=-std=c18 target=x86_64_v3"
|
||||
- "%gcc"
|
||||
gromacs:
|
||||
require:
|
||||
- one_of:
|
||||
- "+intel_provided_gcc %intel ^intel-oneapi-mkl target=x86_64_v4"
|
||||
- "+intel_provided_gcc %intel ^intel-oneapi-mkl target=x86_64_v3"
|
||||
- "%gcc"
|
||||
intel-mpi:
|
||||
variants: +external-libfabric
|
||||
intel-oneapi-compilers:
|
||||
require: "intel-oneapi-compilers %gcc target=x86_64_v3"
|
||||
intel-oneapi-mpi:
|
||||
variants: +external-libfabric generic-names=True
|
||||
lammps:
|
||||
require:
|
||||
- one_of:
|
||||
- "lammps_sizes=bigbig +molecule +kspace +rigid +asphere +opt +openmp +openmp-package +intel %intel ^intel-oneapi-mkl target=x86_64_v4"
|
||||
- "lammps_sizes=bigbig +molecule +kspace +rigid +asphere +opt +openmp +openmp-package %intel ^intel-oneapi-mkl target=x86_64_v3"
|
||||
- "%gcc"
|
||||
libidn2:
|
||||
require:
|
||||
- one_of:
|
||||
- "cflags=-std=c18 target=x86_64_v4"
|
||||
- "cflags=-std=c18 target=x86_64_v3"
|
||||
- '%gcc'
|
||||
libfabric:
|
||||
buildable: true
|
||||
externals:
|
||||
- prefix: /opt/amazon/efa/
|
||||
spec: libfabric@${LIBFABRIC_VERSION}
|
||||
require: ['fabrics=shm,efa']
|
||||
libunistring:
|
||||
require:
|
||||
- one_of:
|
||||
- "cflags=-std=c18 target=x86_64_v4"
|
||||
- "cflags=-std=c18 target=x86_64_v3"
|
||||
- "%gcc"
|
||||
mpas-model:
|
||||
require:
|
||||
- one_of:
|
||||
- "precision=single %intel ^parallelio+pnetcdf target=x86_64_v4"
|
||||
- "precision=single %intel ^parallelio+pnetcdf target=x86_64_v3"
|
||||
- "%gcc"
|
||||
mpich:
|
||||
require:
|
||||
- one_of:
|
||||
- "mpich pmi=pmi2 device=ch4 netmod=ofi +slurm target=x86_64_v4"
|
||||
- "mpich pmi=pmi2 device=ch4 netmod=ofi +slurm target=x86_64_v3"
|
||||
openfoam:
|
||||
require:
|
||||
- one_of:
|
||||
- "openfoam %gcc ^scotch@6.0.9 target=x86_64_v4"
|
||||
- "openfoam %gcc ^scotch@6.0.9 target=x86_64_v3"
|
||||
openmpi:
|
||||
variants: ~atomics ~cuda ~cxx ~cxx_exceptions ~internal-hwloc ~java +legacylaunchers ~lustre ~memchecker +pmi +romio ~singularity +vt +wrapper-rpath fabrics=ofi schedulers=slurm
|
||||
require:
|
||||
- one_of:
|
||||
- "openmpi @4: target=x86_64_v4"
|
||||
- "openmpi @4: target=x86_64_v3"
|
||||
palace:
|
||||
require:
|
||||
- one_of:
|
||||
- "palace %oneapi ^fmt@9.1.0 target=x86_64_v4"
|
||||
- "palace %oneapi ^fmt@9.1.0 target=x86_64_v3"
|
||||
- "%gcc ^fmt@9.1.0"
|
||||
pmix:
|
||||
require:
|
||||
- one_of:
|
||||
- "pmix@3: target=x86_64_v4"
|
||||
- "pmix@3: target=x86_64_v3"
|
||||
quantum-espresso:
|
||||
require:
|
||||
- one_of:
|
||||
- "quantum-espresso@6.6 %intel ^intel-oneapi-mkl+cluster target=x86_64_v4"
|
||||
- "quantum-espresso@6.6 %intel ^intel-oneapi-mkl+cluster target=x86_64_v3"
|
||||
- "%gcc"
|
||||
slurm:
|
||||
buildable: false
|
||||
externals:
|
||||
- prefix: /opt/slurm/
|
||||
spec: slurm@${SLURM_VERSION} +pmix
|
||||
wrf:
|
||||
require:
|
||||
- one_of:
|
||||
- "wrf@4 build_type=dm+sm %intel target=x86_64_v4"
|
||||
- "wrf@4 build_type=dm+sm %intel target=x86_64_v3"
|
||||
- "wrf@4.2.2 +netcdf_classic fflags=\"-fp-model fast=2 -no-heap-arrays -no-prec-div -no-prec-sqrt -fno-common\" build_type=dm+sm %intel target=x86_64_v3"
|
||||
- "%gcc"
|
||||
all:
|
||||
compiler: [intel, gcc]
|
||||
permissions:
|
||||
read: world
|
||||
write: user
|
||||
providers:
|
||||
blas: [intel-oneapi-mkl, intel-mkl]
|
||||
daal: [intel-oneapi-dal, intel-daal]
|
||||
fftw-api: [intel-oneapi-mkl, intel-mkl]
|
||||
ipp: [intel-oneapi-ipp, intel-ipp]
|
||||
lapack: [intel-oneapi-mkl, intel-mkl]
|
||||
mkl: [intel-oneapi-mkl, intel-mkl]
|
||||
mpi: [intel-oneapi-mpi, openmpi, mpich]
|
||||
tbb: [intel-oneapi-tbb, intel-tbb]
|
||||
scalapack: [intel-oneapi-mkl, intel-mkl]
|
@ -0,0 +1,32 @@
|
||||
spack:
|
||||
view: false
|
||||
|
||||
definitions:
|
||||
|
||||
- optimized_configs:
|
||||
- palace target=x86_64_v4
|
||||
- palace target=x86_64_v3
|
||||
|
||||
specs:
|
||||
- $optimized_configs
|
||||
|
||||
ci:
|
||||
pipeline-gen:
|
||||
- build-job:
|
||||
image: { "name": "ghcr.io/spack/pcluster-amazonlinux-2:v2024-01-29", "entrypoint": [""] }
|
||||
before_script:
|
||||
- - . "./share/spack/setup-env.sh"
|
||||
- . /etc/profile.d/modules.sh
|
||||
- spack --version
|
||||
- spack arch
|
||||
# Use gcc from pre-installed spack store
|
||||
- - cp share/spack/gitlab/cloud_pipelines/configs/config.yaml etc/spack/
|
||||
- /bin/bash "${SPACK_ROOT}/share/spack/gitlab/cloud_pipelines/scripts/pcluster/setup-pcluster.sh"
|
||||
- rm etc/spack/config.yaml
|
||||
- signing-job:
|
||||
before_script:
|
||||
# Do not distribute Intel & ARM binaries
|
||||
- - for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep intel-oneapi | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
- for i in $(aws s3 ls --recursive ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/ | grep armpl | awk '{print $4}' | sed -e 's?^.*build_cache/??g'); do aws s3 rm ${SPACK_REMOTE_MIRROR_OVERRIDE}/build_cache/$i; done
|
||||
cdash:
|
||||
build-group: AWS Packages
|
Loading…
Reference in New Issue
Block a user