release workflow: Add build scripts for jobs and means to upload pkgs
This commit is contained in:
		 Scott Wittenburg
					Scott Wittenburg
				
			
				
					committed by
					
						 Peter Scheibel
						Peter Scheibel
					
				
			
			
				
	
			
			
			 Peter Scheibel
						Peter Scheibel
					
				
			
						parent
						
							a6e8e889b2
						
					
				
				
					commit
					5600c9f0d9
				
			
							
								
								
									
										13
									
								
								bin/rebuild-index.sh
									
									
									
									
									
										Executable file
									
								
							
							
						
						
									
										13
									
								
								bin/rebuild-index.sh
									
									
									
									
									
										Executable file
									
								
							| @@ -0,0 +1,13 @@ | ||||
| #!/bin/bash | ||||
|  | ||||
| # Copyright 2013-2019 Lawrence Livermore National Security, LLC and other | ||||
| # Spack Project Developers. See the top-level COPYRIGHT file for details. | ||||
| # | ||||
| # SPDX-License-Identifier: (Apache-2.0 OR MIT) | ||||
|  | ||||
| set -x | ||||
|  | ||||
| SPACK_BIN_DIR="${CI_PROJECT_DIR}/bin" | ||||
| export PATH="${SPACK_BIN_DIR}:${PATH}" | ||||
|  | ||||
| spack upload-s3 index | ||||
							
								
								
									
										316
									
								
								bin/rebuild-package.sh
									
									
									
									
									
										Executable file
									
								
							
							
						
						
									
										316
									
								
								bin/rebuild-package.sh
									
									
									
									
									
										Executable file
									
								
							| @@ -0,0 +1,316 @@ | ||||
| #!/bin/bash | ||||
|  | ||||
| # Copyright 2013-2019 Lawrence Livermore National Security, LLC and other | ||||
| # Spack Project Developers. See the top-level COPYRIGHT file for details. | ||||
| # | ||||
| # SPDX-License-Identifier: (Apache-2.0 OR MIT) | ||||
|  | ||||
| ### | ||||
| ### This script represents a gitlab-ci job, corresponding to a single release | ||||
| ### spec.  As such this script must first decide whether or not the spec it | ||||
| ### has been assigned is up to date on the remote binary mirror.  If it is | ||||
| ### not (i.e. the source code has changed in a way that caused a change in the | ||||
| ### full_hash of the spec), this script will build the package, create a | ||||
| ### binary cache for it, and then push all related files to the remote binary | ||||
| ### mirror.  This script also communicates with a remote CDash instance to | ||||
| ### share status on the package build process. | ||||
| ### | ||||
| ### The following environment variables are expected to be set in order for | ||||
| ### the various elements in this script to function properly.  Listed first | ||||
| ### are two defaults we rely on from gitlab, then three we set up in the | ||||
| ### variables section of gitlab ourselves, and finally four variables | ||||
| ### written into the .gitlab-ci.yml file. | ||||
| ### | ||||
| ### CI_PROJECT_DIR | ||||
| ### CI_JOB_NAME | ||||
| ### | ||||
| ### AWS_ACCESS_KEY_ID | ||||
| ### AWS_SECRET_ACCESS_KEY | ||||
| ### SPACK_SIGNING_KEY | ||||
| ### | ||||
| ### CDASH_BASE_URL | ||||
| ### ROOT_SPEC | ||||
| ### DEPENDENCIES | ||||
| ### MIRROR_URL | ||||
| ### | ||||
|  | ||||
| shopt -s expand_aliases | ||||
|  | ||||
| export FORCE_UNSAFE_CONFIGURE=1 | ||||
|  | ||||
| TEMP_DIR="${CI_PROJECT_DIR}/jobs_scratch_dir" | ||||
|  | ||||
| JOB_LOG_DIR="${TEMP_DIR}/logs" | ||||
| SPEC_DIR="${TEMP_DIR}/specs" | ||||
| LOCAL_MIRROR="${CI_PROJECT_DIR}/local_mirror" | ||||
| BUILD_CACHE_DIR="${LOCAL_MIRROR}/build_cache" | ||||
| SPACK_BIN_DIR="${CI_PROJECT_DIR}/bin" | ||||
| CDASH_UPLOAD_URL="${CDASH_BASE_URL}/submit.php?project=Spack" | ||||
| DEP_JOB_RELATEBUILDS_URL="${CDASH_BASE_URL}/api/v1/relateBuilds.php" | ||||
| declare -a JOB_DEPS_PKG_NAMES | ||||
|  | ||||
| export SPACK_ROOT=${CI_PROJECT_DIR} | ||||
| export PATH="${SPACK_BIN_DIR}:${PATH}" | ||||
| export GNUPGHOME="${CI_PROJECT_DIR}/opt/spack/gpg" | ||||
|  | ||||
| mkdir -p ${JOB_LOG_DIR} | ||||
| mkdir -p ${SPEC_DIR} | ||||
|  | ||||
| cleanup() { | ||||
|     set +x | ||||
|  | ||||
|     if [ -z "$exit_code" ] ; then | ||||
|  | ||||
|         exit_code=$1 | ||||
|         if [ -z "$exit_code" ] ; then | ||||
|             exit_code=0 | ||||
|         fi | ||||
|  | ||||
|         restore_io | ||||
|  | ||||
|         if [ "$( type -t finalize )" '=' 'function' ] ; then | ||||
|             finalize "$JOB_LOG_DIR/cdash_log.txt" | ||||
|         fi | ||||
|  | ||||
|         # We can clean these out later on, once we have a good sense for | ||||
|         # how the logging infrastructure is working | ||||
|         # rm -rf "$JOB_LOG_DIR" | ||||
|     fi | ||||
|  | ||||
|     \exit $exit_code | ||||
| } | ||||
|  | ||||
| alias exit='cleanup' | ||||
|  | ||||
| begin_logging() { | ||||
|     trap "cleanup 1; \\exit \$exit_code" INT TERM QUIT | ||||
|     trap "cleanup 0; \\exit \$exit_code" EXIT | ||||
|  | ||||
|     rm -rf "$JOB_LOG_DIR/cdash_log.txt" | ||||
|  | ||||
|     # NOTE: Here, some redirects are set up | ||||
|     exec 3>&1 # fd 3 is now a dup of stdout | ||||
|     exec 4>&2 # fd 4 is now a dup of stderr | ||||
|  | ||||
|     # stdout and stderr are joined and redirected to the log | ||||
|     exec &> "$JOB_LOG_DIR/cdash_log.txt" | ||||
|  | ||||
|     set -x | ||||
| } | ||||
|  | ||||
| restore_io() { | ||||
|     exec  >&- | ||||
|     exec 2>&- | ||||
|  | ||||
|     exec  >&3 | ||||
|     exec 2>&4 | ||||
|  | ||||
|     exec 3>&- | ||||
|     exec 4>&- | ||||
| } | ||||
|  | ||||
| finalize() { | ||||
|     # If you define a finalize function: | ||||
|     #  - it will always be called at the very end of the script | ||||
|     #  - the log file will be passed in as the first argument, and | ||||
|     #  - the code in this function will not be logged. | ||||
|     echo "The full log file is located at $1" | ||||
|     # TODO: send this log data to cdash! | ||||
| } | ||||
|  | ||||
| check_error() | ||||
| { | ||||
|     local last_exit_code=$1 | ||||
|     local last_cmd=$2 | ||||
|     if [[ ${last_exit_code} -ne 0 ]]; then | ||||
|         echo "${last_cmd} exited with code ${last_exit_code}" | ||||
|         echo "TERMINATING JOB" | ||||
|         exit 1 | ||||
|     else | ||||
|         echo "${last_cmd} completed successfully" | ||||
|     fi | ||||
| } | ||||
|  | ||||
| extract_build_id() | ||||
| { | ||||
|     LINES_TO_SEARCH=$1 | ||||
|     regex="buildSummary\.php\?buildid=([[:digit:]]+)" | ||||
|     SINGLE_LINE_OUTPUT=$(echo ${LINES_TO_SEARCH} | tr -d '\n') | ||||
|  | ||||
|     if [[ ${SINGLE_LINE_OUTPUT} =~ ${regex} ]]; then | ||||
|         echo "${BASH_REMATCH[1]}" | ||||
|     else | ||||
|         echo "NONE" | ||||
|     fi | ||||
| } | ||||
|  | ||||
| get_relate_builds_post_data() | ||||
| { | ||||
|   cat <<EOF | ||||
| { | ||||
|   "project": "${1}", | ||||
|   "buildid": ${2}, | ||||
|   "relatedid": ${3}, | ||||
|   "relationship": "depends on" | ||||
| } | ||||
| EOF | ||||
| } | ||||
|  | ||||
| gen_full_specs_for_job_and_deps() { | ||||
|  | ||||
|     read -ra PARTSARRAY <<< "${CI_JOB_NAME}" | ||||
|     local pkgName="${PARTSARRAY[0]}" | ||||
|     local pkgVersion="${PARTSARRAY[1]}" | ||||
|     local compiler="${PARTSARRAY[2]}" | ||||
|     local osarch="${PARTSARRAY[3]}" | ||||
|  | ||||
|     JOB_SPEC_NAME="${pkgName}@${pkgVersion}%${compiler} arch=${osarch}" | ||||
|     JOB_PKG_NAME="${pkgName}" | ||||
|     SPEC_YAML_PATH="${SPEC_DIR}/${pkgName}.yaml" | ||||
|     local root_spec_name="${ROOT_SPEC}" | ||||
|     local spec_names_to_save="${pkgName}" | ||||
|  | ||||
|     IFS=';' read -ra DEPS <<< "${DEPENDENCIES}" | ||||
|     for i in "${DEPS[@]}"; do | ||||
|         read -ra PARTSARRAY <<< "${i}" | ||||
|         pkgName="${PARTSARRAY[0]}" | ||||
|         spec_names_to_save="${spec_names_to_save} ${pkgName}" | ||||
|         JOB_DEPS_PKG_NAMES+=("${pkgName}") | ||||
|     done | ||||
|  | ||||
|     spack -d buildcache save-yaml --specs "${spec_names_to_save}" --root-spec "${root_spec_name}" --yaml-dir "${SPEC_DIR}" | ||||
| } | ||||
|  | ||||
| begin_logging | ||||
|  | ||||
| gen_full_specs_for_job_and_deps | ||||
|  | ||||
| echo "Building package ${JOB_SPEC_NAME}, ${HASH}, ${MIRROR_URL}" | ||||
|  | ||||
| # Finally, list the compilers spack knows about | ||||
| echo "Compiler Configurations:" | ||||
| spack config get compilers | ||||
|  | ||||
| # Make the build_cache directory if it doesn't exist | ||||
| mkdir -p "${BUILD_CACHE_DIR}" | ||||
|  | ||||
| # Get buildcache name so we can write a CDash build id file in the right place. | ||||
| # If we're unable to get the buildcache name, we may have encountered a problem | ||||
| # concretizing the spec, or some other issue that will eventually cause the job | ||||
| # to fail. | ||||
| JOB_BUILD_CACHE_ENTRY_NAME=`spack -d buildcache get-buildcache-name --spec-yaml "${SPEC_YAML_PATH}"` | ||||
| if [[ $? -ne 0 ]]; then | ||||
|     echo "ERROR, unable to get buildcache entry name for job ${CI_JOB_NAME} (spec: ${JOB_SPEC_NAME})" | ||||
|     exit 1 | ||||
| fi | ||||
|  | ||||
| # This should create the directory we referred to as GNUPGHOME earlier | ||||
| spack gpg list | ||||
|  | ||||
| # Importing the secret key using gpg2 directly should allow to | ||||
| # sign and verify both | ||||
| set +x | ||||
| KEY_IMPORT_RESULT=`echo ${SPACK_SIGNING_KEY} | base64 --decode | gpg2 --import` | ||||
| check_error $? "gpg2 --import" | ||||
| set -x | ||||
|  | ||||
| spack gpg list --trusted | ||||
| spack gpg list --signing | ||||
|  | ||||
| # Whether we have to build the spec or download it pre-built, we expect to find | ||||
| # the cdash build id file sitting in this location afterwards. | ||||
| JOB_CDASH_ID_FILE="${BUILD_CACHE_DIR}/${JOB_BUILD_CACHE_ENTRY_NAME}.cdashid" | ||||
|  | ||||
| # Finally, we can check the spec we have been tasked with build against | ||||
| # the built binary on the remote mirror to see if it needs to be rebuilt | ||||
| spack -d buildcache check --spec-yaml "${SPEC_YAML_PATH}" --mirror-url "${MIRROR_URL}" --rebuild-on-error | ||||
|  | ||||
| if [[ $? -ne 0 ]]; then | ||||
|     # Configure mirror | ||||
|     spack mirror add local_artifact_mirror "file://${LOCAL_MIRROR}" | ||||
|  | ||||
|     JOB_CDASH_ID="NONE" | ||||
|  | ||||
|     # Install package, using the buildcache from the local mirror to | ||||
|     # satisfy dependencies. | ||||
|     BUILD_ID_LINE=`spack -d -k -v install --use-cache --cdash-upload-url "${CDASH_UPLOAD_URL}" --cdash-build "${JOB_SPEC_NAME}" --cdash-site "Spack AWS Gitlab Instance" --cdash-track "Experimental" -f "${SPEC_YAML_PATH}" | grep "buildSummary\\.php"` | ||||
|     check_error $? "spack install" | ||||
|  | ||||
|     # By parsing the output of the "spack install" command, we can get the | ||||
|     # buildid generated for us by CDash | ||||
|     JOB_CDASH_ID=$(extract_build_id "${BUILD_ID_LINE}") | ||||
|  | ||||
|     # Create buildcache entry for this package, reading the spec from the yaml | ||||
|     # file. | ||||
|     spack -d buildcache create --spec-yaml "${SPEC_YAML_PATH}" -a -f -d "${LOCAL_MIRROR}" --no-rebuild-index | ||||
|     check_error $? "spack buildcache create" | ||||
|  | ||||
|     # Write the .cdashid file to the buildcache as well | ||||
|     echo "${JOB_CDASH_ID}" >> ${JOB_CDASH_ID_FILE} | ||||
|  | ||||
|     # TODO: The upload-s3 command should eventually be replaced with something | ||||
|     # like: "spack buildcache put <mirror> <spec>", when that subcommand is | ||||
|     # properly implemented. | ||||
|     spack -d upload-s3 spec --base-dir "${LOCAL_MIRROR}" --spec-yaml "${SPEC_YAML_PATH}" | ||||
|     check_error $? "spack upload-s3 spec" | ||||
| else | ||||
|     echo "spec ${JOB_SPEC_NAME} is already up to date on remote mirror, downloading it" | ||||
|  | ||||
|     # Configure remote mirror so we can download buildcache entry | ||||
|     spack mirror add remote_binary_mirror ${MIRROR_URL} | ||||
|  | ||||
|     # Now download it | ||||
|     spack -d buildcache download --spec-yaml "${SPEC_YAML_PATH}" --path "${BUILD_CACHE_DIR}/" --require-cdashid | ||||
|     check_error $? "spack buildcache download" | ||||
| fi | ||||
|  | ||||
| # The next step is to relate this job to the jobs it depends on | ||||
| if [ -f "${JOB_CDASH_ID_FILE}" ]; then | ||||
|     JOB_CDASH_BUILD_ID=$(<${JOB_CDASH_ID_FILE}) | ||||
|  | ||||
|     if [ "${JOB_CDASH_BUILD_ID}" == "NONE" ]; then | ||||
|         echo "ERROR: unable to read this jobs id from ${JOB_CDASH_ID_FILE}" | ||||
|         exit 1 | ||||
|     fi | ||||
|  | ||||
|     # Now get CDash ids for dependencies and "relate" each dependency build | ||||
|     # with this jobs build | ||||
|     for DEP_PKG_NAME in "${JOB_DEPS_PKG_NAMES[@]}"; do | ||||
|         echo "Getting cdash id for dependency --> ${DEP_PKG_NAME} <--" | ||||
|         DEP_SPEC_YAML_PATH="${SPEC_DIR}/${DEP_PKG_NAME}.yaml" | ||||
|         DEP_JOB_BUILDCACHE_NAME=`spack -d buildcache get-buildcache-name --spec-yaml "${DEP_SPEC_YAML_PATH}"` | ||||
|  | ||||
|         if [[ $? -eq 0 ]]; then | ||||
|             DEP_JOB_ID_FILE="${BUILD_CACHE_DIR}/${DEP_JOB_BUILDCACHE_NAME}.cdashid" | ||||
|             echo "DEP_JOB_ID_FILE path = ${DEP_JOB_ID_FILE}" | ||||
|  | ||||
|             if [ -f "${DEP_JOB_ID_FILE}" ]; then | ||||
|                 DEP_JOB_CDASH_BUILD_ID=$(<${DEP_JOB_ID_FILE}) | ||||
|                 echo "File ${DEP_JOB_ID_FILE} contained value ${DEP_JOB_CDASH_BUILD_ID}" | ||||
|                 echo "Relating builds -> ${JOB_SPEC_NAME} (buildid=${JOB_CDASH_BUILD_ID}) depends on ${DEP_PKG_NAME} (buildid=${DEP_JOB_CDASH_BUILD_ID})" | ||||
|                 relateBuildsPostBody="$(get_relate_builds_post_data "Spack" ${JOB_CDASH_BUILD_ID} ${DEP_JOB_CDASH_BUILD_ID})" | ||||
|                 relateBuildsResult=`curl "${DEP_JOB_RELATEBUILDS_URL}" -H "Content-Type: application/json" -H "Accept: application/json" -d "${relateBuildsPostBody}"` | ||||
|                 echo "Result of curl request: ${relateBuildsResult}" | ||||
|             else | ||||
|                 echo "ERROR: Did not find expected .cdashid file for dependency: ${DEP_JOB_ID_FILE}" | ||||
|                 exit 1 | ||||
|             fi | ||||
|         else | ||||
|             echo "ERROR: Unable to get buildcache entry name for ${DEP_SPEC_NAME}" | ||||
|             exit 1 | ||||
|         fi | ||||
|     done | ||||
| else | ||||
|     echo "ERROR: Did not find expected .cdashid file ${JOB_CDASH_ID_FILE}" | ||||
|     exit 1 | ||||
| fi | ||||
|  | ||||
| # Show the size of the buildcache and a list of what's in it, directly | ||||
| # in the gitlab log output | ||||
| ( | ||||
|     restore_io | ||||
|     du -sh ${BUILD_CACHE_DIR} | ||||
|     find ${BUILD_CACHE_DIR} -maxdepth 3 -type d -ls | ||||
| ) | ||||
|  | ||||
| echo "End of rebuild package script" | ||||
							
								
								
									
										212
									
								
								lib/spack/spack/cmd/upload_s3.py
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										212
									
								
								lib/spack/spack/cmd/upload_s3.py
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,212 @@ | ||||
| # Copyright 2013-2019 Lawrence Livermore National Security, LLC and other | ||||
| # Spack Project Developers. See the top-level COPYRIGHT file for details. | ||||
| # | ||||
| # SPDX-License-Identifier: (Apache-2.0 OR MIT) | ||||
|  | ||||
| # TODO: This will be merged into the buildcache command once | ||||
| # everything is working. | ||||
|  | ||||
| import os | ||||
| import re | ||||
| import sys | ||||
|  | ||||
| try: | ||||
|     import boto3 | ||||
|     import botocore | ||||
|     have_boto3_support = True | ||||
| except ImportError: | ||||
|     have_boto3_support = False | ||||
|  | ||||
| import llnl.util.tty as tty | ||||
|  | ||||
| from spack.error import SpackError | ||||
| import spack.tengine as template_engine | ||||
| from spack.spec import Spec | ||||
|  | ||||
|  | ||||
| import spack.binary_distribution as bindist | ||||
|  | ||||
|  | ||||
| description = "temporary command to upload buildcaches to 's3.spack.io'" | ||||
| section = "packaging" | ||||
| level = "long" | ||||
|  | ||||
|  | ||||
| def setup_parser(subparser): | ||||
|     setup_parser.parser = subparser | ||||
|     subparsers = subparser.add_subparsers(help='upload-s3 sub-commands') | ||||
|  | ||||
|     # sub-command to upload a built spec to s3 | ||||
|     spec = subparsers.add_parser('spec', help=upload_spec.__doc__) | ||||
|  | ||||
|     spec.add_argument('-s', '--spec', default=None, | ||||
|                       help='Spec to upload') | ||||
|  | ||||
|     spec.add_argument('-y', '--spec-yaml', default=None, | ||||
|                       help='Path to spec yaml file containing spec to upload') | ||||
|  | ||||
|     spec.add_argument('-b', '--base-dir', default=None, | ||||
|                       help='Path to root of buildcaches') | ||||
|  | ||||
|     spec.add_argument('-e', '--endpoint-url', | ||||
|                       default='https://s3.spack.io', help='URL of mirror') | ||||
|  | ||||
|     spec.set_defaults(func=upload_spec) | ||||
|  | ||||
|     # sub-command to update the index of a buildcache on s3 | ||||
|     index = subparsers.add_parser('index', help=update_index.__doc__) | ||||
|  | ||||
|     index.add_argument('-e', '--endpoint-url', | ||||
|         default='https://s3.spack.io', help='URL of mirror') | ||||
|  | ||||
|     index.set_defaults(func=update_index) | ||||
|  | ||||
|  | ||||
| def get_s3_session(endpoint_url): | ||||
|     if not have_boto3_support: | ||||
|         raise SpackError('boto3 module not available') | ||||
|  | ||||
|     session = boto3.Session() | ||||
|     s3 = session.resource('s3') | ||||
|  | ||||
|     bucket_names = [] | ||||
|     for bucket in s3.buckets.all(): | ||||
|         bucket_names.append(bucket.name) | ||||
|  | ||||
|     if len(bucket_names) > 1: | ||||
|         raise SpackError('More than one bucket associated with credentials') | ||||
|  | ||||
|     bucket_name = bucket_names[0] | ||||
|  | ||||
|     return s3, bucket_name | ||||
|  | ||||
|  | ||||
| def update_index(args): | ||||
|     """Update the index of an s3 buildcache""" | ||||
|     s3, bucket_name = get_s3_session(args.endpoint_url) | ||||
|  | ||||
|     bucket = s3.Bucket(bucket_name) | ||||
|     exists = True | ||||
|  | ||||
|     try: | ||||
|         s3.meta.client.head_bucket(Bucket=bucket_name) | ||||
|     except botocore.exceptions.ClientError as e: | ||||
|         # If a client error is thrown, then check that it was a 404 error. | ||||
|         # If it was a 404 error, then the bucket does not exist. | ||||
|         error_code = e.response['Error']['Code'] | ||||
|         if error_code == '404': | ||||
|             exists = False | ||||
|  | ||||
|     if not exists: | ||||
|         tty.error('S3 bucket "{0}" does not exist'.format(bucket_name)) | ||||
|         sys.exit(1) | ||||
|  | ||||
|     build_cache_dir = os.path.join( | ||||
|         'mirror', bindist.build_cache_relative_path()) | ||||
|  | ||||
|     spec_yaml_regex = re.compile('{0}/(.+\\.spec\\.yaml)$'.format( | ||||
|         build_cache_dir)) | ||||
|     spack_regex = re.compile('{0}/([^/]+)/.+\\.spack$'.format( | ||||
|         build_cache_dir)) | ||||
|  | ||||
|     top_level_keys = set() | ||||
|  | ||||
|     for key in bucket.objects.all(): | ||||
|         m = spec_yaml_regex.search(key.key) | ||||
|         if m: | ||||
|             top_level_keys.add(m.group(1)) | ||||
|             print(m.group(1)) | ||||
|             continue | ||||
|  | ||||
|         m = spack_regex.search(key.key) | ||||
|         if m: | ||||
|             top_level_keys.add(m.group(1)) | ||||
|             print(m.group(1)) | ||||
|             continue | ||||
|  | ||||
|     index_data = { | ||||
|         'top_level_keys': top_level_keys, | ||||
|     } | ||||
|  | ||||
|     env = template_engine.make_environment() | ||||
|     template_dir = 'misc' | ||||
|     index_template = os.path.join(template_dir, 'buildcache_index.html') | ||||
|     t = env.get_template(index_template) | ||||
|     contents = t.render(index_data) | ||||
|  | ||||
|     index_key = os.path.join(build_cache_dir, 'index.html') | ||||
|  | ||||
|     tty.debug('Generated index:') | ||||
|     tty.debug(contents) | ||||
|     tty.debug('Pushing it to {0} -> {1}'.format(bucket_name, index_key)) | ||||
|  | ||||
|     s3_obj = s3.Object(bucket_name, index_key) | ||||
|     s3_obj.put(Body=contents, ACL='public-read') | ||||
|  | ||||
|  | ||||
| def upload_spec(args): | ||||
|     """Upload a spec to s3 bucket""" | ||||
|     if not args.spec and not args.spec_yaml: | ||||
|         tty.error('Cannot upload spec without spec arg or path to spec yaml') | ||||
|         sys.exit(1) | ||||
|  | ||||
|     if not args.base_dir: | ||||
|         tty.error('No base directory for buildcache specified') | ||||
|         sys.exit(1) | ||||
|  | ||||
|     if args.spec: | ||||
|         try: | ||||
|             spec = Spec(args.spec) | ||||
|             spec.concretize() | ||||
|         except Exception: | ||||
|             tty.error('Unable to concrectize spec from string {0}'.format( | ||||
|                 args.spec)) | ||||
|             sys.exit(1) | ||||
|     else: | ||||
|         try: | ||||
|             with open(args.spec_yaml, 'r') as fd: | ||||
|                 spec = Spec.from_yaml(fd.read()) | ||||
|         except Exception: | ||||
|             tty.error('Unable to concrectize spec from yaml {0}'.format( | ||||
|                 args.spec_yaml)) | ||||
|             sys.exit(1) | ||||
|  | ||||
|     s3, bucket_name = get_s3_session(args.endpoint_url) | ||||
|  | ||||
|     build_cache_dir = bindist.build_cache_relative_path() | ||||
|  | ||||
|     tarball_key = os.path.join( | ||||
|         build_cache_dir, bindist.tarball_path_name(spec, '.spack')) | ||||
|     tarball_path = os.path.join(args.base_dir, tarball_key) | ||||
|  | ||||
|     specfile_key = os.path.join( | ||||
|         build_cache_dir, bindist.tarball_name(spec, '.spec.yaml')) | ||||
|     specfile_path = os.path.join(args.base_dir, specfile_key) | ||||
|  | ||||
|     cdashidfile_key = os.path.join( | ||||
|         build_cache_dir, bindist.tarball_name(spec, '.cdashid')) | ||||
|     cdashidfile_path = os.path.join(args.base_dir, cdashidfile_key) | ||||
|  | ||||
|     tty.msg('Uploading {0}'.format(tarball_key)) | ||||
|     s3.meta.client.upload_file( | ||||
|         tarball_path, bucket_name, | ||||
|         os.path.join('mirror', tarball_key), | ||||
|         ExtraArgs={'ACL': 'public-read'}) | ||||
|  | ||||
|     tty.msg('Uploading {0}'.format(specfile_key)) | ||||
|     s3.meta.client.upload_file( | ||||
|         specfile_path, bucket_name, | ||||
|         os.path.join('mirror', specfile_key), | ||||
|         ExtraArgs={'ACL': 'public-read'}) | ||||
|  | ||||
|     if os.path.exists(cdashidfile_path): | ||||
|         tty.msg('Uploading {0}'.format(cdashidfile_key)) | ||||
|         s3.meta.client.upload_file( | ||||
|             cdashidfile_path, bucket_name, | ||||
|             os.path.join('mirror', cdashidfile_key), | ||||
|             ExtraArgs={'ACL': 'public-read'}) | ||||
|  | ||||
|  | ||||
| def upload_s3(parser, args): | ||||
|     if args.func: | ||||
|         args.func(args) | ||||
		Reference in New Issue
	
	Block a user