ci: Don't raise an exception if we fail to relate CDash builds (#24299)
ci: Do not allow cdash-related errors to cause pipeline failure
This commit is contained in:
parent
b8512983d9
commit
0bbd71d561
@ -81,7 +81,8 @@ def _create_buildgroup(opener, headers, url, project, group_name, group_type):
|
|||||||
if response_code != 200 and response_code != 201:
|
if response_code != 200 and response_code != 201:
|
||||||
msg = 'Creating buildgroup failed (response code = {0}'.format(
|
msg = 'Creating buildgroup failed (response code = {0}'.format(
|
||||||
response_code)
|
response_code)
|
||||||
raise SpackError(msg)
|
tty.warn(msg)
|
||||||
|
return None
|
||||||
|
|
||||||
response_text = response.read()
|
response_text = response.read()
|
||||||
response_json = json.loads(response_text)
|
response_json = json.loads(response_text)
|
||||||
@ -110,7 +111,8 @@ def populate_buildgroup(job_names, group_name, project, site,
|
|||||||
if not parent_group_id or not group_id:
|
if not parent_group_id or not group_id:
|
||||||
msg = 'Failed to create or retrieve buildgroups for {0}'.format(
|
msg = 'Failed to create or retrieve buildgroups for {0}'.format(
|
||||||
group_name)
|
group_name)
|
||||||
raise SpackError(msg)
|
tty.warn(msg)
|
||||||
|
return
|
||||||
|
|
||||||
data = {
|
data = {
|
||||||
'project': project,
|
'project': project,
|
||||||
@ -133,7 +135,7 @@ def populate_buildgroup(job_names, group_name, project, site,
|
|||||||
if response_code != 200:
|
if response_code != 200:
|
||||||
msg = 'Error response code ({0}) in populate_buildgroup'.format(
|
msg = 'Error response code ({0}) in populate_buildgroup'.format(
|
||||||
response_code)
|
response_code)
|
||||||
raise SpackError(msg)
|
tty.warn(msg)
|
||||||
|
|
||||||
|
|
||||||
def is_main_phase(phase_name):
|
def is_main_phase(phase_name):
|
||||||
@ -1259,7 +1261,8 @@ def register_cdash_build(build_name, base_url, project, site, track):
|
|||||||
|
|
||||||
if response_code != 200 and response_code != 201:
|
if response_code != 200 and response_code != 201:
|
||||||
msg = 'Adding build failed (response code = {0}'.format(response_code)
|
msg = 'Adding build failed (response code = {0}'.format(response_code)
|
||||||
raise SpackError(msg)
|
tty.warn(msg)
|
||||||
|
return (None, None)
|
||||||
|
|
||||||
response_text = response.read()
|
response_text = response.read()
|
||||||
response_json = json.loads(response_text)
|
response_json = json.loads(response_text)
|
||||||
@ -1296,8 +1299,9 @@ def relate_cdash_builds(spec_map, cdash_base_url, job_build_id, cdash_project,
|
|||||||
tty.debug('Did not find cdashid for {0} on {1}'.format(
|
tty.debug('Did not find cdashid for {0} on {1}'.format(
|
||||||
dep_pkg_name, url))
|
dep_pkg_name, url))
|
||||||
else:
|
else:
|
||||||
raise SpackError('Did not find cdashid for {0} anywhere'.format(
|
tty.warn('Did not find cdashid for {0} anywhere'.format(
|
||||||
dep_pkg_name))
|
dep_pkg_name))
|
||||||
|
return
|
||||||
|
|
||||||
payload = {
|
payload = {
|
||||||
"project": cdash_project,
|
"project": cdash_project,
|
||||||
@ -1318,7 +1322,8 @@ def relate_cdash_builds(spec_map, cdash_base_url, job_build_id, cdash_project,
|
|||||||
if response_code != 200 and response_code != 201:
|
if response_code != 200 and response_code != 201:
|
||||||
msg = 'Relate builds ({0} -> {1}) failed (resp code = {2})'.format(
|
msg = 'Relate builds ({0} -> {1}) failed (resp code = {2})'.format(
|
||||||
job_build_id, dep_build_id, response_code)
|
job_build_id, dep_build_id, response_code)
|
||||||
raise SpackError(msg)
|
tty.warn(msg)
|
||||||
|
return
|
||||||
|
|
||||||
response_text = response.read()
|
response_text = response.read()
|
||||||
tty.debug('Relate builds response: {0}'.format(response_text))
|
tty.debug('Relate builds response: {0}'.format(response_text))
|
||||||
@ -1341,7 +1346,16 @@ def write_cdashid_to_mirror(cdashid, spec, mirror_url):
|
|||||||
tty.debug('pushing cdashid to url')
|
tty.debug('pushing cdashid to url')
|
||||||
tty.debug(' local file path: {0}'.format(local_cdash_path))
|
tty.debug(' local file path: {0}'.format(local_cdash_path))
|
||||||
tty.debug(' remote url: {0}'.format(remote_url))
|
tty.debug(' remote url: {0}'.format(remote_url))
|
||||||
web_util.push_to_url(local_cdash_path, remote_url)
|
|
||||||
|
try:
|
||||||
|
web_util.push_to_url(local_cdash_path, remote_url)
|
||||||
|
except Exception as inst:
|
||||||
|
# No matter what went wrong here, don't allow the pipeline to fail
|
||||||
|
# just because there was an issue storing the cdashid on the mirror
|
||||||
|
msg = 'Failed to write cdashid {0} to mirror {1}'.format(
|
||||||
|
cdashid, mirror_url)
|
||||||
|
tty.warn(inst)
|
||||||
|
tty.warn(msg)
|
||||||
|
|
||||||
|
|
||||||
def read_cdashid_from_mirror(spec, mirror_url):
|
def read_cdashid_from_mirror(spec, mirror_url):
|
||||||
@ -1359,40 +1373,34 @@ def read_cdashid_from_mirror(spec, mirror_url):
|
|||||||
return int(contents)
|
return int(contents)
|
||||||
|
|
||||||
|
|
||||||
def push_mirror_contents(env, spec, yaml_path, mirror_url, build_id,
|
def push_mirror_contents(env, spec, yaml_path, mirror_url, sign_binaries):
|
||||||
sign_binaries):
|
try:
|
||||||
if mirror_url:
|
unsigned = not sign_binaries
|
||||||
try:
|
tty.debug('Creating buildcache ({0})'.format(
|
||||||
unsigned = not sign_binaries
|
'unsigned' if unsigned else 'signed'))
|
||||||
tty.debug('Creating buildcache ({0})'.format(
|
spack.cmd.buildcache._createtarball(
|
||||||
'unsigned' if unsigned else 'signed'))
|
env, spec_yaml=yaml_path, add_deps=False,
|
||||||
spack.cmd.buildcache._createtarball(
|
output_location=mirror_url, force=True, allow_root=True,
|
||||||
env, spec_yaml=yaml_path, add_deps=False,
|
unsigned=unsigned)
|
||||||
output_location=mirror_url, force=True, allow_root=True,
|
except Exception as inst:
|
||||||
unsigned=unsigned)
|
# If the mirror we're pushing to is on S3 and there's some
|
||||||
if build_id:
|
# permissions problem, for example, we can't just target
|
||||||
tty.debug('Writing cdashid ({0}) to remote mirror: {1}'.format(
|
# that exception type here, since users of the
|
||||||
build_id, mirror_url))
|
# `spack ci rebuild' may not need or want any dependency
|
||||||
write_cdashid_to_mirror(build_id, spec, mirror_url)
|
# on boto3. So we use the first non-boto exception type
|
||||||
except Exception as inst:
|
# in the heirarchy:
|
||||||
# If the mirror we're pushing to is on S3 and there's some
|
# boto3.exceptions.S3UploadFailedError
|
||||||
# permissions problem, for example, we can't just target
|
# boto3.exceptions.Boto3Error
|
||||||
# that exception type here, since users of the
|
# Exception
|
||||||
# `spack ci rebuild' may not need or want any dependency
|
# BaseException
|
||||||
# on boto3. So we use the first non-boto exception type
|
# object
|
||||||
# in the heirarchy:
|
err_msg = 'Error msg: {0}'.format(inst)
|
||||||
# boto3.exceptions.S3UploadFailedError
|
if 'Access Denied' in err_msg:
|
||||||
# boto3.exceptions.Boto3Error
|
tty.msg('Permission problem writing to {0}'.format(
|
||||||
# Exception
|
mirror_url))
|
||||||
# BaseException
|
tty.msg(err_msg)
|
||||||
# object
|
else:
|
||||||
err_msg = 'Error msg: {0}'.format(inst)
|
raise inst
|
||||||
if 'Access Denied' in err_msg:
|
|
||||||
tty.msg('Permission problem writing to {0}'.format(
|
|
||||||
mirror_url))
|
|
||||||
tty.msg(err_msg)
|
|
||||||
else:
|
|
||||||
raise inst
|
|
||||||
|
|
||||||
|
|
||||||
def copy_stage_logs_to_artifacts(job_spec, job_log_dir):
|
def copy_stage_logs_to_artifacts(job_spec, job_log_dir):
|
||||||
|
@ -230,7 +230,6 @@ def ci_rebuild(args):
|
|||||||
eq_idx = proj_enc.find('=') + 1
|
eq_idx = proj_enc.find('=') + 1
|
||||||
cdash_project_enc = proj_enc[eq_idx:]
|
cdash_project_enc = proj_enc[eq_idx:]
|
||||||
cdash_site = ci_cdash['site']
|
cdash_site = ci_cdash['site']
|
||||||
cdash_id_path = os.path.join(repro_dir, 'cdash_id.txt')
|
|
||||||
tty.debug('cdash_base_url = {0}'.format(cdash_base_url))
|
tty.debug('cdash_base_url = {0}'.format(cdash_base_url))
|
||||||
tty.debug('cdash_project = {0}'.format(cdash_project))
|
tty.debug('cdash_project = {0}'.format(cdash_project))
|
||||||
tty.debug('cdash_project_enc = {0}'.format(cdash_project_enc))
|
tty.debug('cdash_project_enc = {0}'.format(cdash_project_enc))
|
||||||
@ -398,7 +397,7 @@ def ci_rebuild(args):
|
|||||||
job_spec_pkg_name, matching_mirror))
|
job_spec_pkg_name, matching_mirror))
|
||||||
tty.debug('Downloading to {0}'.format(build_cache_dir))
|
tty.debug('Downloading to {0}'.format(build_cache_dir))
|
||||||
buildcache.download_buildcache_files(
|
buildcache.download_buildcache_files(
|
||||||
job_spec, build_cache_dir, True, matching_mirror)
|
job_spec, build_cache_dir, False, matching_mirror)
|
||||||
|
|
||||||
# Now we are done and successful
|
# Now we are done and successful
|
||||||
sys.exit(0)
|
sys.exit(0)
|
||||||
@ -433,24 +432,21 @@ def ci_rebuild(args):
|
|||||||
cdash_build_name, cdash_base_url, cdash_project,
|
cdash_build_name, cdash_base_url, cdash_project,
|
||||||
cdash_site, job_spec_buildgroup)
|
cdash_site, job_spec_buildgroup)
|
||||||
|
|
||||||
cdash_upload_url = '{0}/submit.php?project={1}'.format(
|
if cdash_build_id is not None:
|
||||||
cdash_base_url, cdash_project_enc)
|
cdash_upload_url = '{0}/submit.php?project={1}'.format(
|
||||||
|
cdash_base_url, cdash_project_enc)
|
||||||
|
|
||||||
install_args.extend([
|
install_args.extend([
|
||||||
'--cdash-upload-url', cdash_upload_url,
|
'--cdash-upload-url', cdash_upload_url,
|
||||||
'--cdash-build', cdash_build_name,
|
'--cdash-build', cdash_build_name,
|
||||||
'--cdash-site', cdash_site,
|
'--cdash-site', cdash_site,
|
||||||
'--cdash-buildstamp', cdash_build_stamp,
|
'--cdash-buildstamp', cdash_build_stamp,
|
||||||
])
|
])
|
||||||
|
|
||||||
tty.debug('CDash: Relating build with dependency builds')
|
tty.debug('CDash: Relating build with dependency builds')
|
||||||
spack_ci.relate_cdash_builds(
|
spack_ci.relate_cdash_builds(
|
||||||
spec_map, cdash_base_url, cdash_build_id, cdash_project,
|
spec_map, cdash_base_url, cdash_build_id, cdash_project,
|
||||||
[pipeline_mirror_url, pr_mirror_url, remote_mirror_url])
|
[pipeline_mirror_url, pr_mirror_url, remote_mirror_url])
|
||||||
|
|
||||||
# store the cdash build id on disk for later
|
|
||||||
with open(cdash_id_path, 'w') as fd:
|
|
||||||
fd.write(cdash_build_id)
|
|
||||||
|
|
||||||
# A compiler action of 'FIND_ANY' means we are building a bootstrap
|
# A compiler action of 'FIND_ANY' means we are building a bootstrap
|
||||||
# compiler or one of its deps.
|
# compiler or one of its deps.
|
||||||
@ -546,17 +542,31 @@ def ci_rebuild(args):
|
|||||||
|
|
||||||
# Create buildcache in either the main remote mirror, or in the
|
# Create buildcache in either the main remote mirror, or in the
|
||||||
# per-PR mirror, if this is a PR pipeline
|
# per-PR mirror, if this is a PR pipeline
|
||||||
spack_ci.push_mirror_contents(
|
if buildcache_mirror_url:
|
||||||
env, job_spec, job_spec_yaml_path, buildcache_mirror_url,
|
spack_ci.push_mirror_contents(
|
||||||
cdash_build_id, sign_binaries)
|
env, job_spec, job_spec_yaml_path, buildcache_mirror_url,
|
||||||
|
sign_binaries)
|
||||||
|
|
||||||
|
if cdash_build_id:
|
||||||
|
tty.debug('Writing cdashid ({0}) to remote mirror: {1}'.format(
|
||||||
|
cdash_build_id, buildcache_mirror_url))
|
||||||
|
spack_ci.write_cdashid_to_mirror(
|
||||||
|
cdash_build_id, job_spec, buildcache_mirror_url)
|
||||||
|
|
||||||
# Create another copy of that buildcache in the per-pipeline
|
# Create another copy of that buildcache in the per-pipeline
|
||||||
# temporary storage mirror (this is only done if either
|
# temporary storage mirror (this is only done if either
|
||||||
# artifacts buildcache is enabled or a temporary storage url
|
# artifacts buildcache is enabled or a temporary storage url
|
||||||
# prefix is set)
|
# prefix is set)
|
||||||
spack_ci.push_mirror_contents(
|
if pipeline_mirror_url:
|
||||||
env, job_spec, job_spec_yaml_path, pipeline_mirror_url,
|
spack_ci.push_mirror_contents(
|
||||||
cdash_build_id, sign_binaries)
|
env, job_spec, job_spec_yaml_path, pipeline_mirror_url,
|
||||||
|
sign_binaries)
|
||||||
|
|
||||||
|
if cdash_build_id:
|
||||||
|
tty.debug('Writing cdashid ({0}) to remote mirror: {1}'.format(
|
||||||
|
cdash_build_id, pipeline_mirror_url))
|
||||||
|
spack_ci.write_cdashid_to_mirror(
|
||||||
|
cdash_build_id, job_spec, pipeline_mirror_url)
|
||||||
else:
|
else:
|
||||||
tty.debug('spack install exited non-zero, will not create buildcache')
|
tty.debug('spack install exited non-zero, will not create buildcache')
|
||||||
|
|
||||||
|
@ -179,7 +179,7 @@ def test_register_cdash_build(monkeypatch):
|
|||||||
|
|
||||||
|
|
||||||
def test_relate_cdash_builds(config, mutable_mock_env_path, mock_packages,
|
def test_relate_cdash_builds(config, mutable_mock_env_path, mock_packages,
|
||||||
monkeypatch):
|
monkeypatch, capfd):
|
||||||
e = ev.create('test1')
|
e = ev.create('test1')
|
||||||
e.add('dyninst')
|
e.add('dyninst')
|
||||||
e.concretize()
|
e.concretize()
|
||||||
@ -228,9 +228,11 @@ def test_relate_cdash_builds(config, mutable_mock_env_path, mock_packages,
|
|||||||
}
|
}
|
||||||
|
|
||||||
fake_responder._resp_code = 400
|
fake_responder._resp_code = 400
|
||||||
with pytest.raises(spack.error.SpackError):
|
ci.relate_cdash_builds(spec_map, cdash_api_url, job_build_id,
|
||||||
ci.relate_cdash_builds(spec_map, cdash_api_url, job_build_id,
|
cdash_project, [cdashids_mirror_url])
|
||||||
cdash_project, [cdashids_mirror_url])
|
out, err = capfd.readouterr()
|
||||||
|
assert('Warning: Relate builds' in err)
|
||||||
|
assert('failed' in err)
|
||||||
|
|
||||||
dep_cdash_ids = {}
|
dep_cdash_ids = {}
|
||||||
|
|
||||||
|
@ -894,8 +894,9 @@ def fake_dl_method(spec, dest, require_cdashid, m_url=None):
|
|||||||
|
|
||||||
@pytest.mark.disable_clean_stage_check
|
@pytest.mark.disable_clean_stage_check
|
||||||
def test_push_mirror_contents(tmpdir, mutable_mock_env_path, env_deactivate,
|
def test_push_mirror_contents(tmpdir, mutable_mock_env_path, env_deactivate,
|
||||||
install_mockery, mock_packages, mock_fetch,
|
install_mockery_mutable_config, mock_packages,
|
||||||
mock_stage, mock_gnupghome, project_dir_env):
|
mock_fetch, mock_stage, mock_gnupghome,
|
||||||
|
project_dir_env):
|
||||||
project_dir_env(tmpdir.strpath)
|
project_dir_env(tmpdir.strpath)
|
||||||
working_dir = tmpdir.join('working_dir')
|
working_dir = tmpdir.join('working_dir')
|
||||||
|
|
||||||
@ -951,7 +952,9 @@ def test_push_mirror_contents(tmpdir, mutable_mock_env_path, env_deactivate,
|
|||||||
|
|
||||||
# env, spec, yaml_path, mirror_url, build_id, sign_binaries
|
# env, spec, yaml_path, mirror_url, build_id, sign_binaries
|
||||||
ci.push_mirror_contents(
|
ci.push_mirror_contents(
|
||||||
env, concrete_spec, yaml_path, mirror_url, '42', True)
|
env, concrete_spec, yaml_path, mirror_url, True)
|
||||||
|
|
||||||
|
ci.write_cdashid_to_mirror('42', concrete_spec, mirror_url)
|
||||||
|
|
||||||
buildcache_path = os.path.join(mirror_dir.strpath, 'build_cache')
|
buildcache_path = os.path.join(mirror_dir.strpath, 'build_cache')
|
||||||
|
|
||||||
@ -1054,7 +1057,7 @@ def faked(env, spec_yaml=None, packages=None, add_spec=True,
|
|||||||
monkeypatch.setattr(buildcache, '_createtarball', faked)
|
monkeypatch.setattr(buildcache, '_createtarball', faked)
|
||||||
|
|
||||||
url = 'fakejunk'
|
url = 'fakejunk'
|
||||||
ci.push_mirror_contents(None, None, None, url, None, None)
|
ci.push_mirror_contents(None, None, None, url, None)
|
||||||
|
|
||||||
captured = capsys.readouterr()
|
captured = capsys.readouterr()
|
||||||
std_out = captured[0]
|
std_out = captured[0]
|
||||||
|
Loading…
Reference in New Issue
Block a user