
Currently as part of installing a package, we lock a prefix, check if it exists, and create it if not; the logic for creating the prefix included a check for the existence of that prefix (and raised an exception if it did), which was redundant. This also includes removal of tests which were not verifying anything (they pass with or without the modifications in this PR).
637 lines
24 KiB
Python
637 lines
24 KiB
Python
# Copyright 2013-2021 Lawrence Livermore National Security, LLC and other
|
|
# Spack Project Developers. See the top-level COPYRIGHT file for details.
|
|
#
|
|
# SPDX-License-Identifier: (Apache-2.0 OR MIT)
|
|
|
|
import errno
|
|
import glob
|
|
import os
|
|
import shutil
|
|
import tempfile
|
|
from contextlib import contextmanager
|
|
|
|
import ruamel.yaml as yaml
|
|
|
|
import llnl.util.filesystem as fs
|
|
import llnl.util.tty as tty
|
|
|
|
import spack.config
|
|
import spack.hash_types as ht
|
|
import spack.spec
|
|
import spack.util.spack_json as sjson
|
|
from spack.error import SpackError
|
|
|
|
default_projections = {'all': ('{architecture}/'
|
|
'{compiler.name}-{compiler.version}/'
|
|
'{name}-{version}-{hash}')}
|
|
|
|
|
|
def _check_concrete(spec):
|
|
"""If the spec is not concrete, raise a ValueError"""
|
|
if not spec.concrete:
|
|
raise ValueError('Specs passed to a DirectoryLayout must be concrete!')
|
|
|
|
|
|
class DirectoryLayout(object):
|
|
"""A directory layout is used to associate unique paths with specs.
|
|
Different installations are going to want different layouts for their
|
|
install, and they can use this to customize the nesting structure of
|
|
spack installs. The default layout is:
|
|
|
|
* <install root>/
|
|
|
|
* <platform-os-target>/
|
|
|
|
* <compiler>-<compiler version>/
|
|
|
|
* <name>-<version>-<hash>
|
|
|
|
The hash here is a SHA-1 hash for the full DAG plus the build
|
|
spec.
|
|
|
|
The installation directory projections can be modified with the
|
|
projections argument.
|
|
"""
|
|
|
|
def __init__(self, root, **kwargs):
|
|
self.root = root
|
|
self.check_upstream = True
|
|
projections = kwargs.get('projections') or default_projections
|
|
self.projections = dict((key, projection.lower())
|
|
for key, projection in projections.items())
|
|
|
|
# apply hash length as appropriate
|
|
self.hash_length = kwargs.get('hash_length', None)
|
|
if self.hash_length is not None:
|
|
for when_spec, projection in self.projections.items():
|
|
if '{hash}' not in projection:
|
|
if '{hash' in projection:
|
|
raise InvalidDirectoryLayoutParametersError(
|
|
"Conflicting options for installation layout hash"
|
|
" length")
|
|
else:
|
|
raise InvalidDirectoryLayoutParametersError(
|
|
"Cannot specify hash length when the hash is not"
|
|
" part of all install_tree projections")
|
|
self.projections[when_spec] = projection.replace(
|
|
"{hash}", "{hash:%d}" % self.hash_length)
|
|
|
|
# If any of these paths change, downstream databases may not be able to
|
|
# locate files in older upstream databases
|
|
self.metadata_dir = '.spack'
|
|
self.deprecated_dir = 'deprecated'
|
|
self.spec_file_name = 'spec.json'
|
|
# Use for checking yaml and deprecated types
|
|
self._spec_file_name_yaml = 'spec.yaml'
|
|
self.extension_file_name = 'extensions.yaml'
|
|
self.packages_dir = 'repos' # archive of package.py files
|
|
self.manifest_file_name = 'install_manifest.json'
|
|
|
|
@property
|
|
def hidden_file_paths(self):
|
|
return (self.metadata_dir,)
|
|
|
|
def relative_path_for_spec(self, spec):
|
|
_check_concrete(spec)
|
|
|
|
projection = spack.projections.get_projection(self.projections, spec)
|
|
path = spec.format(projection)
|
|
return path
|
|
|
|
def write_spec(self, spec, path):
|
|
"""Write a spec out to a file."""
|
|
_check_concrete(spec)
|
|
with open(path, 'w') as f:
|
|
# The hash the the projection is the DAG hash but we write out the
|
|
# full provenance by full hash so it's availabe if we want it later
|
|
# extension = os.path.splitext(path)[-1].lower()
|
|
# if 'json' in extension:
|
|
spec.to_json(f, hash=ht.full_hash)
|
|
# elif 'yaml' in extension:
|
|
# spec.to_yaml(f, hash=ht.full_hash)
|
|
|
|
def write_host_environment(self, spec):
|
|
"""The host environment is a json file with os, kernel, and spack
|
|
versioning. We use it in the case that an analysis later needs to
|
|
easily access this information.
|
|
"""
|
|
from spack.util.environment import get_host_environment_metadata
|
|
env_file = self.env_metadata_path(spec)
|
|
environ = get_host_environment_metadata()
|
|
with open(env_file, 'w') as fd:
|
|
sjson.dump(environ, fd)
|
|
|
|
def read_spec(self, path):
|
|
"""Read the contents of a file and parse them as a spec"""
|
|
try:
|
|
with open(path) as f:
|
|
extension = os.path.splitext(path)[-1].lower()
|
|
if extension == '.json':
|
|
spec = spack.spec.Spec.from_json(f)
|
|
elif extension == '.yaml':
|
|
# Too late for conversion; spec_file_path() already called.
|
|
spec = spack.spec.Spec.from_yaml(f)
|
|
else:
|
|
raise SpecReadError('Did not recognize spec file extension:'
|
|
' {0}'.format(extension))
|
|
except Exception as e:
|
|
if spack.config.get('config:debug'):
|
|
raise
|
|
raise SpecReadError(
|
|
'Unable to read file: %s' % path, 'Cause: ' + str(e))
|
|
|
|
# Specs read from actual installations are always concrete
|
|
spec._mark_concrete()
|
|
return spec
|
|
|
|
def spec_file_path(self, spec):
|
|
"""Gets full path to spec file"""
|
|
_check_concrete(spec)
|
|
# Attempts to convert to JSON if possible.
|
|
# Otherwise just returns the YAML.
|
|
yaml_path = os.path.join(
|
|
self.metadata_path(spec), self._spec_file_name_yaml)
|
|
json_path = os.path.join(self.metadata_path(spec), self.spec_file_name)
|
|
if os.path.exists(yaml_path) and fs.can_write_to_dir(yaml_path):
|
|
self.write_spec(spec, json_path)
|
|
try:
|
|
os.remove(yaml_path)
|
|
except OSError as err:
|
|
tty.debug('Could not remove deprecated {0}'.format(yaml_path))
|
|
tty.debug(err)
|
|
elif os.path.exists(yaml_path):
|
|
return yaml_path
|
|
return json_path
|
|
|
|
def deprecated_file_path(self, deprecated_spec, deprecator_spec=None):
|
|
"""Gets full path to spec file for deprecated spec
|
|
|
|
If the deprecator_spec is provided, use that. Otherwise, assume
|
|
deprecated_spec is already deprecated and its prefix links to the
|
|
prefix of its deprecator."""
|
|
_check_concrete(deprecated_spec)
|
|
if deprecator_spec:
|
|
_check_concrete(deprecator_spec)
|
|
|
|
# If deprecator spec is None, assume deprecated_spec already deprecated
|
|
# and use its link to find the file.
|
|
base_dir = self.path_for_spec(
|
|
deprecator_spec
|
|
) if deprecator_spec else os.readlink(deprecated_spec.prefix)
|
|
|
|
yaml_path = os.path.join(base_dir, self.metadata_dir,
|
|
self.deprecated_dir, deprecated_spec.dag_hash()
|
|
+ '_' + self._spec_file_name_yaml)
|
|
|
|
json_path = os.path.join(base_dir, self.metadata_dir,
|
|
self.deprecated_dir, deprecated_spec.dag_hash()
|
|
+ '_' + self.spec_file_name)
|
|
|
|
if (os.path.exists(yaml_path) and fs.can_write_to_dir(yaml_path)):
|
|
self.write_spec(deprecated_spec, json_path)
|
|
try:
|
|
os.remove(yaml_path)
|
|
except (IOError, OSError) as err:
|
|
tty.debug('Could not remove deprecated {0}'.format(yaml_path))
|
|
tty.debug(err)
|
|
elif os.path.exists(yaml_path):
|
|
return yaml_path
|
|
|
|
return json_path
|
|
|
|
@contextmanager
|
|
def disable_upstream_check(self):
|
|
self.check_upstream = False
|
|
yield
|
|
self.check_upstream = True
|
|
|
|
def metadata_path(self, spec):
|
|
return os.path.join(spec.prefix, self.metadata_dir)
|
|
|
|
def env_metadata_path(self, spec):
|
|
return os.path.join(self.metadata_path(spec), "install_environment.json")
|
|
|
|
def build_packages_path(self, spec):
|
|
return os.path.join(self.metadata_path(spec), self.packages_dir)
|
|
|
|
def create_install_directory(self, spec):
|
|
_check_concrete(spec)
|
|
|
|
# Create install directory with properly configured permissions
|
|
# Cannot import at top of file
|
|
from spack.package_prefs import get_package_dir_permissions, get_package_group
|
|
|
|
# Each package folder can have its own specific permissions, while
|
|
# intermediate folders (arch/compiler) are set with access permissions
|
|
# equivalent to the root permissions of the layout.
|
|
group = get_package_group(spec)
|
|
perms = get_package_dir_permissions(spec)
|
|
|
|
fs.mkdirp(spec.prefix, mode=perms, group=group, default_perms='parents')
|
|
fs.mkdirp(self.metadata_path(spec), mode=perms, group=group) # in prefix
|
|
|
|
self.write_spec(spec, self.spec_file_path(spec))
|
|
|
|
def check_installed(self, spec):
|
|
_check_concrete(spec)
|
|
path = self.path_for_spec(spec)
|
|
spec_file_path = self.spec_file_path(spec)
|
|
|
|
if not os.path.isdir(path):
|
|
return None
|
|
|
|
if not os.path.isfile(spec_file_path):
|
|
raise InconsistentInstallDirectoryError(
|
|
'Install prefix exists but contains no spec.json:',
|
|
" " + path)
|
|
|
|
installed_spec = self.read_spec(spec_file_path)
|
|
if installed_spec == spec:
|
|
return path
|
|
|
|
# DAG hashes currently do not include build dependencies.
|
|
#
|
|
# TODO: remove this when we do better concretization and don't
|
|
# ignore build-only deps in hashes.
|
|
elif (installed_spec.copy(deps=('link', 'run')) ==
|
|
spec.copy(deps=('link', 'run'))):
|
|
# The directory layout prefix is based on the dag hash, so among
|
|
# specs with differing full-hash but matching dag-hash, only one
|
|
# may be installed. This means for example that for two instances
|
|
# that differ only in CMake version used to build, only one will
|
|
# be installed.
|
|
return path
|
|
|
|
if spec.dag_hash() == installed_spec.dag_hash():
|
|
raise SpecHashCollisionError(spec, installed_spec)
|
|
else:
|
|
raise InconsistentInstallDirectoryError(
|
|
'Spec file in %s does not match hash!' % spec_file_path)
|
|
|
|
def all_specs(self):
|
|
if not os.path.isdir(self.root):
|
|
return []
|
|
|
|
specs = []
|
|
for _, path_scheme in self.projections.items():
|
|
path_elems = ["*"] * len(path_scheme.split(os.sep))
|
|
# NOTE: Does not validate filename extension; should happen later
|
|
path_elems += [self.metadata_dir, 'spec.json']
|
|
pattern = os.path.join(self.root, *path_elems)
|
|
spec_files = glob.glob(pattern)
|
|
if not spec_files: # we're probably looking at legacy yaml...
|
|
path_elems += [self.metadata_dir, 'spec.yaml']
|
|
pattern = os.path.join(self.root, *path_elems)
|
|
spec_files = glob.glob(pattern)
|
|
specs.extend([self.read_spec(s) for s in spec_files])
|
|
return specs
|
|
|
|
def all_deprecated_specs(self):
|
|
if not os.path.isdir(self.root):
|
|
return []
|
|
|
|
deprecated_specs = set()
|
|
for _, path_scheme in self.projections.items():
|
|
path_elems = ["*"] * len(path_scheme.split(os.sep))
|
|
# NOTE: Does not validate filename extension; should happen later
|
|
path_elems += [self.metadata_dir, self.deprecated_dir,
|
|
'*_spec.*'] # + self.spec_file_name]
|
|
pattern = os.path.join(self.root, *path_elems)
|
|
spec_files = glob.glob(pattern)
|
|
get_depr_spec_file = lambda x: os.path.join(
|
|
os.path.dirname(os.path.dirname(x)), self.spec_file_name)
|
|
deprecated_specs |= set((self.read_spec(s),
|
|
self.read_spec(get_depr_spec_file(s)))
|
|
for s in spec_files)
|
|
return deprecated_specs
|
|
|
|
def specs_by_hash(self):
|
|
by_hash = {}
|
|
for spec in self.all_specs():
|
|
by_hash[spec.dag_hash()] = spec
|
|
return by_hash
|
|
|
|
def path_for_spec(self, spec):
|
|
"""Return absolute path from the root to a directory for the spec."""
|
|
_check_concrete(spec)
|
|
|
|
if spec.external:
|
|
return spec.external_path
|
|
if self.check_upstream:
|
|
upstream, record = spack.store.db.query_by_spec_hash(
|
|
spec.dag_hash())
|
|
if upstream:
|
|
raise SpackError(
|
|
"Internal error: attempted to call path_for_spec on"
|
|
" upstream-installed package.")
|
|
|
|
path = self.relative_path_for_spec(spec)
|
|
assert(not path.startswith(self.root))
|
|
return os.path.join(self.root, path)
|
|
|
|
def remove_install_directory(self, spec, deprecated=False):
|
|
"""Removes a prefix and any empty parent directories from the root.
|
|
Raised RemoveFailedError if something goes wrong.
|
|
"""
|
|
path = self.path_for_spec(spec)
|
|
assert(path.startswith(self.root))
|
|
|
|
if deprecated:
|
|
if os.path.exists(path):
|
|
try:
|
|
metapath = self.deprecated_file_path(spec)
|
|
os.unlink(path)
|
|
os.remove(metapath)
|
|
except OSError as e:
|
|
raise RemoveFailedError(spec, path, e)
|
|
|
|
elif os.path.exists(path):
|
|
try:
|
|
shutil.rmtree(path)
|
|
except OSError as e:
|
|
raise RemoveFailedError(spec, path, e)
|
|
|
|
path = os.path.dirname(path)
|
|
while path != self.root:
|
|
if os.path.isdir(path):
|
|
try:
|
|
os.rmdir(path)
|
|
except OSError as e:
|
|
if e.errno == errno.ENOENT:
|
|
# already deleted, continue with parent
|
|
pass
|
|
elif e.errno == errno.ENOTEMPTY:
|
|
# directory wasn't empty, done
|
|
return
|
|
else:
|
|
raise e
|
|
path = os.path.dirname(path)
|
|
|
|
|
|
class ExtensionsLayout(object):
|
|
"""A directory layout is used to associate unique paths with specs for
|
|
package extensions.
|
|
Keeps track of which extensions are activated for what package.
|
|
Depending on the use case, this can mean globally activated extensions
|
|
directly in the installation folder - or extensions activated in
|
|
filesystem views.
|
|
"""
|
|
def __init__(self, view, **kwargs):
|
|
self.view = view
|
|
|
|
def add_extension(self, spec, ext_spec):
|
|
"""Add to the list of currently installed extensions."""
|
|
raise NotImplementedError()
|
|
|
|
def check_activated(self, spec, ext_spec):
|
|
"""Ensure that ext_spec can be removed from spec.
|
|
|
|
If not, raise NoSuchExtensionError.
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
def check_extension_conflict(self, spec, ext_spec):
|
|
"""Ensure that ext_spec can be activated in spec.
|
|
|
|
If not, raise ExtensionAlreadyInstalledError or
|
|
ExtensionConflictError.
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
def extension_map(self, spec):
|
|
"""Get a dict of currently installed extension packages for a spec.
|
|
|
|
Dict maps { name : extension_spec }
|
|
Modifying dict does not affect internals of this layout.
|
|
"""
|
|
raise NotImplementedError()
|
|
|
|
def extendee_target_directory(self, extendee):
|
|
"""Specify to which full path extendee should link all files
|
|
from extensions."""
|
|
raise NotImplementedError
|
|
|
|
def remove_extension(self, spec, ext_spec):
|
|
"""Remove from the list of currently installed extensions."""
|
|
raise NotImplementedError()
|
|
|
|
|
|
class YamlViewExtensionsLayout(ExtensionsLayout):
|
|
"""Maintain extensions within a view.
|
|
"""
|
|
def __init__(self, view, layout):
|
|
"""layout is the corresponding YamlDirectoryLayout object for which
|
|
we implement extensions.
|
|
"""
|
|
super(YamlViewExtensionsLayout, self).__init__(view)
|
|
self.layout = layout
|
|
self.extension_file_name = 'extensions.yaml'
|
|
|
|
# Cache of already written/read extension maps.
|
|
self._extension_maps = {}
|
|
|
|
def add_extension(self, spec, ext_spec):
|
|
_check_concrete(spec)
|
|
_check_concrete(ext_spec)
|
|
|
|
# Check whether it's already installed or if it's a conflict.
|
|
exts = self._extension_map(spec)
|
|
self.check_extension_conflict(spec, ext_spec)
|
|
|
|
# do the actual adding.
|
|
exts[ext_spec.name] = ext_spec
|
|
self._write_extensions(spec, exts)
|
|
|
|
def check_extension_conflict(self, spec, ext_spec):
|
|
exts = self._extension_map(spec)
|
|
if ext_spec.name in exts:
|
|
installed_spec = exts[ext_spec.name].copy(deps=('link', 'run'))
|
|
if ext_spec.copy(deps=('link', 'run')) == installed_spec:
|
|
raise ExtensionAlreadyInstalledError(spec, ext_spec)
|
|
else:
|
|
raise ExtensionConflictError(spec, ext_spec, installed_spec)
|
|
|
|
def check_activated(self, spec, ext_spec):
|
|
exts = self._extension_map(spec)
|
|
if (ext_spec.name not in exts) or (ext_spec != exts[ext_spec.name]):
|
|
raise NoSuchExtensionError(spec, ext_spec)
|
|
|
|
def extension_file_path(self, spec):
|
|
"""Gets full path to an installed package's extension file, which
|
|
keeps track of all the extensions for that package which have been
|
|
added to this view.
|
|
"""
|
|
_check_concrete(spec)
|
|
normalize_path = lambda p: (
|
|
os.path.abspath(p).rstrip(os.path.sep))
|
|
|
|
view_prefix = self.view.get_projection_for_spec(spec)
|
|
if normalize_path(spec.prefix) == normalize_path(view_prefix):
|
|
# For backwards compatibility, when the view is the extended
|
|
# package's installation directory, do not include the spec name
|
|
# as a subdirectory.
|
|
components = [view_prefix, self.layout.metadata_dir,
|
|
self.extension_file_name]
|
|
else:
|
|
components = [view_prefix, self.layout.metadata_dir, spec.name,
|
|
self.extension_file_name]
|
|
|
|
return os.path.join(*components)
|
|
|
|
def extension_map(self, spec):
|
|
"""Defensive copying version of _extension_map() for external API."""
|
|
_check_concrete(spec)
|
|
return self._extension_map(spec).copy()
|
|
|
|
def remove_extension(self, spec, ext_spec):
|
|
_check_concrete(spec)
|
|
_check_concrete(ext_spec)
|
|
|
|
# Make sure it's installed before removing.
|
|
exts = self._extension_map(spec)
|
|
self.check_activated(spec, ext_spec)
|
|
|
|
# do the actual removing.
|
|
del exts[ext_spec.name]
|
|
self._write_extensions(spec, exts)
|
|
|
|
def _extension_map(self, spec):
|
|
"""Get a dict<name -> spec> for all extensions currently
|
|
installed for this package."""
|
|
_check_concrete(spec)
|
|
|
|
if spec not in self._extension_maps:
|
|
path = self.extension_file_path(spec)
|
|
if not os.path.exists(path):
|
|
self._extension_maps[spec] = {}
|
|
|
|
else:
|
|
by_hash = self.layout.specs_by_hash()
|
|
exts = {}
|
|
with open(path) as ext_file:
|
|
yaml_file = yaml.load(ext_file)
|
|
for entry in yaml_file['extensions']:
|
|
name = next(iter(entry))
|
|
dag_hash = entry[name]['hash']
|
|
prefix = entry[name]['path']
|
|
|
|
if dag_hash not in by_hash:
|
|
raise InvalidExtensionSpecError(
|
|
"Spec %s not found in %s" % (dag_hash, prefix))
|
|
|
|
ext_spec = by_hash[dag_hash]
|
|
if prefix != ext_spec.prefix:
|
|
raise InvalidExtensionSpecError(
|
|
"Prefix %s does not match spec hash %s: %s"
|
|
% (prefix, dag_hash, ext_spec))
|
|
|
|
exts[ext_spec.name] = ext_spec
|
|
self._extension_maps[spec] = exts
|
|
|
|
return self._extension_maps[spec]
|
|
|
|
def _write_extensions(self, spec, extensions):
|
|
path = self.extension_file_path(spec)
|
|
|
|
if not extensions:
|
|
# Remove the empty extensions file
|
|
os.remove(path)
|
|
return
|
|
|
|
# Create a temp file in the same directory as the actual file.
|
|
dirname, basename = os.path.split(path)
|
|
fs.mkdirp(dirname)
|
|
|
|
tmp = tempfile.NamedTemporaryFile(
|
|
prefix=basename, dir=dirname, delete=False)
|
|
|
|
# write tmp file
|
|
with tmp:
|
|
yaml.dump({
|
|
'extensions': [
|
|
{ext.name: {
|
|
'hash': ext.dag_hash(),
|
|
'path': str(ext.prefix)
|
|
}} for ext in sorted(extensions.values())]
|
|
}, tmp, default_flow_style=False, encoding='utf-8')
|
|
|
|
# Atomic update by moving tmpfile on top of old one.
|
|
os.rename(tmp.name, path)
|
|
|
|
|
|
class DirectoryLayoutError(SpackError):
|
|
"""Superclass for directory layout errors."""
|
|
|
|
def __init__(self, message, long_msg=None):
|
|
super(DirectoryLayoutError, self).__init__(message, long_msg)
|
|
|
|
|
|
class SpecHashCollisionError(DirectoryLayoutError):
|
|
"""Raised when there is a hash collision in an install layout."""
|
|
|
|
def __init__(self, installed_spec, new_spec):
|
|
super(SpecHashCollisionError, self).__init__(
|
|
'Specs %s and %s have the same SHA-1 prefix!'
|
|
% (installed_spec, new_spec))
|
|
|
|
|
|
class RemoveFailedError(DirectoryLayoutError):
|
|
"""Raised when a DirectoryLayout cannot remove an install prefix."""
|
|
|
|
def __init__(self, installed_spec, prefix, error):
|
|
super(RemoveFailedError, self).__init__(
|
|
'Could not remove prefix %s for %s : %s'
|
|
% (prefix, installed_spec.short_spec, error))
|
|
self.cause = error
|
|
|
|
|
|
class InconsistentInstallDirectoryError(DirectoryLayoutError):
|
|
"""Raised when a package seems to be installed to the wrong place."""
|
|
|
|
def __init__(self, message, long_msg=None):
|
|
super(InconsistentInstallDirectoryError, self).__init__(
|
|
message, long_msg)
|
|
|
|
|
|
class SpecReadError(DirectoryLayoutError):
|
|
"""Raised when directory layout can't read a spec."""
|
|
|
|
|
|
class InvalidDirectoryLayoutParametersError(DirectoryLayoutError):
|
|
"""Raised when a invalid directory layout parameters are supplied"""
|
|
|
|
def __init__(self, message, long_msg=None):
|
|
super(InvalidDirectoryLayoutParametersError, self).__init__(
|
|
message, long_msg)
|
|
|
|
|
|
class InvalidExtensionSpecError(DirectoryLayoutError):
|
|
"""Raised when an extension file has a bad spec in it."""
|
|
|
|
|
|
class ExtensionAlreadyInstalledError(DirectoryLayoutError):
|
|
"""Raised when an extension is added to a package that already has it."""
|
|
|
|
def __init__(self, spec, ext_spec):
|
|
super(ExtensionAlreadyInstalledError, self).__init__(
|
|
"%s is already installed in %s"
|
|
% (ext_spec.short_spec, spec.short_spec))
|
|
|
|
|
|
class ExtensionConflictError(DirectoryLayoutError):
|
|
"""Raised when an extension is added to a package that already has it."""
|
|
|
|
def __init__(self, spec, ext_spec, conflict):
|
|
super(ExtensionConflictError, self).__init__(
|
|
"%s cannot be installed in %s because it conflicts with %s"
|
|
% (ext_spec.short_spec, spec.short_spec, conflict.short_spec))
|
|
|
|
|
|
class NoSuchExtensionError(DirectoryLayoutError):
|
|
"""Raised when an extension isn't there on deactivate."""
|
|
|
|
def __init__(self, spec, ext_spec):
|
|
super(NoSuchExtensionError, self).__init__(
|
|
"%s cannot be removed from %s because it's not activated."
|
|
% (ext_spec.short_spec, spec.short_spec))
|