blob: dd2183021b78b737c63aabf879cf140fbc8d687e [file] [log] [blame]
#!/usr/bin/env python3
#
# Copyright (C) 2016-2018 Codethink Limited
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library. If not, see <http://www.gnu.org/licenses/>.
#
# Authors:
# Tristan Van Berkom <tristan.vanberkom@codethink.co.uk>
import os
import multiprocessing # for cpu_count()
from collections import Mapping
from pluginbase import PluginBase
from . import utils
from . import _cachekey
from . import _site
from . import _yaml
from ._profile import Topics, profile_start, profile_end
from ._exceptions import LoadError, LoadErrorReason
from ._options import OptionPool
from ._artifactcache import artifact_cache_specs_from_config_node
from ._elementfactory import ElementFactory
from ._sourcefactory import SourceFactory
# The base BuildStream format version
#
# This version is bumped whenever enhancements are made
# to the `project.conf` format or the core element format.
#
BST_FORMAT_VERSION = 2
# The separator we use for user specified aliases
_ALIAS_SEPARATOR = ':'
# Project()
#
# The Project Configuration
#
class Project():
def __init__(self, directory, context, *, junction=None, cli_options=None):
# The project name
self.name = None
# The project directory
self.directory = os.path.abspath(directory)
# Absolute path to where elements are loaded from within the project
self.element_path = None
self._context = context # The invocation Context
self._variables = {} # The default variables overridden with project wide overrides
self._environment = {} # The base sandbox environment
self._elements = {} # Element specific configurations
self._sources = {} # Source specific configurations
self._aliases = {} # Aliases dictionary
self._workspaces = {} # Workspaces
self._plugin_source_origins = [] # Origins of custom sources
self._plugin_element_origins = [] # Origins of custom elements
self._options = None # Project options, the OptionPool
self._junction = junction # The junction element, if this is a subproject
self._cli_options = cli_options
self._cache_key = None
self._source_format_versions = {}
self._element_format_versions = {}
self._fail_on_overlap = False
# Shell options
self._shell_command = [] # The default interactive shell command
self._shell_env_inherit = [] # Environment vars to inherit when non-isolated
self._shell_host_files = {} # Mapping of sandbox paths to host paths
profile_start(Topics.LOAD_PROJECT, self.directory.replace(os.sep, '-'))
self._load()
profile_end(Topics.LOAD_PROJECT, self.directory.replace(os.sep, '-'))
self._context._add_project(self)
# translate_url():
#
# Translates the given url which may be specified with an alias
# into a fully qualified url.
#
# Args:
# url (str): A url, which may be using an alias
#
# Returns:
# str: The fully qualified url, with aliases resolved
#
# This method is provided for :class:`.Source` objects to resolve
# fully qualified urls based on the shorthand which is allowed
# to be specified in the YAML
def translate_url(self, url):
if url and _ALIAS_SEPARATOR in url:
url_alias, url_body = url.split(_ALIAS_SEPARATOR, 1)
alias_url = self._aliases.get(url_alias)
if alias_url:
url = alias_url + url_body
return url
# _load():
#
# Loads the project configuration file in the project directory.
#
# Raises: LoadError if there was a problem with the project.conf
#
def _load(self):
# Load builtin default
projectfile = os.path.join(self.directory, "project.conf")
config = _yaml.load(_site.default_project_config)
# Load project local config and override the builtin
project_conf = _yaml.load(projectfile)
_yaml.composite(config, project_conf)
# Element and Source type configurations will be composited later onto
# element/source types, so we delete it from here and run our final
# assertion after.
self._elements = _yaml.node_get(config, Mapping, 'elements', default_value={})
self._sources = _yaml.node_get(config, Mapping, 'sources', default_value={})
config.pop('elements', None)
config.pop('sources', None)
_yaml.node_final_assertions(config)
_yaml.node_validate(config, [
'format-version',
'element-path', 'variables',
'environment', 'environment-nocache',
'split-rules', 'elements', 'plugins',
'aliases', 'name',
'artifacts', 'options',
'fail-on-overlap', 'shell'
])
# The project name, element path and option declarations
# are constant and cannot be overridden by option conditional statements
self.name = _yaml.node_get(config, str, 'name')
self.element_path = os.path.join(
self.directory,
_yaml.node_get(config, str, 'element-path', default_value='.')
)
# Load project options
options_node = _yaml.node_get(config, Mapping, 'options', default_value={})
self._options = OptionPool(self.element_path)
self._options.load(options_node)
if self._junction:
# load before user configuration
self._options.load_yaml_values(self._junction.options, transform=self._junction._subst_string)
# Collect option values specified in the user configuration
overrides = self._context._get_overrides(self.name)
override_options = _yaml.node_get(overrides, Mapping, 'options', default_value={})
self._options.load_yaml_values(override_options)
if self._cli_options:
self._options.load_cli_values(self._cli_options)
# We're done modifying options, now we can use them for substitutions
self._options.resolve()
#
# Now resolve any conditionals in the remaining configuration,
# any conditionals specified for project option declarations,
# or conditionally specifying the project name; will be ignored.
#
self._options.process_node(config)
#
# Now all YAML composition is done, from here on we just load
# the values from our loaded configuration dictionary.
#
# Load artifacts pull/push configuration for this project
self.artifact_cache_specs = artifact_cache_specs_from_config_node(config)
# Workspace configurations
self._workspaces = self._load_workspace_config()
self._ensure_workspace_config_format()
# Assert project version
format_version = _yaml.node_get(config, int, 'format-version', default_value=0)
if BST_FORMAT_VERSION < format_version:
major, minor = utils.get_bst_version()
raise LoadError(
LoadErrorReason.UNSUPPORTED_PROJECT,
"Project requested format version {}, but BuildStream {}.{} only supports up until format version {}"
.format(format_version, major, minor, BST_FORMAT_VERSION))
# Plugin origins and versions
origins = _yaml.node_get(config, list, 'plugins', default_value=[])
for origin in origins:
allowed_origin_fields = [
'origin', 'sources', 'elements',
'package-name', 'path',
]
allowed_origins = ['core', 'local', 'pip']
_yaml.node_validate(origin, allowed_origin_fields)
if origin['origin'] not in allowed_origins:
raise LoadError(
LoadErrorReason.INVALID_YAML,
"Origin '{}' is not one of the allowed types"
.format(origin['origin']))
# Store source versions for checking later
source_versions = _yaml.node_get(origin, Mapping, 'sources', default_value={})
for key, _ in _yaml.node_items(source_versions):
if key in self._source_format_versions:
raise LoadError(
LoadErrorReason.INVALID_YAML,
"Duplicate listing of source '{}'".format(key))
self._source_format_versions[key] = _yaml.node_get(source_versions, int, key)
# Store element versions for checking later
element_versions = _yaml.node_get(origin, Mapping, 'elements', default_value={})
for key, _ in _yaml.node_items(element_versions):
if key in self._element_format_versions:
raise LoadError(
LoadErrorReason.INVALID_YAML,
"Duplicate listing of element '{}'".format(key))
self._element_format_versions[key] = _yaml.node_get(element_versions, int, key)
# Store the origins if they're not 'core'.
# core elements are loaded by default, so storing is unnecessary.
if _yaml.node_get(origin, str, 'origin') != 'core':
self._store_origin(origin, 'sources', self._plugin_source_origins)
self._store_origin(origin, 'elements', self._plugin_element_origins)
pluginbase = PluginBase(package='buildstream.plugins')
self._element_factory = ElementFactory(pluginbase, self._plugin_element_origins)
self._source_factory = SourceFactory(pluginbase, self._plugin_source_origins)
# Source url aliases
self._aliases = _yaml.node_get(config, Mapping, 'aliases', default_value={})
# Load base variables
self._variables = _yaml.node_get(config, Mapping, 'variables')
# Add the project name as a default variable
self._variables['project-name'] = self.name
# Extend variables with automatic variables and option exports
# Initialize it as a string as all variables are processed as strings.
self._variables['max-jobs'] = str(multiprocessing.cpu_count())
# Export options into variables, if that was requested
for _, option in self._options.options.items():
if option.variable:
self._variables[option.variable] = option.get_value()
# Load sandbox configuration
self._environment = _yaml.node_get(config, Mapping, 'environment')
self._env_nocache = _yaml.node_get(config, list, 'environment-nocache')
# Load project split rules
self._splits = _yaml.node_get(config, Mapping, 'split-rules')
# Fail on overlap
self._fail_on_overlap = _yaml.node_get(config, bool, 'fail-on-overlap',
default_value=False)
# Parse shell options
shell_options = _yaml.node_get(config, Mapping, 'shell', default_value={})
_yaml.node_validate(shell_options, ['command', 'environment-inherit', 'host-files'])
self._shell_command = _yaml.node_get(shell_options, list, 'command',
default_value=['sh', '-i'])
self._shell_env_inherit = _yaml.node_get(shell_options, list, 'environment-inherit',
default_value=[])
# Host files is parsed as a list for convenience
host_files = _yaml.node_get(shell_options, list, 'host-files', default_value=[])
for host_file in host_files:
if isinstance(host_file, str):
self._shell_host_files[host_file] = host_file
else:
index = host_files.index(host_file)
host_file_desc = _yaml.node_get(shell_options, Mapping, 'host-files', indices=[index])
_yaml.node_validate(host_file_desc, ['host', 'sandbox'])
host_path = _yaml.node_get(host_file_desc, str, 'host')
sandbox_path = _yaml.node_get(host_file_desc, str, 'sandbox')
self._shell_host_files[sandbox_path] = host_path
# _store_origin()
#
# Helper function to store plugin origins
#
# Args:
# origin (dict) - a dictionary indicating the origin of a group of
# plugins.
# plugin_group (str) - The name of the type of plugin that is being
# loaded
# destination (list) - A list of dicts to store the origins in
#
# Raises:
# LoadError if 'origin' is an unexpected value
def _store_origin(self, origin, plugin_group, destination):
expected_groups = ['sources', 'elements']
if plugin_group not in expected_groups:
raise LoadError(LoadErrorReason.INVALID_DATA,
"Unexpected plugin group: {}, expecting {}"
.format(plugin_group, expected_groups))
if plugin_group in origin:
origin_dict = _yaml.node_copy(origin)
plugins = _yaml.node_get(origin, Mapping, plugin_group, default_value={})
origin_dict['plugins'] = [k for k, _ in _yaml.node_items(plugins)]
for group in expected_groups:
if group in origin_dict:
del origin_dict[group]
if origin_dict['origin'] == 'local':
# paths are passed in relative to the project, but must be absolute
origin_dict['path'] = os.path.join(self.directory, origin_dict['path'])
destination.append(origin_dict)
# _list_workspaces()
#
# Generator function to enumerate workspaces.
#
# Yields:
# A tuple in the following format: (element, path).
def _list_workspaces(self):
for element, _ in _yaml.node_items(self._workspaces):
yield (element, self._workspaces[element])
# _get_workspace()
#
# Get the path of the workspace source associated with the given
# element's source at the given index
#
# Args:
# element (str) - The element name
#
# Returns:
# None if no workspace is open, the path to the workspace
# otherwise
#
def _get_workspace(self, element):
try:
return self._workspaces[element]
except KeyError:
return None
# _set_workspace()
#
# Set the path of the workspace associated with the given
# element's source at the given index
#
# Args:
# element (str) - The element name
# path (str) - The path to set the workspace to
#
def _set_workspace(self, element, path):
if element.name not in self._workspaces:
self._workspaces[element.name] = {}
self._workspaces[element.name] = path
element._set_source_workspaces(path)
# _delete_workspace()
#
# Remove the workspace from the workspace element. Note that this
# does *not* remove the workspace from the stored yaml
# configuration, call _save_workspace_config() afterwards.
#
# Args:
# element (str) - The element name
#
def _delete_workspace(self, element):
del self._workspaces[element]
# _load_workspace_config()
#
# Load the workspace configuration and return a node containing
# all open workspaces for the project
#
# Returns:
#
# A node containing a dict that assigns elements to their
# workspaces. For example:
#
# alpha.bst: /home/me/alpha
# bravo.bst: /home/me/bravo
#
def _load_workspace_config(self):
os.makedirs(os.path.join(self.directory, ".bst"), exist_ok=True)
workspace_file = os.path.join(self.directory, ".bst", "workspaces.yml")
try:
open(workspace_file, "a").close()
except IOError as e:
raise LoadError(LoadErrorReason.MISSING_FILE,
"Could not load workspace config: {}".format(e)) from e
return _yaml.load(workspace_file)
# _ensure_workspace_config_format()
#
# If workspace config is in old-style format, i.e. it is using
# source-specific workspaces, try to convert it to element-specific
# workspaces.
#
# This method will rewrite workspace config, if it is in old format.
#
# Args:
# workspaces (dict): current workspace config, usually output of _load_workspace_config()
#
# Raises: LoadError if there was a problem with the workspace config
#
def _ensure_workspace_config_format(self):
needs_rewrite = False
for element, config in _yaml.node_items(self._workspaces):
if isinstance(config, str):
pass
elif isinstance(config, dict):
sources = list(_yaml.node_items(config))
if len(sources) > 1:
detail = "There are multiple workspaces open for '{}'.\n" + \
"This is not supported anymore.\n" + \
"Please remove this element from '{}'."
raise LoadError(LoadErrorReason.INVALID_DATA,
detail.format(element,
os.path.join(self.directory, ".bst", "workspaces.yml")))
self._workspaces[element] = sources[0][1]
needs_rewrite = True
else:
raise LoadError(LoadErrorReason.INVALID_DATA,
"Workspace config is in unexpected format.")
if needs_rewrite:
self._save_workspace_config()
# _save_workspace_config()
#
# Dump the current workspace element to the project configuration
# file. This makes any changes performed with _delete_workspace or
# _set_workspace permanent
#
def _save_workspace_config(self):
_yaml.dump(_yaml.node_sanitize(self._workspaces),
os.path.join(self.directory, ".bst", "workspaces.yml"))
def _extract_plugin_paths(self, node, name):
if not node:
return
path_list = _yaml.node_get(node, list, name, default_value=[])
for i in range(len(path_list)):
path = _yaml.node_get(node, str, name, indices=[i])
yield path
# _get_cache_key():
#
# Returns the cache key, calculating it if necessary
#
# Returns:
# (str): A hex digest cache key for the Context
#
def _get_cache_key(self):
if self._cache_key is None:
# Anything that alters the build goes into the unique key
# (currently nothing here)
self._cache_key = _cachekey.generate_key({})
return self._cache_key
def _create_element(self, kind, artifacts, meta):
return self._element_factory.create(kind, self._context, self, artifacts, meta)
def _create_source(self, kind, meta):
return self._source_factory.create(kind, self._context, self, meta)