blob: 14e35b991c5ac4a2c2e1db201f3e6d18d7119273 [file] [log] [blame]
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
# Authors:
# Tristan Van Berkom <tristan.vanberkom@codethink.co.uk>
"""
Plugin - Base plugin class
==========================
BuildStream supports third party plugins to define additional kinds of
:mod:`Elements <buildstream.element>` and :mod:`Sources <buildstream.source>`.
The common API is documented here, along with some information on how
external plugin packages are structured.
.. _core_plugin_abstract_methods:
Abstract Methods
----------------
For both :mod:`Elements <buildstream.element>` and :mod:`Sources <buildstream.source>`,
it is mandatory to implement the following abstract methods:
* :func:`Plugin.configure() <buildstream.plugin.Plugin.configure>`
Loads the user provided configuration YAML for the given source or element
* :func:`Plugin.preflight() <buildstream.plugin.Plugin.preflight>`
Early preflight checks allow plugins to bail out early with an error
in the case that it can predict that failure is inevitable.
* :func:`Plugin.get_unique_key() <buildstream.plugin.Plugin.get_unique_key>`
Once all configuration has been loaded and preflight checks have passed,
this method is used to inform the core of a plugin's unique configuration.
Configurable Warnings
---------------------
Warnings raised through calling :func:`Plugin.warn() <buildstream.plugin.Plugin.warn>` can provide an optional
parameter ``warning_token``, this will raise a :class:`PluginError` if the warning is configured as fatal within
the project configuration.
Configurable warnings will be prefixed with :func:`Plugin.get_kind() <buildstream.plugin.Plugin.get_kind>`
within buildstream and must be prefixed as such in project configurations. For more detail on project configuration
see :ref:`Configurable Warnings <configurable_warnings>`.
It is important to document these warnings in your plugin documentation to allow users to make full use of them
while configuring their projects.
Example
~~~~~~~
If the :class:`git <buildstream.plugins.sources.git.GitSource>` plugin uses the warning ``"inconsistent-submodule"``
then it could be referenced in project configuration as ``"git:inconsistent-submodule"``.
Plugin Structure
----------------
A plugin should consist of a `setuptools package
<http://setuptools.readthedocs.io/en/latest/setuptools.html>`_ that
advertises contained plugins using `entry points
<http://setuptools.readthedocs.io/en/latest/setuptools.html#dynamic-discovery-of-services-and-plugins>`_.
A plugin entry point must be a module that extends a class in the
:ref:`core_framework` to be discovered by BuildStream. A YAML file
defining plugin default settings with the same name as the module can
also be defined in the same directory as the plugin module.
.. note::
BuildStream does not support function/class entry points.
A sample plugin could be structured as such:
.. code-block:: text
.
├── elements
│   ├── autotools.py
│   ├── autotools.yaml
│   └── __init__.py
├── MANIFEST.in
└── setup.py
The setuptools configuration should then contain at least:
setup.py:
.. literalinclude:: ../source/sample_plugin/setup.py
:language: python
MANIFEST.in:
.. literalinclude:: ../source/sample_plugin/MANIFEST.in
:language: text
Class Reference
---------------
"""
import itertools
#
# In some cases (on Debian with python 3.9 it has been seen) when multiple threads
# enter the Plugin.blocking_activity() context manager simultaneously, we get ImportErrors
# from the multiprocessing submodules complaining that we are importing symbols from
# partially initialized submodules (hinting at possible circular imports).
#
# Ensuring that these submodules have been initialized up front circumvents these edge
# case stack trace bugs from occurring.
#
import multiprocessing
import multiprocessing.queues
import multiprocessing.synchronize
import multiprocessing.popen_forkserver # type: ignore
import os
import queue
import signal
import subprocess
import sys
import traceback
from contextlib import contextmanager, suppress
from typing import IO, Any, Callable, Generator, Optional, Sequence, Dict, Tuple, TypeVar, Union, TYPE_CHECKING
from weakref import WeakValueDictionary
from . import utils, _signals
from ._exceptions import PluginError, ImplError
from ._message import Message, MessageType
from .node import Node, MappingNode
from .types import CoreWarnings, SourceRef
if TYPE_CHECKING:
# pylint: disable=cyclic-import
from ._context import Context
from ._project import Project
# pylint: enable=cyclic-import
T1 = TypeVar("T1")
#
# Some types used for subprocess.Popen() wrappers
#
# This recipe was gleaned from observing:
# https://github.com/python/typeshed/blob/master/stdlib/_typeshed/__init__.pyi
# https://github.com/python/typeshed/blob/master/stdlib/subprocess.pyi
#
_FILE = Union[None, int, IO[Any]]
_TXT = Union[bytes, str]
_STR_BYTES_PATH = Union[str, bytes, "os.PathLike[str]", "os.PathLike[bytes]"]
if sys.version_info >= (3, 8):
_CMD = Union[_STR_BYTES_PATH, Sequence[_STR_BYTES_PATH]]
else:
# Python 3.6 doesn't support _CMD being a single PathLike.
# See: https://bugs.python.org/issue31961
_CMD = Union[_TXT, Sequence[_STR_BYTES_PATH]]
# _background_job_wrapper()
#
# Wrapper for running jobs in the background, transparently for users
#
# This method will put on the queue a response of the form:
# (PickleError, OtherError, Result)
#
# Args:
# result_queue: The queue in which to pass back the result
# target: function to execute in the background
# args: positional arguments to give to the target function
#
def _background_job_wrapper(result_queue: multiprocessing.Queue, target: Callable[..., T1], args: Any) -> None:
result = None
try:
result = target(*args)
result_queue.put((None, result))
except Exception: # pylint: disable=broad-except
result_queue.put((traceback.format_exc(), None))
class Plugin:
"""Plugin()
Base Plugin class.
Some common features to both Sources and Elements are found
in this class.
.. note::
Derivation of plugins is not supported. Plugins may only
derive from the base :mod:`Source <buildstream.source>` and
:mod:`Element <buildstream.element>` types, and any convenience
subclasses (like :mod:`BuildElement <buildstream.buildelement>`)
which are included in the buildstream namespace.
"""
BST_MIN_VERSION: Optional[str] = None
"""The minimum required version of BuildStream required by this plugin.
The version must be expressed as the string *"<major>.<minor>"*, where the
*major* version number is the API version and the *minor* version number is
the revision of the same BuildStream API where new symbols might have been
added to the API.
**Example:**
The following statement means that this plugin works with *BuildStream 2.X*,
only if *X >= 2*:
.. code:: python
class Foo(Source):
# Our plugin requires 2.2
BST_MIN_VERSION = "2.2"
.. note::
This version works exactly the same was as the :ref:`min-version <project_min_version>`
which must be specified in the project.conf file.
"""
BST_PLUGIN_DEPRECATED = False
"""True if this element plugin has been deprecated.
If this is set to true, BuildStream will emit a deprecation warning
in any place where this plugin is used.
The deprecation warnings can be suppressed when defining the
:ref:`plugin origins in your project configuration <project_plugins_deprecation>`
"""
BST_PLUGIN_DEPRECATION_MESSAGE = None
"""An additional message to report when a plugin is deprecated
This can be used to refer the user to a suitable replacement or
alternative approach when the plugin is deprecated.
"""
# Unique id generator for Plugins
#
# Each plugin gets a unique id at creation.
#
# Ids are a monotically increasing integer which
# starts as 1 (a falsy plugin ID is considered unset
# in various parts of the codebase).
#
__id_generator = itertools.count(1)
# Hold on to a lookup table by counter of all instantiated plugins.
# We use this to send the id back from child processes so we can lookup
# corresponding element/source in the master process.
#
# Use WeakValueDictionary() so the map we use to lookup objects does not
# keep the plugins alive after pipeline destruction.
#
# Note that Plugins can only be instantiated in the main process before
# scheduling tasks.
__TABLE = WeakValueDictionary() # type: WeakValueDictionary[int, Plugin]
try:
__multiprocessing_context: multiprocessing.context.BaseContext = multiprocessing.get_context("forkserver")
except ValueError:
# We are on a system without `forkserver` support. Let's default to
# spawn. This seems to be hanging however in some rare cases.
#
# Support is not as critical for now, since we do not work on
# platforms not supporting forkserver for now (mainly Windows)
#
# XXX: investigate why we sometimes get deadlocks there
__multiprocessing_context = multiprocessing.get_context("spawn")
def __init__(
self,
name: str,
context: "Context",
project: "Project",
provenance_node: Node,
type_tag: str,
unique_id: Optional[int] = None,
):
self.name = name
"""The plugin name
For elements, this is the project relative bst filename,
for sources this is the owning element's name with a suffix
indicating its index on the owning element.
For sources this is for display purposes only.
"""
# Unique ID
#
# This id allows to uniquely identify a plugin.
#
# /!\ the unique id must be an increasing value /!\
# This is because we are depending on it in buildstream.element.Element
# to give us a topological sort over all elements.
# Modifying how we handle ids here will modify the behavior of the
# Element's state handling.
if unique_id is None:
# Register ourself in the table containing all existing plugins
self._unique_id = next(self.__id_generator)
self.__TABLE[self._unique_id] = self
else:
# If the unique ID is passed in the constructor, then it is a cloned
# plugin in a subprocess and should use the same ID.
self._unique_id = unique_id
self.__context = context # The Context object
# Note that when pickling jobs over to a child process, we rely on this
# reference to the Project, it keeps the plugin factory alive. If the
# factory were to be GC'd then we would see undefined behaviour. Make
# sure to test plugin pickling if this reference is to be removed.
self.__project = project # The Project object
self.__provenance_node = provenance_node # The originating YAML node
self.__type_tag = type_tag # The type of plugin (element or source)
self.__configuring = False # Whether we are currently configuring
# Get the full_name as project & type_tag are resolved
self.__full_name = self.__get_full_name()
# Our message kwargs
self._message_kwargs = {"element_name": self._get_full_name()}
# Infer the kind identifier
modulename = type(self).__module__
self.__kind = modulename.rsplit(".", maxsplit=1)[-1]
self.debug("Created: {}".format(self))
def __del__(self):
# Dont send anything through the Message() pipeline at destruction time,
# any subsequent lookup of plugin by unique id would raise KeyError.
if self.__context.log_debug:
sys.stderr.write("DEBUG: Destroyed: {}\n".format(self))
def __str__(self):
return "{kind} {typetag} at {provenance}".format(
kind=self.__kind, typetag=self.__type_tag, provenance=self._get_provenance()
)
#############################################################
# Abstract Methods #
#############################################################
def configure(self, node: MappingNode) -> None:
"""Configure the Plugin from loaded configuration data
Args:
node: The loaded configuration dictionary
Raises:
:class:`.SourceError`: If it's a :class:`.Source` implementation
:class:`.ElementError`: If it's an :class:`.Element` implementation
Plugin implementors should implement this method to read configuration
data and store it.
The :func:`MappingNode.validate_keys() <buildstream.node.MappingNode.validate_keys>` method
should be used to ensure that the user has not specified keys in `node` which are unsupported
by the plugin.
"""
raise ImplError(
"{tag} plugin '{kind}' does not implement configure()".format(tag=self.__type_tag, kind=self.get_kind())
)
def preflight(self) -> None:
"""Preflight Check
Raises:
:class:`.SourceError`: If it's a :class:`.Source` implementation
:class:`.ElementError`: If it's an :class:`.Element` implementation
This method is run after :func:`Plugin.configure() <buildstream.plugin.Plugin.configure>`
and after the pipeline is fully constructed.
Implementors should simply raise :class:`.SourceError` or :class:`.ElementError`
with an informative message in the case that the host environment is
unsuitable for operation.
Plugins which require host tools (only sources usually) should obtain
them with :func:`utils.get_host_tool() <buildstream.utils.get_host_tool>` which
will raise an error automatically informing the user that a host tool is needed.
"""
raise ImplError(
"{tag} plugin '{kind}' does not implement preflight()".format(tag=self.__type_tag, kind=self.get_kind())
)
def get_unique_key(self) -> SourceRef:
"""Return something which uniquely identifies the plugin input
Returns:
A string, list or dictionary which uniquely identifies the input
This is used to construct unique :ref:`cache keys <cachekeys>` for elements
and sources, sources should return something which uniquely identifies the payload,
such as an sha256 sum of a tarball content.
Elements and Sources should implement this by collecting any configurations
which could possibly affect the output and return a dictionary of these settings.
For Sources, this is guaranteed to only be called if
:func:`Source.is_resolved() <buildstream.source.Source.is_resolved>` has returned
``True`` which is to say that the :class:`.Source` is expected to have an exact
:ref:`source ref <core_source_ref>` indicating exactly what source is going to be staged.
.. note::
If your plugin is concerned with API stability, then future extensions of your
plugin YAML configuration which affect the unique key returned here should be added
to this key with care.
A good rule of thumb is to only compute the new value in the returned key if
the value of the newly added YAML key is not equal to it's default value.
"""
raise ImplError(
"{tag} plugin '{kind}' does not implement get_unique_key()".format(
tag=self.__type_tag, kind=self.get_kind()
)
)
#############################################################
# Public Methods #
#############################################################
def get_kind(self) -> str:
"""Fetches the kind of this plugin
Returns:
The kind of this plugin
"""
return self.__kind
def node_get_project_path(self, node, *, check_is_file=False, check_is_dir=False) -> str:
"""Fetches a project path from a dictionary node and validates it
Paths are asserted to never lead to a directory outside of the
project directory. In addition, paths can not point to symbolic
links, fifos, sockets and block/character devices.
The `check_is_file` and `check_is_dir` parameters can be used to
perform additional validations on the path. Note that an
exception will always be raised if both parameters are set to
``True``.
Args:
node (ScalarNode): A Node loaded from YAML containing the path to validate
check_is_file (bool): If ``True`` an error will also be raised
if path does not point to a regular file.
Defaults to ``False``
check_is_dir (bool): If ``True`` an error will also be raised
if path does not point to a directory.
Defaults to ``False``
Returns:
(str): The project path
Raises:
:class:`.LoadError`: In the case that the project path is not
valid or does not exist
**Example:**
.. code:: python
path = self.node_get_project_path(node, 'path')
"""
return self.__project.get_path_from_node(node, check_is_file=check_is_file, check_is_dir=check_is_dir)
def debug(self, brief: str, *, detail: Optional[str] = None) -> None:
"""Print a debugging message
Args:
brief: The brief message
detail: An optional detailed message, can be multiline output
"""
if self.__context.log_debug:
self.__message(MessageType.DEBUG, brief, detail=detail)
def status(self, brief: str, *, detail: Optional[str] = None) -> None:
"""Print a status message
Args:
brief: The brief message
detail: An optional detailed message, can be multiline output
Note: Status messages tell about what a plugin is currently doing
"""
self.__message(MessageType.STATUS, brief, detail=detail)
def info(self, brief: str, *, detail: Optional[str] = None) -> None:
"""Print an informative message
Args:
brief: The brief message
detail: An optional detailed message, can be multiline output
Note: Informative messages tell the user something they might want
to know, like if refreshing an element caused it to change.
The instance full name of the plugin will be generated with the
message, this being the name of the given element, as appose to
the class name of the underlying plugin __kind identifier.
"""
self.__message(MessageType.INFO, brief, detail=detail)
def warn(self, brief: str, *, detail: Optional[str] = None, warning_token: Optional[str] = None) -> None:
"""Print a warning message, checks warning_token against project configuration
Args:
brief: The brief message
detail: An optional detailed message, can be multiline output
warning_token: An optional configurable warning assosciated with this warning,
this will cause PluginError to be raised if this warning is configured as fatal.
Raises:
(:class:`.PluginError`): When warning_token is considered fatal by the project configuration
"""
if warning_token:
warning_token = _prefix_warning(self, warning_token)
brief = "[{}]: {}".format(warning_token, brief)
project = self._get_project()
if project._warning_is_fatal(warning_token):
detail = detail if detail else ""
raise PluginError(message="{}\n{}".format(brief, detail), reason=warning_token)
self.__message(MessageType.WARN, brief=brief, detail=detail)
def log(self, brief: str, *, detail: Optional[str] = None) -> None:
"""Log a message into the plugin's log file
The message will not be shown in the master log at all (so it will not
be displayed to the user on the console).
Args:
brief: The brief message
detail: An optional detailed message, can be multiline output
"""
self.__message(MessageType.LOG, brief, detail=detail)
@contextmanager
def timed_activity(
self, activity_name: str, *, detail: Optional[str] = None, silent_nested: bool = False
) -> Generator[None, None, None]:
"""Context manager for performing timed activities in plugins
Args:
activity_name: The name of the activity
detail: An optional detailed message, can be multiline output
silent_nested: If specified, nested messages will be silenced
This function lets you perform timed tasks in your plugin,
the core will take care of timing the duration of your
task and printing start / fail / success messages.
**Example**
.. code:: python
# Activity will be logged and timed
with self.timed_activity("Mirroring {}".format(self.url)):
# This will raise SourceError on its own
self.call(... command which takes time ...)
"""
# Get the plugin kwargs and pass them along
plugin_kwargs = self._message_kwargs
with self.__context.messenger.timed_activity(
activity_name, detail=detail, silent_nested=silent_nested, **plugin_kwargs
):
yield
def blocking_activity(
self,
target: Callable[..., T1],
args: Sequence[Any],
activity_name: str,
*,
detail: Optional[str] = None,
silent_nested: bool = False
) -> T1:
"""Execute a blocking activity in the background.
This is to execute potentially blocking methods in the background,
in order to avoid starving the scheduler.
The function, its arguments and return value must all be pickleable,
as it will be run in another process. The function should not raise
an exception.
This should be used whenever there is a potential for a blocking
syscall to not return in a reasonable (<1s) amount of time.
For example, you would use this if you were doing a request to a
remote server, without a timeout.
Args:
target: the function to execute in the background
args: positional arguments to pass to the method to call
activity_name: The name of the activity
detail: An optional detailed message, can be multiline output
silent_nested: If specified, nested messages will be silenced
Returns:
the return value from `target`.
"""
with self.__context.messenger.timed_activity(
activity_name, element_name=self._get_full_name(), detail=detail, silent_nested=silent_nested
):
result_queue = self.__multiprocessing_context.Queue()
proc = None
def kill_proc():
if proc and proc.is_alive():
proc.kill()
proc.join()
def suspend_proc():
if proc and proc.is_alive():
with suppress(ProcessLookupError):
os.kill(proc.pid, signal.SIGSTOP)
def resume_proc():
if proc and proc.is_alive():
with suppress(ProcessLookupError):
os.kill(proc.pid, signal.SIGCONT)
with _signals.suspendable(suspend_proc, resume_proc), _signals.terminator(kill_proc):
proc = self.__multiprocessing_context.Process(
target=_background_job_wrapper, args=(result_queue, target, args)
)
proc.start()
should_continue = True
last_check = False
while should_continue or last_check:
last_check = False
try:
err, result = result_queue.get(timeout=1)
break
except queue.Empty:
if not proc.is_alive() and should_continue:
# Let's check one last time, just in case it stopped
# between our last check and now
last_check = True
should_continue = False
continue
else:
raise PluginError("Background process died with error code {}".format(proc.exitcode))
try:
proc.join(timeout=15)
proc.terminate()
except TimeoutError:
raise PluginError("Background process didn't exit after 15 seconds and got killed.")
if err is not None:
raise PluginError("An error happened while running a blocking activity", detail=err)
return result
def call(
self,
args: _CMD,
fail: Optional[str] = None,
fail_temporarily: bool = False,
cwd: Optional[_STR_BYTES_PATH] = None,
env: Optional[Dict[str, str]] = None,
stdin: Optional[_FILE] = None,
stdout: Optional[_FILE] = None,
stderr: Optional[_FILE] = None,
) -> int:
"""A wrapper for subprocess.call()
Args:
args: A sequence of program arguments or else a string or `path-like-object <https://docs.python.org/3/glossary.html#term-path-like-object>`_
fail: A message to display if the process returns a non zero exit code
fail_temporarily: Whether any exceptions should be raised as temporary.
cwd: Optionally specify the working directory for the command
env: Optionally specify some environment variables for the command
stdin: Optionally specify standard input for the command
stdout: Optionally specify standard output for the command
stderr: Optionally specify standard error for the command
Returns:
The process exit code.
Raises:
(:class:`.PluginError`): If a non-zero return code is received and *fail* is specified
.. attention::
If *fail* is not specified, then the return value of subprocess.call()
is returned even on error, and no exception is automatically raised.
.. attention::
If *stderr* and/or *stdout* are specified, then these streams are delivered
directly to the plugin instead of being delivered to the associated log file.
**Example**
.. code:: python
# Call some host tool
self.tool = utils.get_host_tool('toolname')
self.call(
[self.tool, '--download-ponies', self.mirror_directory],
"Failed to download ponies from {}".format(
self.mirror_directory))
"""
exit_code, _ = self.__call(
args,
fail=fail,
fail_temporarily=fail_temporarily,
cwd=cwd,
env=env,
stdin=stdin,
stdout=stdout,
stderr=stderr,
)
return exit_code
def check_output(
self,
args: _CMD,
fail: Optional[str] = None,
fail_temporarily: bool = False,
cwd: Optional[_STR_BYTES_PATH] = None,
env: Optional[Dict[str, str]] = None,
stdin: Optional[_FILE] = None,
stderr: Optional[_FILE] = None,
) -> Tuple[int, str]:
"""A wrapper for subprocess.check_output()
Args:
args: A sequence of program arguments or else a string or `path-like-object <https://docs.python.org/3/glossary.html#term-path-like-object>`_
fail: A message to display if the process returns a non zero exit code
fail_temporarily: Whether any exceptions should be raised as temporary.
cwd: Optionally specify the working directory for the command
env: Optionally specify some environment variables for the command
stdin: Optionally specify standard input for the command
stderr: Optionally specify standard error for the command
Returns:
A 2-tuple of form (process exit code, process standard output)
Raises:
(:class:`.PluginError`): If a non-zero return code is received and *fail* is specified
.. attention::
If *fail* is not specified, then the return value of subprocess.call()
is returned even on error, and no exception is automatically raised.
.. attention::
If *stderr* and/or *stdout* are specified, then these streams are delivered
directly to the plugin instead of being delivered to the associated log file.
**Example**
.. code:: python
# Get the tool at preflight time
self.tool = utils.get_host_tool('toolname')
# Call the tool, automatically raise an error
_, output = self.check_output(
[self.tool, '--print-ponies'],
"Failed to print the ponies in {}".format(
self.mirror_directory),
cwd=self.mirror_directory)
# Call the tool, inspect exit code
exit_code, output = self.check_output(
[self.tool, 'get-ref', tracking],
cwd=self.mirror_directory)
if exit_code == 128:
return
elif exit_code != 0:
fmt = "{plugin}: Failed to get ref for tracking: {track}"
raise SourceError(
fmt.format(plugin=self, track=tracking)) from e
"""
return self.__call(
args,
collect_stdout=True,
fail=fail,
fail_temporarily=fail_temporarily,
cwd=cwd,
env=env,
stdin=stdin,
stderr=stderr,
)
#############################################################
# Private Methods used in BuildStream #
#############################################################
# _lookup():
#
# Fetch a plugin in the current process by its
# unique identifier
#
# Args:
# unique_id: The unique identifier as returned by
# plugin._unique_id
#
# Returns:
# (Plugin): The plugin for the given ID, or None
#
@classmethod
def _lookup(cls, unique_id):
assert unique_id != 0, "Looking up invalid plugin ID 0, ID counter starts at 1"
try:
return cls.__TABLE[unique_id]
except KeyError:
assert False, "Could not find plugin with ID {}".format(unique_id)
raise # In case a user is running with "python -O"
# _get_context()
#
# Fetches the invocation context
#
def _get_context(self):
return self.__context
# _get_project()
#
# Fetches the project object associated with this plugin
#
def _get_project(self):
return self.__project
# _get_provenance():
#
# Fetch bst file, line and column of the entity
#
def _get_provenance(self):
return self.__provenance_node.get_provenance()
# Context manager for getting the open file handle to this
# plugin's log. Used in the child context to add stuff to
# a log.
#
@contextmanager
def _output_file(self):
log = self.__context.messenger.get_log_handle()
if log is None:
with open(os.devnull, "w", encoding="utf-8") as output:
yield output
else:
yield log
# _configure():
#
# Calls configure() for the plugin, this must be called by
# the core instead of configure() directly, so that the
# _get_configuring() state is up to date.
#
# Args:
# node (buildstream.node.MappingNode): The loaded configuration dictionary
#
def _configure(self, node):
self.__configuring = True
self.configure(node)
self.__configuring = False
# _get_configuring():
#
# Checks whether the plugin is in the middle of having
# its Plugin.configure() method called
#
# Returns:
# (bool): Whether we are currently configuring
def _get_configuring(self):
return self.__configuring
# _preflight():
#
# Calls preflight() for the plugin, and allows generic preflight
# checks to be added
#
# Raises:
# SourceError: If it's a Source implementation
# ElementError: If it's an Element implementation
# ProgramNotFoundError: If a required host tool is not found
#
def _preflight(self):
self.preflight()
# _get_full_name():
#
# The instance full name of the plugin prepended with the owning
# junction if appropriate. This being the name of the given element,
# as appose to the class name of the underlying plugin __kind identifier.
#
# Returns:
# (str): element full name, with prepended owning junction if appropriate
#
def _get_full_name(self):
return self.__full_name
#############################################################
# Local Private Methods #
#############################################################
# Internal subprocess implementation for the call() and check_output() APIs
#
def __call(
self,
args: _CMD,
fail: Optional[str] = None,
fail_temporarily: bool = False,
collect_stdout: bool = False,
cwd: Optional[_STR_BYTES_PATH] = None,
env: Optional[Dict[str, str]] = None,
stdin: Optional[_FILE] = None,
stdout: Optional[_FILE] = None,
stderr: Optional[_FILE] = None,
):
with self._output_file() as output_file:
if stdout is None:
stdout = output_file
if stderr is None:
stderr = output_file
if collect_stdout:
stdout = subprocess.PIPE
self.__note_command(output_file, args, cwd)
exit_code, output = utils._call(args, cwd=cwd, env=env, stdin=stdin, stdout=stdout, stderr=stderr)
if fail and exit_code:
raise PluginError("{plugin}: {message}".format(plugin=self, message=fail), temporary=fail_temporarily)
return (exit_code, output)
# __message():
#
# The plugin level focal point for issuing messages.
#
# Args:
# message_type (MessageType): The message type
# brief (str): The brief message
# kwargs: The remaining Message attributes
#
def __message(self, message_type, brief, **kwargs):
#
# Merge the plugin kwargs with the explicitly passed kwargs, give
# precedence to the explicitly passed kwargs.
#
plugin_kwargs = self._message_kwargs.copy()
plugin_kwargs.update(kwargs)
message = Message(message_type, brief, **plugin_kwargs)
self.__context.messenger.message(message)
# __note_command():
#
# Make a note in the logs that we are running a process on the host.
#
# Args:
# output: The output logging file
# args: The command with args which we're about to run
# cwd: The host side working directory to run the command, or None
#
def __note_command(self, output, args, cwd):
workdir = cwd or os.getcwd()
command = " ".join(args)
output.write("Running host command {}: {}\n".format(workdir, command))
output.flush()
self.status("Running host command", detail=command)
def __get_full_name(self):
project = self.__project
# Set the name, depending on element or source plugin type
name = self._element_name if self.__type_tag == "source" else self.name # pylint: disable=no-member
if project.junction:
return "{}:{}".format(project.junction._get_full_name(), name)
else:
return name
# A local table for _prefix_warning()
#
__CORE_WARNINGS = [value for name, value in CoreWarnings.__dict__.items() if not name.startswith("__")]
# _prefix_warning():
#
# Prefix a warning with the plugin kind. CoreWarnings are not prefixed.
#
# Args:
# plugin (Plugin): The plugin which raised the warning
# warning (str): The warning to prefix
#
# Returns:
# (str): A prefixed warning
#
def _prefix_warning(plugin, warning):
if any((warning is core_warning for core_warning in __CORE_WARNINGS)):
return warning
return "{}:{}".format(plugin.get_kind(), warning)