"""Miscellaneous utility functions and classes.""" from __future__ import annotations import abc import collections.abc as c import enum import fcntl import importlib.util import inspect import json import keyword import os import platform import pkgutil import random import re import shutil import stat import string import subprocess import sys import time import functools import shlex import typing as t import warnings from struct import unpack, pack from termios import TIOCGWINSZ # CAUTION: Avoid third-party imports in this module whenever possible. # Any third-party imports occurring here will result in an error if they are vendored by ansible-core. from .locale_util import ( LOCALE_WARNING, CONFIGURED_LOCALE, ) from .encoding import ( to_bytes, to_optional_bytes, to_optional_text, ) from .io import ( open_binary_file, read_text_file, ) from .thread import ( mutex, WrappedThread, ) from .constants import ( SUPPORTED_PYTHON_VERSIONS, ) C = t.TypeVar('C') TBase = t.TypeVar('TBase') TKey = t.TypeVar('TKey') TValue = t.TypeVar('TValue') PYTHON_PATHS: dict[str, str] = {} COVERAGE_CONFIG_NAME = 'coveragerc' ANSIBLE_TEST_ROOT = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # assume running from install ANSIBLE_ROOT = os.path.dirname(ANSIBLE_TEST_ROOT) ANSIBLE_LIB_ROOT = os.path.join(ANSIBLE_ROOT, 'ansible') ANSIBLE_SOURCE_ROOT = None if not os.path.exists(ANSIBLE_LIB_ROOT): # running from source ANSIBLE_ROOT = os.path.dirname(os.path.dirname(os.path.dirname(ANSIBLE_TEST_ROOT))) ANSIBLE_LIB_ROOT = os.path.join(ANSIBLE_ROOT, 'lib', 'ansible') ANSIBLE_SOURCE_ROOT = ANSIBLE_ROOT ANSIBLE_TEST_DATA_ROOT = os.path.join(ANSIBLE_TEST_ROOT, '_data') ANSIBLE_TEST_UTIL_ROOT = os.path.join(ANSIBLE_TEST_ROOT, '_util') ANSIBLE_TEST_CONFIG_ROOT = os.path.join(ANSIBLE_TEST_ROOT, 'config') ANSIBLE_TEST_CONTROLLER_ROOT = os.path.join(ANSIBLE_TEST_UTIL_ROOT, 'controller') ANSIBLE_TEST_TARGET_ROOT = os.path.join(ANSIBLE_TEST_UTIL_ROOT, 'target') ANSIBLE_TEST_TOOLS_ROOT = os.path.join(ANSIBLE_TEST_CONTROLLER_ROOT, 'tools') ANSIBLE_TEST_TARGET_TOOLS_ROOT = os.path.join(ANSIBLE_TEST_TARGET_ROOT, 'tools') # Modes are set to allow all users the same level of access. # This permits files to be used in tests that change users. # The only exception is write access to directories for the user creating them. # This avoids having to modify the directory permissions a second time. MODE_READ = stat.S_IRUSR | stat.S_IRGRP | stat.S_IROTH MODE_FILE = MODE_READ MODE_FILE_EXECUTE = MODE_FILE | stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH MODE_FILE_WRITE = MODE_FILE | stat.S_IWUSR | stat.S_IWGRP | stat.S_IWOTH MODE_DIRECTORY = MODE_READ | stat.S_IWUSR | stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH MODE_DIRECTORY_WRITE = MODE_DIRECTORY | stat.S_IWGRP | stat.S_IWOTH class OutputStream(enum.Enum): """The output stream to use when running a subprocess and redirecting/capturing stdout or stderr.""" ORIGINAL = enum.auto() AUTO = enum.auto() def get_buffer(self, original: t.BinaryIO) -> t.BinaryIO: """Return the correct output buffer to use, taking into account the given original buffer.""" if self == OutputStream.ORIGINAL: return original if self == OutputStream.AUTO: return display.fd.buffer raise NotImplementedError(str(self)) class Architecture: """ Normalized architecture names. These are the architectures supported by ansible-test, such as when provisioning remote instances. """ X86_64 = 'x86_64' AARCH64 = 'aarch64' REMOTE_ARCHITECTURES = list(value for key, value in Architecture.__dict__.items() if not key.startswith('__')) WINDOWS_CONNECTION_VARIABLES: dict[str, t.Any] = { 'psrp+http': dict( ansible_port=5985, ansible_psrp_protocol='http', use_password=True, ), 'psrp+https': dict( ansible_port=5986, ansible_psrp_protocol='https', ansible_psrp_cert_validation='ignore', use_password=True, ), 'ssh+key': dict( ansible_port=22, ansible_shell_type='powershell', use_password=False, ), 'ssh+password': dict( ansible_port=22, ansible_shell_type='powershell', use_password=True, ), 'winrm+http': dict( ansible_port=5985, ansible_winrm_scheme='http', ansible_winrm_transport='ntlm', use_password=True, ), 'winrm+https': dict( ansible_port=5986, ansible_winrm_scheme='https', ansible_winrm_server_cert_validation='ignore', use_password=True, ), } """Dictionary of Windows connection types and variables required to use them.""" WINDOWS_CONNECTIONS = list(WINDOWS_CONNECTION_VARIABLES) def is_valid_identifier(value: str) -> bool: """Return True if the given value is a valid non-keyword Python identifier, otherwise return False.""" return value.isidentifier() and not keyword.iskeyword(value) def cache(func: c.Callable[[], TValue]) -> c.Callable[[], TValue]: """Enforce exclusive access on a decorated function and cache the result.""" storage: dict[None, TValue] = {} sentinel = object() @functools.wraps(func) def cache_func(): """Cache the return value from func.""" if (value := storage.get(None, sentinel)) is sentinel: value = storage[None] = func() return value wrapper = mutex(cache_func) return wrapper @mutex def detect_architecture(python: str) -> t.Optional[str]: """Detect the architecture of the specified Python and return a normalized version, or None if it cannot be determined.""" results: dict[str, t.Optional[str]] try: results = detect_architecture.results # type: ignore[attr-defined] except AttributeError: results = detect_architecture.results = {} # type: ignore[attr-defined] if python in results: return results[python] if python == sys.executable or os.path.realpath(python) == os.path.realpath(sys.executable): uname = platform.uname() else: data = raw_command([python, '-c', 'import json, platform; print(json.dumps(platform.uname()));'], capture=True)[0] uname = json.loads(data) translation = { 'x86_64': Architecture.X86_64, # Linux, macOS 'amd64': Architecture.X86_64, # FreeBSD 'aarch64': Architecture.AARCH64, # Linux, FreeBSD 'arm64': Architecture.AARCH64, # FreeBSD } candidates = [] if len(uname) >= 5: candidates.append(uname[4]) if len(uname) >= 6: candidates.append(uname[5]) candidates = sorted(set(candidates)) architectures = sorted(set(arch for arch in [translation.get(candidate) for candidate in candidates] if arch)) architecture: t.Optional[str] = None if not architectures: display.warning(f'Unable to determine architecture for Python interpreter "{python}" from: {candidates}') elif len(architectures) == 1: architecture = architectures[0] display.info(f'Detected architecture {architecture} for Python interpreter: {python}', verbosity=1) else: display.warning(f'Conflicting architectures detected ({architectures}) for Python interpreter "{python}" from: {candidates}') results[python] = architecture return architecture def filter_args(args: list[str], filters: dict[str, int]) -> list[str]: """Return a filtered version of the given command line arguments.""" remaining = 0 result = [] for arg in args: if not arg.startswith('-') and remaining: remaining -= 1 continue remaining = 0 parts = arg.split('=', 1) key = parts[0] if key in filters: remaining = filters[key] - len(parts) + 1 continue result.append(arg) return result def read_lines_without_comments(path: str, remove_blank_lines: bool = False, optional: bool = False) -> list[str]: """ Returns lines from the specified text file with comments removed. Comments are any content from a hash symbol to the end of a line. Any spaces immediately before a comment are also removed. """ if optional and not os.path.exists(path): return [] lines = read_text_file(path).splitlines() lines = [re.sub(r' *#.*$', '', line) for line in lines] if remove_blank_lines: lines = [line for line in lines if line] return lines def exclude_none_values(data: dict[TKey, t.Optional[TValue]]) -> dict[TKey, TValue]: """Return the provided dictionary with any None values excluded.""" return dict((key, value) for key, value in data.items() if value is not None) def find_executable(executable: str, cwd: t.Optional[str] = None, path: t.Optional[str] = None, required: t.Union[bool, str] = True) -> t.Optional[str]: """ Find the specified executable and return the full path, or None if it could not be found. If required is True an exception will be raised if the executable is not found. If required is set to 'warning' then a warning will be shown if the executable is not found. """ match = None real_cwd = os.getcwd() if not cwd: cwd = real_cwd if os.path.dirname(executable): target = os.path.join(cwd, executable) if os.path.exists(target) and os.access(target, os.F_OK | os.X_OK): match = executable else: if path is None: path = os.environ.get('PATH', os.path.defpath) if path: path_dirs = path.split(os.path.pathsep) seen_dirs = set() for path_dir in path_dirs: if path_dir in seen_dirs: continue seen_dirs.add(path_dir) if os.path.abspath(path_dir) == real_cwd: path_dir = cwd candidate = os.path.join(path_dir, executable) if os.path.exists(candidate) and os.access(candidate, os.F_OK | os.X_OK): match = candidate break if not match and required: message = 'Required program "%s" not found.' % executable if required != 'warning': raise ApplicationError(message) display.warning(message) return match def find_python(version: str, path: t.Optional[str] = None, required: bool = True) -> t.Optional[str]: """ Find and return the full path to the specified Python version. If required, an exception will be raised not found. If not required, None will be returned if not found. """ version_info = str_to_version(version) if not path and version_info == sys.version_info[:len(version_info)]: python_bin = sys.executable else: python_bin = find_executable('python%s' % version, path=path, required=required) return python_bin @cache def get_ansible_version() -> str: """Return the Ansible version.""" # ansible may not be in our sys.path # avoids a symlink to release.py since ansible placement relative to ansible-test may change during delegation load_module(os.path.join(ANSIBLE_LIB_ROOT, 'release.py'), 'ansible_release') # noinspection PyUnresolvedReferences from ansible_release import __version__ as ansible_version # pylint: disable=import-error return ansible_version def _enable_vendoring() -> None: """Enable support for loading Python packages vendored by ansible-core.""" # Load the vendoring code by file path, since ansible may not be in our sys.path. # Convert warnings into errors, to avoid problems from surfacing later. with warnings.catch_warnings(): warnings.filterwarnings('error') load_module(os.path.join(ANSIBLE_LIB_ROOT, '_vendor', '__init__.py'), 'ansible_vendor') @cache def get_available_python_versions() -> dict[str, str]: """Return a dictionary indicating which supported Python versions are available.""" return dict((version, path) for version, path in ((version, find_python(version, required=False)) for version in SUPPORTED_PYTHON_VERSIONS) if path) def raw_command( cmd: c.Iterable[str], capture: bool, env: t.Optional[dict[str, str]] = None, data: t.Optional[str] = None, cwd: t.Optional[str] = None, explain: bool = False, stdin: t.Optional[t.Union[t.IO[bytes], int]] = None, stdout: t.Optional[t.Union[t.IO[bytes], int]] = None, interactive: bool = False, output_stream: t.Optional[OutputStream] = None, cmd_verbosity: int = 1, str_errors: str = 'strict', error_callback: t.Optional[c.Callable[[SubprocessError], None]] = None, ) -> tuple[t.Optional[str], t.Optional[str]]: """Run the specified command and return stdout and stderr as a tuple.""" output_stream = output_stream or OutputStream.AUTO if capture and interactive: raise InternalError('Cannot combine capture=True with interactive=True.') if data and interactive: raise InternalError('Cannot combine data with interactive=True.') if stdin and interactive: raise InternalError('Cannot combine stdin with interactive=True.') if stdout and interactive: raise InternalError('Cannot combine stdout with interactive=True.') if stdin and data: raise InternalError('Cannot combine stdin with data.') if stdout and not capture: raise InternalError('Redirection of stdout requires capture=True to avoid redirection of stderr to stdout.') if output_stream != OutputStream.AUTO and capture: raise InternalError(f'Cannot combine {output_stream=} with capture=True.') if output_stream != OutputStream.AUTO and interactive: raise InternalError(f'Cannot combine {output_stream=} with interactive=True.') if not cwd: cwd = os.getcwd() if not env: env = common_environment() cmd = list(cmd) escaped_cmd = shlex.join(cmd) if capture: description = 'Run' elif interactive: description = 'Interactive' else: description = 'Stream' description += ' command' with_types = [] if data: with_types.append('data') if stdin: with_types.append('stdin') if stdout: with_types.append('stdout') if with_types: description += f' with {"/".join(with_types)}' display.info(f'{description}: {escaped_cmd}', verbosity=cmd_verbosity, truncate=True) display.info('Working directory: %s' % cwd, verbosity=2) program = find_executable(cmd[0], cwd=cwd, path=env['PATH'], required=False) if program: display.info('Program found: %s' % program, verbosity=2) for key in sorted(env.keys()): display.info('%s=%s' % (key, env[key]), verbosity=2) if explain: return None, None communicate = False if stdin is not None: data = None elif data is not None: stdin = subprocess.PIPE communicate = True elif interactive: pass # allow the subprocess access to our stdin else: stdin = subprocess.DEVNULL if not interactive: # When not running interactively, send subprocess stdout/stderr through a pipe. # This isolates the stdout/stderr of the subprocess from the current process, and also hides the current TTY from it, if any. # This prevents subprocesses from sharing stdout/stderr with the current process or each other. # Doing so allows subprocesses to safely make changes to their file handles, such as making them non-blocking (ssh does this). # This also maintains consistency between local testing and CI systems, which typically do not provide a TTY. # To maintain output ordering, a single pipe is used for both stdout/stderr when not capturing output unless the output stream is ORIGINAL. stdout = stdout or subprocess.PIPE stderr = subprocess.PIPE if capture or output_stream == OutputStream.ORIGINAL else subprocess.STDOUT communicate = True else: stderr = None start = time.time() process = None try: try: cmd_bytes = [to_bytes(arg) for arg in cmd] env_bytes = dict((to_bytes(k), to_bytes(v)) for k, v in env.items()) process = subprocess.Popen(cmd_bytes, env=env_bytes, stdin=stdin, stdout=stdout, stderr=stderr, cwd=cwd) # pylint: disable=consider-using-with except FileNotFoundError as ex: raise ApplicationError('Required program "%s" not found.' % cmd[0]) from ex if communicate: data_bytes = to_optional_bytes(data) stdout_bytes, stderr_bytes = communicate_with_process(process, data_bytes, stdout == subprocess.PIPE, stderr == subprocess.PIPE, capture=capture, output_stream=output_stream) stdout_text = to_optional_text(stdout_bytes, str_errors) or '' stderr_text = to_optional_text(stderr_bytes, str_errors) or '' else: process.wait() stdout_text, stderr_text = None, None finally: if process and process.returncode is None: process.kill() display.info('') # the process we're interrupting may have completed a partial line of output display.notice('Killed command to avoid an orphaned child process during handling of an unexpected exception.') status = process.returncode runtime = time.time() - start display.info('Command exited with status %s after %s seconds.' % (status, runtime), verbosity=4) if status == 0: return stdout_text, stderr_text raise SubprocessError(cmd, status, stdout_text, stderr_text, runtime, error_callback) def communicate_with_process( process: subprocess.Popen, stdin: t.Optional[bytes], stdout: bool, stderr: bool, capture: bool, output_stream: OutputStream, ) -> tuple[bytes, bytes]: """Communicate with the specified process, handling stdin/stdout/stderr as requested.""" threads: list[WrappedThread] = [] reader: t.Type[ReaderThread] if capture: reader = CaptureThread else: reader = OutputThread if stdin is not None: threads.append(WriterThread(process.stdin, stdin)) if stdout: stdout_reader = reader(process.stdout, output_stream.get_buffer(sys.stdout.buffer)) threads.append(stdout_reader) else: stdout_reader = None if stderr: stderr_reader = reader(process.stderr, output_stream.get_buffer(sys.stderr.buffer)) threads.append(stderr_reader) else: stderr_reader = None for thread in threads: thread.start() for thread in threads: try: thread.wait_for_result() except Exception as ex: # pylint: disable=broad-except display.error(str(ex)) if isinstance(stdout_reader, ReaderThread): stdout_bytes = b''.join(stdout_reader.lines) else: stdout_bytes = b'' if isinstance(stderr_reader, ReaderThread): stderr_bytes = b''.join(stderr_reader.lines) else: stderr_bytes = b'' process.wait() return stdout_bytes, stderr_bytes class WriterThread(WrappedThread): """Thread to write data to stdin of a subprocess.""" def __init__(self, handle: t.IO[bytes], data: bytes) -> None: super().__init__(self._run) self.handle = handle self.data = data def _run(self) -> None: """Workload to run on a thread.""" try: self.handle.write(self.data) self.handle.flush() finally: self.handle.close() class ReaderThread(WrappedThread, metaclass=abc.ABCMeta): """Thread to read stdout from a subprocess.""" def __init__(self, handle: t.IO[bytes], buffer: t.BinaryIO) -> None: super().__init__(self._run) self.handle = handle self.buffer = buffer self.lines: list[bytes] = [] @abc.abstractmethod def _run(self) -> None: """Workload to run on a thread.""" class CaptureThread(ReaderThread): """Thread to capture stdout from a subprocess into a buffer.""" def _run(self) -> None: """Workload to run on a thread.""" src = self.handle dst = self.lines try: for line in src: dst.append(line) finally: src.close() class OutputThread(ReaderThread): """Thread to pass stdout from a subprocess to stdout.""" def _run(self) -> None: """Workload to run on a thread.""" src = self.handle dst = self.buffer try: for line in src: dst.write(line) dst.flush() finally: src.close() def common_environment() -> dict[str, str]: """Common environment used for executing all programs.""" env = dict( LC_ALL=CONFIGURED_LOCALE, PATH=os.environ.get('PATH', os.path.defpath), ) required = ( 'HOME', ) optional = ( 'LD_LIBRARY_PATH', 'SSH_AUTH_SOCK', # MacOS High Sierra Compatibility # http://sealiesoftware.com/blog/archive/2017/6/5/Objective-C_and_fork_in_macOS_1013.html # Example configuration for macOS: # export OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES 'OBJC_DISABLE_INITIALIZE_FORK_SAFETY', 'ANSIBLE_KEEP_REMOTE_FILES', # MacOS Homebrew Compatibility # https://cryptography.io/en/latest/installation/#building-cryptography-on-macos # This may also be required to install pyyaml with libyaml support when installed in non-standard locations. # Example configuration for brew on macOS: # export LDFLAGS="-L$(brew --prefix openssl)/lib/ -L$(brew --prefix libyaml)/lib/" # export CFLAGS="-I$(brew --prefix openssl)/include/ -I$(brew --prefix libyaml)/include/" 'LDFLAGS', 'CFLAGS', ) # FreeBSD Compatibility # This is required to include libyaml support in PyYAML. # The header /usr/local/include/yaml.h isn't in the default include path for the compiler. # It is included here so that tests can take advantage of it, rather than only ansible-test during managed pip installs. # If CFLAGS has been set in the environment that value will take precedence due to being an optional var when calling pass_vars. if os.path.exists('/etc/freebsd-update.conf'): env.update(CFLAGS='-I/usr/local/include/') env.update(pass_vars(required=required, optional=optional)) return env def report_locale(show_warning: bool) -> None: """Report the configured locale and the locale warning, if applicable.""" display.info(f'Configured locale: {CONFIGURED_LOCALE}', verbosity=1) if LOCALE_WARNING and show_warning: display.warning(LOCALE_WARNING) def pass_vars(required: c.Collection[str], optional: c.Collection[str]) -> dict[str, str]: """Return a filtered dictionary of environment variables based on the current environment.""" env = {} for name in required: if name not in os.environ: raise MissingEnvironmentVariable(name) env[name] = os.environ[name] for name in optional: if name not in os.environ: continue env[name] = os.environ[name] return env def verified_chmod(path: str, mode: int) -> None: """Perform chmod on the specified path and then verify the permissions were applied.""" os.chmod(path, mode) # pylint: disable=ansible-bad-function executable = any(mode & perm for perm in (stat.S_IXUSR, stat.S_IXGRP, stat.S_IXOTH)) if executable and not os.access(path, os.X_OK): raise ApplicationError(f'Path "{path}" should executable, but is not. Is the filesystem mounted with the "noexec" option?') def remove_tree(path: str) -> None: """Remove the specified directory, silently continuing if the directory does not exist.""" try: shutil.rmtree(to_bytes(path)) except FileNotFoundError: pass def is_binary_file(path: str) -> bool: """Return True if the specified file is a binary file, otherwise return False.""" assume_text = { '.cfg', '.conf', '.crt', '.cs', '.css', '.html', '.ini', '.j2', '.js', '.json', '.md', '.pem', '.ps1', '.psm1', '.py', '.rst', '.sh', '.txt', '.xml', '.yaml', '.yml', } assume_binary = { '.bin', '.eot', '.gz', '.ico', '.iso', '.jpg', '.otf', '.p12', '.png', '.pyc', '.rpm', '.ttf', '.woff', '.woff2', '.zip', } ext = os.path.splitext(path)[1] if ext in assume_text: return False if ext in assume_binary: return True with open_binary_file(path) as path_fd: return b'\0' in path_fd.read(4096) def generate_name(length: int = 8) -> str: """Generate and return a random name.""" return ''.join(random.choice(string.ascii_letters + string.digits) for _idx in range(length)) def generate_password() -> str: """Generate and return random password.""" chars = [ string.ascii_letters, string.digits, string.ascii_letters, string.digits, '-', ] * 4 password = ''.join([random.choice(char) for char in chars[:-1]]) display.sensitive.add(password) return password class Display: """Manages color console output.""" clear = '\033[0m' red = '\033[31m' green = '\033[32m' yellow = '\033[33m' blue = '\033[34m' purple = '\033[35m' cyan = '\033[36m' verbosity_colors = { 0: None, 1: green, 2: blue, 3: cyan, } def __init__(self) -> None: self.verbosity = 0 self.color = sys.stdout.isatty() self.warnings: list[str] = [] self.warnings_unique: set[str] = set() self.fd = sys.stderr # default to stderr until config is initialized to avoid early messages going to stdout self.rows = 0 self.columns = 0 self.truncate = 0 self.redact = True self.sensitive: set[str] = set() if os.isatty(0): self.rows, self.columns = unpack('HHHH', fcntl.ioctl(0, TIOCGWINSZ, pack('HHHH', 0, 0, 0, 0)))[:2] def __warning(self, message: str) -> None: """Internal implementation for displaying a warning message.""" self.print_message('WARNING: %s' % message, color=self.purple) def review_warnings(self) -> None: """Review all warnings which previously occurred.""" if not self.warnings: return self.__warning('Reviewing previous %d warning(s):' % len(self.warnings)) for warning in self.warnings: self.__warning(warning) def warning(self, message: str, unique: bool = False, verbosity: int = 0) -> None: """Display a warning level message.""" if verbosity > self.verbosity: return if unique: if message in self.warnings_unique: return self.warnings_unique.add(message) self.__warning(message) self.warnings.append(message) def notice(self, message: str) -> None: """Display a notice level message.""" self.print_message('NOTICE: %s' % message, color=self.purple) def error(self, message: str) -> None: """Display an error level message.""" self.print_message('ERROR: %s' % message, color=self.red) def fatal(self, message: str) -> None: """Display a fatal level message.""" self.print_message('FATAL: %s' % message, color=self.red, stderr=True) def info(self, message: str, verbosity: int = 0, truncate: bool = False) -> None: """Display an info level message.""" if self.verbosity >= verbosity: color = self.verbosity_colors.get(verbosity, self.yellow) self.print_message(message, color=color, truncate=truncate) def print_message( # pylint: disable=locally-disabled, invalid-name self, message: str, color: t.Optional[str] = None, stderr: bool = False, truncate: bool = False, ) -> None: """Display a message.""" if self.redact and self.sensitive: for item in self.sensitive: if not item: continue message = message.replace(item, '*' * len(item)) if truncate: if len(message) > self.truncate > 5: message = message[:self.truncate - 5] + ' ...' if color and self.color: # convert color resets in message to desired color message = message.replace(self.clear, color) message = '%s%s%s' % (color, message, self.clear) fd = sys.stderr if stderr else self.fd print(message, file=fd) fd.flush() class InternalError(Exception): """An unhandled internal error indicating a bug in the code.""" def __init__(self, message: str) -> None: super().__init__(f'An internal error has occurred in ansible-test: {message}') class ApplicationError(Exception): """General application error.""" class ApplicationWarning(Exception): """General application warning which interrupts normal program flow.""" class TimeoutExpiredError(SystemExit): """Error raised when the test timeout has been reached or exceeded.""" class SubprocessError(ApplicationError): """Error resulting from failed subprocess execution.""" def __init__( self, cmd: list[str], status: int = 0, stdout: t.Optional[str] = None, stderr: t.Optional[str] = None, runtime: t.Optional[float] = None, error_callback: t.Optional[c.Callable[[SubprocessError], None]] = None, ) -> None: message = 'Command "%s" returned exit status %s.\n' % (shlex.join(cmd), status) message += format_command_output(stdout, stderr) self.cmd = cmd self.message = message self.status = status self.stdout = stdout self.stderr = stderr self.runtime = runtime if error_callback: error_callback(self) self.message = self.message.strip() super().__init__(self.message) class MissingEnvironmentVariable(ApplicationError): """Error caused by missing environment variable.""" def __init__(self, name: str) -> None: super().__init__('Missing environment variable: %s' % name) self.name = name class HostConnectionError(ApplicationError): """ Raised when the initial connection during host profile setup has failed and all retries have been exhausted. Raised by provisioning code when one or more provisioning threads raise this exception. Also raised when an SSH connection fails for the shell command. """ def __init__(self, message: str, callback: t.Callable[[], None] = None) -> None: super().__init__(message) self._callback = callback def run_callback(self) -> None: """Run the error callback, if any.""" if self._callback: self._callback() def format_command_output(stdout: str, stderr: str) -> str: """Return a formatted string containing the given stdout and stderr (if any).""" message = '' if stderr := stderr.strip(): message += '>>> Standard Error\n' message += f'{stderr}{Display.clear}\n' if stdout := stdout.strip(): message += '>>> Standard Output\n' message += f'{stdout}{Display.clear}\n' return message def retry(func: t.Callable[..., TValue], ex_type: t.Type[BaseException] = SubprocessError, sleep: int = 10, attempts: int = 10, warn: bool = True) -> TValue: """Retry the specified function on failure.""" for dummy in range(1, attempts): try: return func() except ex_type as ex: if warn: display.warning(str(ex)) time.sleep(sleep) return func() def parse_to_list_of_dict(pattern: str, value: str) -> list[dict[str, str]]: """Parse lines from the given value using the specified pattern and return the extracted list of key/value pair dictionaries.""" matched = [] unmatched: list[str] = [] for line in value.splitlines(): match = re.search(pattern, line) if match: matched.append(match.groupdict()) else: unmatched.append(line) if unmatched: raise Exception('Pattern "%s" did not match values:\n%s' % (pattern, '\n'.join(unmatched))) return matched def get_subclasses(class_type: t.Type[C]) -> list[t.Type[C]]: """Returns a list of types that are concrete subclasses of the given type.""" subclasses: set[t.Type[C]] = set() queue: list[t.Type[C]] = [class_type] while queue: parent = queue.pop() for child in parent.__subclasses__(): if child not in subclasses: if not inspect.isabstract(child): subclasses.add(child) queue.append(child) return sorted(subclasses, key=lambda sc: sc.__name__) def is_subdir(candidate_path: str, path: str) -> bool: """Returns true if candidate_path is path or a subdirectory of path.""" if not path.endswith(os.path.sep): path += os.path.sep if not candidate_path.endswith(os.path.sep): candidate_path += os.path.sep return candidate_path.startswith(path) def paths_to_dirs(paths: list[str]) -> list[str]: """Returns a list of directories extracted from the given list of paths.""" dir_names = set() for path in paths: while True: path = os.path.dirname(path) if not path or path == os.path.sep: break dir_names.add(path + os.path.sep) return sorted(dir_names) def str_to_version(version: str) -> tuple[int, ...]: """Return a version tuple from a version string.""" return tuple(int(n) for n in version.split('.')) def version_to_str(version: tuple[int, ...]) -> str: """Return a version string from a version tuple.""" return '.'.join(str(n) for n in version) def sorted_versions(versions: list[str]) -> list[str]: """Return a sorted copy of the given list of versions.""" return [version_to_str(version) for version in sorted(str_to_version(version) for version in versions)] def import_plugins(directory: str, root: t.Optional[str] = None) -> None: """ Import plugins from the given directory relative to the given root. If the root is not provided, the 'lib' directory for the test runner will be used. """ if root is None: root = os.path.dirname(__file__) path = os.path.join(root, directory) package = __name__.rsplit('.', 1)[0] prefix = '%s.%s.' % (package, directory.replace(os.path.sep, '.')) for (_module_loader, name, _ispkg) in pkgutil.iter_modules([path], prefix=prefix): module_path = os.path.join(root, name[len(package) + 1:].replace('.', os.path.sep) + '.py') load_module(module_path, name) def load_plugins(base_type: t.Type[C], database: dict[str, t.Type[C]]) -> None: """ Load plugins of the specified type and track them in the specified database. Only plugins which have already been imported will be loaded. """ plugins: dict[str, t.Type[C]] = dict((sc.__module__.rsplit('.', 1)[1], sc) for sc in get_subclasses(base_type)) for plugin in plugins: database[plugin] = plugins[plugin] def load_module(path: str, name: str) -> None: """Load a Python module using the given name and path.""" if name in sys.modules: return spec = importlib.util.spec_from_file_location(name, path) module = importlib.util.module_from_spec(spec) sys.modules[name] = module spec.loader.exec_module(module) def sanitize_host_name(name: str) -> str: """Return a sanitized version of the given name, suitable for use as a hostname.""" return re.sub('[^A-Za-z0-9]+', '-', name)[:63].strip('-') def get_generic_type(base_type: t.Type, generic_base_type: t.Type[TValue]) -> t.Optional[t.Type[TValue]]: """Return the generic type arg derived from the generic_base_type type that is associated with the base_type type, if any, otherwise return None.""" # noinspection PyUnresolvedReferences type_arg = t.get_args(base_type.__orig_bases__[0])[0] return None if isinstance(type_arg, generic_base_type) else type_arg def get_type_associations(base_type: t.Type[TBase], generic_base_type: t.Type[TValue]) -> list[tuple[t.Type[TValue], t.Type[TBase]]]: """Create and return a list of tuples associating generic_base_type derived types with a corresponding base_type derived type.""" return [item for item in [(get_generic_type(sc_type, generic_base_type), sc_type) for sc_type in get_subclasses(base_type)] if item[1]] def get_type_map(base_type: t.Type[TBase], generic_base_type: t.Type[TValue]) -> dict[t.Type[TValue], t.Type[TBase]]: """Create and return a mapping of generic_base_type derived types to base_type derived types.""" return {item[0]: item[1] for item in get_type_associations(base_type, generic_base_type)} def verify_sys_executable(path: str) -> t.Optional[str]: """Verify that the given path references the current Python interpreter. If not, return the expected path, otherwise return None.""" if path == sys.executable: return None if os.path.realpath(path) == os.path.realpath(sys.executable): return None expected_executable = raw_command([path, '-c', 'import sys; print(sys.executable)'], capture=True)[0] if expected_executable == sys.executable: return None return expected_executable def type_guard(sequence: c.Sequence[t.Any], guard_type: t.Type[C]) -> t.TypeGuard[c.Sequence[C]]: """ Raises an exception if any item in the given sequence does not match the specified guard type. Use with assert so that type checkers are aware of the type guard. """ invalid_types = set(type(item) for item in sequence if not isinstance(item, guard_type)) if not invalid_types: return True invalid_type_names = sorted(str(item) for item in invalid_types) raise Exception(f'Sequence required to contain only {guard_type} includes: {", ".join(invalid_type_names)}') display = Display() # pylint: disable=locally-disabled, invalid-name _enable_vendoring()