mirror of https://github.com/ansible/ansible.git
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1174 lines
37 KiB
Python
1174 lines
37 KiB
Python
"""Miscellaneous utility functions and classes."""
|
|
from __future__ import annotations
|
|
|
|
import abc
|
|
import collections.abc as c
|
|
import enum
|
|
import fcntl
|
|
import importlib.util
|
|
import inspect
|
|
import json
|
|
import keyword
|
|
import os
|
|
import platform
|
|
import pkgutil
|
|
import random
|
|
import re
|
|
import shutil
|
|
import stat
|
|
import string
|
|
import subprocess
|
|
import sys
|
|
import time
|
|
import functools
|
|
import shlex
|
|
import typing as t
|
|
import warnings
|
|
|
|
from struct import unpack, pack
|
|
from termios import TIOCGWINSZ
|
|
|
|
# CAUTION: Avoid third-party imports in this module whenever possible.
|
|
# Any third-party imports occurring here will result in an error if they are vendored by ansible-core.
|
|
|
|
try:
|
|
from typing_extensions import TypeGuard # TypeGuard was added in Python 3.10
|
|
except ImportError:
|
|
TypeGuard = None
|
|
|
|
from .locale_util import (
|
|
LOCALE_WARNING,
|
|
CONFIGURED_LOCALE,
|
|
)
|
|
|
|
from .encoding import (
|
|
to_bytes,
|
|
to_optional_bytes,
|
|
to_optional_text,
|
|
)
|
|
|
|
from .io import (
|
|
open_binary_file,
|
|
read_text_file,
|
|
)
|
|
|
|
from .thread import (
|
|
mutex,
|
|
WrappedThread,
|
|
)
|
|
|
|
from .constants import (
|
|
SUPPORTED_PYTHON_VERSIONS,
|
|
)
|
|
|
|
C = t.TypeVar('C')
|
|
TBase = t.TypeVar('TBase')
|
|
TKey = t.TypeVar('TKey')
|
|
TValue = t.TypeVar('TValue')
|
|
|
|
PYTHON_PATHS: dict[str, str] = {}
|
|
|
|
COVERAGE_CONFIG_NAME = 'coveragerc'
|
|
|
|
ANSIBLE_TEST_ROOT = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
|
|
|
|
# assume running from install
|
|
ANSIBLE_ROOT = os.path.dirname(ANSIBLE_TEST_ROOT)
|
|
ANSIBLE_BIN_PATH = os.path.dirname(os.path.abspath(sys.argv[0]))
|
|
ANSIBLE_LIB_ROOT = os.path.join(ANSIBLE_ROOT, 'ansible')
|
|
ANSIBLE_SOURCE_ROOT = None
|
|
|
|
if not os.path.exists(ANSIBLE_LIB_ROOT):
|
|
# running from source
|
|
ANSIBLE_ROOT = os.path.dirname(os.path.dirname(os.path.dirname(ANSIBLE_TEST_ROOT)))
|
|
ANSIBLE_BIN_PATH = os.path.join(ANSIBLE_ROOT, 'bin')
|
|
ANSIBLE_LIB_ROOT = os.path.join(ANSIBLE_ROOT, 'lib', 'ansible')
|
|
ANSIBLE_SOURCE_ROOT = ANSIBLE_ROOT
|
|
|
|
ANSIBLE_TEST_DATA_ROOT = os.path.join(ANSIBLE_TEST_ROOT, '_data')
|
|
ANSIBLE_TEST_UTIL_ROOT = os.path.join(ANSIBLE_TEST_ROOT, '_util')
|
|
ANSIBLE_TEST_CONFIG_ROOT = os.path.join(ANSIBLE_TEST_ROOT, 'config')
|
|
|
|
ANSIBLE_TEST_CONTROLLER_ROOT = os.path.join(ANSIBLE_TEST_UTIL_ROOT, 'controller')
|
|
ANSIBLE_TEST_TARGET_ROOT = os.path.join(ANSIBLE_TEST_UTIL_ROOT, 'target')
|
|
|
|
ANSIBLE_TEST_TOOLS_ROOT = os.path.join(ANSIBLE_TEST_CONTROLLER_ROOT, 'tools')
|
|
ANSIBLE_TEST_TARGET_TOOLS_ROOT = os.path.join(ANSIBLE_TEST_TARGET_ROOT, 'tools')
|
|
|
|
# Modes are set to allow all users the same level of access.
|
|
# This permits files to be used in tests that change users.
|
|
# The only exception is write access to directories for the user creating them.
|
|
# This avoids having to modify the directory permissions a second time.
|
|
|
|
MODE_READ = stat.S_IRUSR | stat.S_IRGRP | stat.S_IROTH
|
|
|
|
MODE_FILE = MODE_READ
|
|
MODE_FILE_EXECUTE = MODE_FILE | stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH
|
|
MODE_FILE_WRITE = MODE_FILE | stat.S_IWUSR | stat.S_IWGRP | stat.S_IWOTH
|
|
|
|
MODE_DIRECTORY = MODE_READ | stat.S_IWUSR | stat.S_IXUSR | stat.S_IXGRP | stat.S_IXOTH
|
|
MODE_DIRECTORY_WRITE = MODE_DIRECTORY | stat.S_IWGRP | stat.S_IWOTH
|
|
|
|
|
|
class OutputStream(enum.Enum):
|
|
"""The output stream to use when running a subprocess and redirecting/capturing stdout or stderr."""
|
|
|
|
ORIGINAL = enum.auto()
|
|
AUTO = enum.auto()
|
|
|
|
def get_buffer(self, original: t.BinaryIO) -> t.BinaryIO:
|
|
"""Return the correct output buffer to use, taking into account the given original buffer."""
|
|
|
|
if self == OutputStream.ORIGINAL:
|
|
return original
|
|
|
|
if self == OutputStream.AUTO:
|
|
return display.fd.buffer
|
|
|
|
raise NotImplementedError(str(self))
|
|
|
|
|
|
class Architecture:
|
|
"""
|
|
Normalized architecture names.
|
|
These are the architectures supported by ansible-test, such as when provisioning remote instances.
|
|
"""
|
|
|
|
X86_64 = 'x86_64'
|
|
AARCH64 = 'aarch64'
|
|
|
|
|
|
REMOTE_ARCHITECTURES = list(value for key, value in Architecture.__dict__.items() if not key.startswith('__'))
|
|
|
|
|
|
def is_valid_identifier(value: str) -> bool:
|
|
"""Return True if the given value is a valid non-keyword Python identifier, otherwise return False."""
|
|
return value.isidentifier() and not keyword.iskeyword(value)
|
|
|
|
|
|
def cache(func: c.Callable[[], TValue]) -> c.Callable[[], TValue]:
|
|
"""Enforce exclusive access on a decorated function and cache the result."""
|
|
storage: dict[None, TValue] = {}
|
|
sentinel = object()
|
|
|
|
@functools.wraps(func)
|
|
def cache_func():
|
|
"""Cache the return value from func."""
|
|
if (value := storage.get(None, sentinel)) is sentinel:
|
|
value = storage[None] = func()
|
|
|
|
return value
|
|
|
|
wrapper = mutex(cache_func)
|
|
|
|
return wrapper
|
|
|
|
|
|
@mutex
|
|
def detect_architecture(python: str) -> t.Optional[str]:
|
|
"""Detect the architecture of the specified Python and return a normalized version, or None if it cannot be determined."""
|
|
results: dict[str, t.Optional[str]]
|
|
|
|
try:
|
|
results = detect_architecture.results # type: ignore[attr-defined]
|
|
except AttributeError:
|
|
results = detect_architecture.results = {} # type: ignore[attr-defined]
|
|
|
|
if python in results:
|
|
return results[python]
|
|
|
|
if python == sys.executable or os.path.realpath(python) == os.path.realpath(sys.executable):
|
|
uname = platform.uname()
|
|
else:
|
|
data = raw_command([python, '-c', 'import json, platform; print(json.dumps(platform.uname()));'], capture=True)[0]
|
|
uname = json.loads(data)
|
|
|
|
translation = {
|
|
'x86_64': Architecture.X86_64, # Linux, macOS
|
|
'amd64': Architecture.X86_64, # FreeBSD
|
|
'aarch64': Architecture.AARCH64, # Linux, FreeBSD
|
|
'arm64': Architecture.AARCH64, # FreeBSD
|
|
}
|
|
|
|
candidates = []
|
|
|
|
if len(uname) >= 5:
|
|
candidates.append(uname[4])
|
|
|
|
if len(uname) >= 6:
|
|
candidates.append(uname[5])
|
|
|
|
candidates = sorted(set(candidates))
|
|
architectures = sorted(set(arch for arch in [translation.get(candidate) for candidate in candidates] if arch))
|
|
|
|
architecture: t.Optional[str] = None
|
|
|
|
if not architectures:
|
|
display.warning(f'Unable to determine architecture for Python interpreter "{python}" from: {candidates}')
|
|
elif len(architectures) == 1:
|
|
architecture = architectures[0]
|
|
display.info(f'Detected architecture {architecture} for Python interpreter: {python}', verbosity=1)
|
|
else:
|
|
display.warning(f'Conflicting architectures detected ({architectures}) for Python interpreter "{python}" from: {candidates}')
|
|
|
|
results[python] = architecture
|
|
|
|
return architecture
|
|
|
|
|
|
def filter_args(args: list[str], filters: dict[str, int]) -> list[str]:
|
|
"""Return a filtered version of the given command line arguments."""
|
|
remaining = 0
|
|
result = []
|
|
|
|
for arg in args:
|
|
if not arg.startswith('-') and remaining:
|
|
remaining -= 1
|
|
continue
|
|
|
|
remaining = 0
|
|
|
|
parts = arg.split('=', 1)
|
|
key = parts[0]
|
|
|
|
if key in filters:
|
|
remaining = filters[key] - len(parts) + 1
|
|
continue
|
|
|
|
result.append(arg)
|
|
|
|
return result
|
|
|
|
|
|
def read_lines_without_comments(path: str, remove_blank_lines: bool = False, optional: bool = False) -> list[str]:
|
|
"""
|
|
Returns lines from the specified text file with comments removed.
|
|
Comments are any content from a hash symbol to the end of a line.
|
|
Any spaces immediately before a comment are also removed.
|
|
"""
|
|
if optional and not os.path.exists(path):
|
|
return []
|
|
|
|
lines = read_text_file(path).splitlines()
|
|
|
|
lines = [re.sub(r' *#.*$', '', line) for line in lines]
|
|
|
|
if remove_blank_lines:
|
|
lines = [line for line in lines if line]
|
|
|
|
return lines
|
|
|
|
|
|
def exclude_none_values(data: dict[TKey, t.Optional[TValue]]) -> dict[TKey, TValue]:
|
|
"""Return the provided dictionary with any None values excluded."""
|
|
return dict((key, value) for key, value in data.items() if value is not None)
|
|
|
|
|
|
def find_executable(executable: str, cwd: t.Optional[str] = None, path: t.Optional[str] = None, required: t.Union[bool, str] = True) -> t.Optional[str]:
|
|
"""
|
|
Find the specified executable and return the full path, or None if it could not be found.
|
|
If required is True an exception will be raised if the executable is not found.
|
|
If required is set to 'warning' then a warning will be shown if the executable is not found.
|
|
"""
|
|
match = None
|
|
real_cwd = os.getcwd()
|
|
|
|
if not cwd:
|
|
cwd = real_cwd
|
|
|
|
if os.path.dirname(executable):
|
|
target = os.path.join(cwd, executable)
|
|
if os.path.exists(target) and os.access(target, os.F_OK | os.X_OK):
|
|
match = executable
|
|
else:
|
|
if path is None:
|
|
path = os.environ.get('PATH', os.path.defpath)
|
|
|
|
if path:
|
|
path_dirs = path.split(os.path.pathsep)
|
|
seen_dirs = set()
|
|
|
|
for path_dir in path_dirs:
|
|
if path_dir in seen_dirs:
|
|
continue
|
|
|
|
seen_dirs.add(path_dir)
|
|
|
|
if os.path.abspath(path_dir) == real_cwd:
|
|
path_dir = cwd
|
|
|
|
candidate = os.path.join(path_dir, executable)
|
|
|
|
if os.path.exists(candidate) and os.access(candidate, os.F_OK | os.X_OK):
|
|
match = candidate
|
|
break
|
|
|
|
if not match and required:
|
|
message = 'Required program "%s" not found.' % executable
|
|
|
|
if required != 'warning':
|
|
raise ApplicationError(message)
|
|
|
|
display.warning(message)
|
|
|
|
return match
|
|
|
|
|
|
def find_python(version: str, path: t.Optional[str] = None, required: bool = True) -> t.Optional[str]:
|
|
"""
|
|
Find and return the full path to the specified Python version.
|
|
If required, an exception will be raised not found.
|
|
If not required, None will be returned if not found.
|
|
"""
|
|
version_info = str_to_version(version)
|
|
|
|
if not path and version_info == sys.version_info[:len(version_info)]:
|
|
python_bin = sys.executable
|
|
else:
|
|
python_bin = find_executable('python%s' % version, path=path, required=required)
|
|
|
|
return python_bin
|
|
|
|
|
|
@cache
|
|
def get_ansible_version() -> str:
|
|
"""Return the Ansible version."""
|
|
# ansible may not be in our sys.path
|
|
# avoids a symlink to release.py since ansible placement relative to ansible-test may change during delegation
|
|
load_module(os.path.join(ANSIBLE_LIB_ROOT, 'release.py'), 'ansible_release')
|
|
|
|
# noinspection PyUnresolvedReferences
|
|
from ansible_release import __version__ as ansible_version # pylint: disable=import-error
|
|
|
|
return ansible_version
|
|
|
|
|
|
def _enable_vendoring() -> None:
|
|
"""Enable support for loading Python packages vendored by ansible-core."""
|
|
# Load the vendoring code by file path, since ansible may not be in our sys.path.
|
|
# Convert warnings into errors, to avoid problems from surfacing later.
|
|
|
|
with warnings.catch_warnings():
|
|
warnings.filterwarnings('error')
|
|
|
|
load_module(os.path.join(ANSIBLE_LIB_ROOT, '_vendor', '__init__.py'), 'ansible_vendor')
|
|
|
|
|
|
@cache
|
|
def get_available_python_versions() -> dict[str, str]:
|
|
"""Return a dictionary indicating which supported Python versions are available."""
|
|
return dict((version, path) for version, path in ((version, find_python(version, required=False)) for version in SUPPORTED_PYTHON_VERSIONS) if path)
|
|
|
|
|
|
def raw_command(
|
|
cmd: c.Iterable[str],
|
|
capture: bool,
|
|
env: t.Optional[dict[str, str]] = None,
|
|
data: t.Optional[str] = None,
|
|
cwd: t.Optional[str] = None,
|
|
explain: bool = False,
|
|
stdin: t.Optional[t.Union[t.IO[bytes], int]] = None,
|
|
stdout: t.Optional[t.Union[t.IO[bytes], int]] = None,
|
|
interactive: bool = False,
|
|
output_stream: t.Optional[OutputStream] = None,
|
|
cmd_verbosity: int = 1,
|
|
str_errors: str = 'strict',
|
|
error_callback: t.Optional[c.Callable[[SubprocessError], None]] = None,
|
|
) -> tuple[t.Optional[str], t.Optional[str]]:
|
|
"""Run the specified command and return stdout and stderr as a tuple."""
|
|
output_stream = output_stream or OutputStream.AUTO
|
|
|
|
if capture and interactive:
|
|
raise InternalError('Cannot combine capture=True with interactive=True.')
|
|
|
|
if data and interactive:
|
|
raise InternalError('Cannot combine data with interactive=True.')
|
|
|
|
if stdin and interactive:
|
|
raise InternalError('Cannot combine stdin with interactive=True.')
|
|
|
|
if stdout and interactive:
|
|
raise InternalError('Cannot combine stdout with interactive=True.')
|
|
|
|
if stdin and data:
|
|
raise InternalError('Cannot combine stdin with data.')
|
|
|
|
if stdout and not capture:
|
|
raise InternalError('Redirection of stdout requires capture=True to avoid redirection of stderr to stdout.')
|
|
|
|
if output_stream != OutputStream.AUTO and capture:
|
|
raise InternalError(f'Cannot combine {output_stream=} with capture=True.')
|
|
|
|
if output_stream != OutputStream.AUTO and interactive:
|
|
raise InternalError(f'Cannot combine {output_stream=} with interactive=True.')
|
|
|
|
if not cwd:
|
|
cwd = os.getcwd()
|
|
|
|
if not env:
|
|
env = common_environment()
|
|
|
|
cmd = list(cmd)
|
|
|
|
escaped_cmd = shlex.join(cmd)
|
|
|
|
if capture:
|
|
description = 'Run'
|
|
elif interactive:
|
|
description = 'Interactive'
|
|
else:
|
|
description = 'Stream'
|
|
|
|
description += ' command'
|
|
|
|
with_types = []
|
|
|
|
if data:
|
|
with_types.append('data')
|
|
|
|
if stdin:
|
|
with_types.append('stdin')
|
|
|
|
if stdout:
|
|
with_types.append('stdout')
|
|
|
|
if with_types:
|
|
description += f' with {"/".join(with_types)}'
|
|
|
|
display.info(f'{description}: {escaped_cmd}', verbosity=cmd_verbosity, truncate=True)
|
|
display.info('Working directory: %s' % cwd, verbosity=2)
|
|
|
|
program = find_executable(cmd[0], cwd=cwd, path=env['PATH'], required='warning')
|
|
|
|
if program:
|
|
display.info('Program found: %s' % program, verbosity=2)
|
|
|
|
for key in sorted(env.keys()):
|
|
display.info('%s=%s' % (key, env[key]), verbosity=2)
|
|
|
|
if explain:
|
|
return None, None
|
|
|
|
communicate = False
|
|
|
|
if stdin is not None:
|
|
data = None
|
|
elif data is not None:
|
|
stdin = subprocess.PIPE
|
|
communicate = True
|
|
elif interactive:
|
|
pass # allow the subprocess access to our stdin
|
|
else:
|
|
stdin = subprocess.DEVNULL
|
|
|
|
if not interactive:
|
|
# When not running interactively, send subprocess stdout/stderr through a pipe.
|
|
# This isolates the stdout/stderr of the subprocess from the current process, and also hides the current TTY from it, if any.
|
|
# This prevents subprocesses from sharing stdout/stderr with the current process or each other.
|
|
# Doing so allows subprocesses to safely make changes to their file handles, such as making them non-blocking (ssh does this).
|
|
# This also maintains consistency between local testing and CI systems, which typically do not provide a TTY.
|
|
# To maintain output ordering, a single pipe is used for both stdout/stderr when not capturing output unless the output stream is ORIGINAL.
|
|
stdout = stdout or subprocess.PIPE
|
|
stderr = subprocess.PIPE if capture or output_stream == OutputStream.ORIGINAL else subprocess.STDOUT
|
|
communicate = True
|
|
else:
|
|
stderr = None
|
|
|
|
start = time.time()
|
|
process = None
|
|
|
|
try:
|
|
try:
|
|
cmd_bytes = [to_bytes(arg) for arg in cmd]
|
|
env_bytes = dict((to_bytes(k), to_bytes(v)) for k, v in env.items())
|
|
process = subprocess.Popen(cmd_bytes, env=env_bytes, stdin=stdin, stdout=stdout, stderr=stderr, cwd=cwd) # pylint: disable=consider-using-with
|
|
except FileNotFoundError as ex:
|
|
raise ApplicationError('Required program "%s" not found.' % cmd[0]) from ex
|
|
|
|
if communicate:
|
|
data_bytes = to_optional_bytes(data)
|
|
stdout_bytes, stderr_bytes = communicate_with_process(process, data_bytes, stdout == subprocess.PIPE, stderr == subprocess.PIPE, capture=capture,
|
|
output_stream=output_stream)
|
|
stdout_text = to_optional_text(stdout_bytes, str_errors) or ''
|
|
stderr_text = to_optional_text(stderr_bytes, str_errors) or ''
|
|
else:
|
|
process.wait()
|
|
stdout_text, stderr_text = None, None
|
|
finally:
|
|
if process and process.returncode is None:
|
|
process.kill()
|
|
display.info('') # the process we're interrupting may have completed a partial line of output
|
|
display.notice('Killed command to avoid an orphaned child process during handling of an unexpected exception.')
|
|
|
|
status = process.returncode
|
|
runtime = time.time() - start
|
|
|
|
display.info('Command exited with status %s after %s seconds.' % (status, runtime), verbosity=4)
|
|
|
|
if status == 0:
|
|
return stdout_text, stderr_text
|
|
|
|
raise SubprocessError(cmd, status, stdout_text, stderr_text, runtime, error_callback)
|
|
|
|
|
|
def communicate_with_process(
|
|
process: subprocess.Popen,
|
|
stdin: t.Optional[bytes],
|
|
stdout: bool,
|
|
stderr: bool,
|
|
capture: bool,
|
|
output_stream: OutputStream,
|
|
) -> tuple[bytes, bytes]:
|
|
"""Communicate with the specified process, handling stdin/stdout/stderr as requested."""
|
|
threads: list[WrappedThread] = []
|
|
reader: t.Type[ReaderThread]
|
|
|
|
if capture:
|
|
reader = CaptureThread
|
|
else:
|
|
reader = OutputThread
|
|
|
|
if stdin is not None:
|
|
threads.append(WriterThread(process.stdin, stdin))
|
|
|
|
if stdout:
|
|
stdout_reader = reader(process.stdout, output_stream.get_buffer(sys.stdout.buffer))
|
|
threads.append(stdout_reader)
|
|
else:
|
|
stdout_reader = None
|
|
|
|
if stderr:
|
|
stderr_reader = reader(process.stderr, output_stream.get_buffer(sys.stderr.buffer))
|
|
threads.append(stderr_reader)
|
|
else:
|
|
stderr_reader = None
|
|
|
|
for thread in threads:
|
|
thread.start()
|
|
|
|
for thread in threads:
|
|
try:
|
|
thread.wait_for_result()
|
|
except Exception as ex: # pylint: disable=broad-except
|
|
display.error(str(ex))
|
|
|
|
if isinstance(stdout_reader, ReaderThread):
|
|
stdout_bytes = b''.join(stdout_reader.lines)
|
|
else:
|
|
stdout_bytes = b''
|
|
|
|
if isinstance(stderr_reader, ReaderThread):
|
|
stderr_bytes = b''.join(stderr_reader.lines)
|
|
else:
|
|
stderr_bytes = b''
|
|
|
|
process.wait()
|
|
|
|
return stdout_bytes, stderr_bytes
|
|
|
|
|
|
class WriterThread(WrappedThread):
|
|
"""Thread to write data to stdin of a subprocess."""
|
|
|
|
def __init__(self, handle: t.IO[bytes], data: bytes) -> None:
|
|
super().__init__(self._run)
|
|
|
|
self.handle = handle
|
|
self.data = data
|
|
|
|
def _run(self) -> None:
|
|
"""Workload to run on a thread."""
|
|
try:
|
|
self.handle.write(self.data)
|
|
self.handle.flush()
|
|
finally:
|
|
self.handle.close()
|
|
|
|
|
|
class ReaderThread(WrappedThread, metaclass=abc.ABCMeta):
|
|
"""Thread to read stdout from a subprocess."""
|
|
|
|
def __init__(self, handle: t.IO[bytes], buffer: t.BinaryIO) -> None:
|
|
super().__init__(self._run)
|
|
|
|
self.handle = handle
|
|
self.buffer = buffer
|
|
self.lines: list[bytes] = []
|
|
|
|
@abc.abstractmethod
|
|
def _run(self) -> None:
|
|
"""Workload to run on a thread."""
|
|
|
|
|
|
class CaptureThread(ReaderThread):
|
|
"""Thread to capture stdout from a subprocess into a buffer."""
|
|
|
|
def _run(self) -> None:
|
|
"""Workload to run on a thread."""
|
|
src = self.handle
|
|
dst = self.lines
|
|
|
|
try:
|
|
for line in src:
|
|
dst.append(line)
|
|
finally:
|
|
src.close()
|
|
|
|
|
|
class OutputThread(ReaderThread):
|
|
"""Thread to pass stdout from a subprocess to stdout."""
|
|
|
|
def _run(self) -> None:
|
|
"""Workload to run on a thread."""
|
|
src = self.handle
|
|
dst = self.buffer
|
|
|
|
try:
|
|
for line in src:
|
|
dst.write(line)
|
|
dst.flush()
|
|
finally:
|
|
src.close()
|
|
|
|
|
|
def common_environment() -> dict[str, str]:
|
|
"""Common environment used for executing all programs."""
|
|
env = dict(
|
|
LC_ALL=CONFIGURED_LOCALE,
|
|
PATH=os.environ.get('PATH', os.path.defpath),
|
|
)
|
|
|
|
required = (
|
|
'HOME',
|
|
)
|
|
|
|
optional = (
|
|
'LD_LIBRARY_PATH',
|
|
'SSH_AUTH_SOCK',
|
|
# MacOS High Sierra Compatibility
|
|
# http://sealiesoftware.com/blog/archive/2017/6/5/Objective-C_and_fork_in_macOS_1013.html
|
|
# Example configuration for macOS:
|
|
# export OBJC_DISABLE_INITIALIZE_FORK_SAFETY=YES
|
|
'OBJC_DISABLE_INITIALIZE_FORK_SAFETY',
|
|
'ANSIBLE_KEEP_REMOTE_FILES',
|
|
# MacOS Homebrew Compatibility
|
|
# https://cryptography.io/en/latest/installation/#building-cryptography-on-macos
|
|
# This may also be required to install pyyaml with libyaml support when installed in non-standard locations.
|
|
# Example configuration for brew on macOS:
|
|
# export LDFLAGS="-L$(brew --prefix openssl)/lib/ -L$(brew --prefix libyaml)/lib/"
|
|
# export CFLAGS="-I$(brew --prefix openssl)/include/ -I$(brew --prefix libyaml)/include/"
|
|
'LDFLAGS',
|
|
'CFLAGS',
|
|
)
|
|
|
|
# FreeBSD Compatibility
|
|
# This is required to include libyaml support in PyYAML.
|
|
# The header /usr/local/include/yaml.h isn't in the default include path for the compiler.
|
|
# It is included here so that tests can take advantage of it, rather than only ansible-test during managed pip installs.
|
|
# If CFLAGS has been set in the environment that value will take precedence due to being an optional var when calling pass_vars.
|
|
if os.path.exists('/etc/freebsd-update.conf'):
|
|
env.update(CFLAGS='-I/usr/local/include/')
|
|
|
|
env.update(pass_vars(required=required, optional=optional))
|
|
|
|
return env
|
|
|
|
|
|
def report_locale(show_warning: bool) -> None:
|
|
"""Report the configured locale and the locale warning, if applicable."""
|
|
|
|
display.info(f'Configured locale: {CONFIGURED_LOCALE}', verbosity=1)
|
|
|
|
if LOCALE_WARNING and show_warning:
|
|
display.warning(LOCALE_WARNING)
|
|
|
|
|
|
def pass_vars(required: c.Collection[str], optional: c.Collection[str]) -> dict[str, str]:
|
|
"""Return a filtered dictionary of environment variables based on the current environment."""
|
|
env = {}
|
|
|
|
for name in required:
|
|
if name not in os.environ:
|
|
raise MissingEnvironmentVariable(name)
|
|
env[name] = os.environ[name]
|
|
|
|
for name in optional:
|
|
if name not in os.environ:
|
|
continue
|
|
env[name] = os.environ[name]
|
|
|
|
return env
|
|
|
|
|
|
def verified_chmod(path: str, mode: int) -> None:
|
|
"""Perform chmod on the specified path and then verify the permissions were applied."""
|
|
os.chmod(path, mode) # pylint: disable=ansible-bad-function
|
|
|
|
executable = any(mode & perm for perm in (stat.S_IXUSR, stat.S_IXGRP, stat.S_IXOTH))
|
|
|
|
if executable and not os.access(path, os.X_OK):
|
|
raise ApplicationError(f'Path "{path}" should executable, but is not. Is the filesystem mounted with the "noexec" option?')
|
|
|
|
|
|
def remove_tree(path: str) -> None:
|
|
"""Remove the specified directory, silently continuing if the directory does not exist."""
|
|
try:
|
|
shutil.rmtree(to_bytes(path))
|
|
except FileNotFoundError:
|
|
pass
|
|
|
|
|
|
def is_binary_file(path: str) -> bool:
|
|
"""Return True if the specified file is a binary file, otherwise return False."""
|
|
assume_text = {
|
|
'.cfg',
|
|
'.conf',
|
|
'.crt',
|
|
'.cs',
|
|
'.css',
|
|
'.html',
|
|
'.ini',
|
|
'.j2',
|
|
'.js',
|
|
'.json',
|
|
'.md',
|
|
'.pem',
|
|
'.ps1',
|
|
'.psm1',
|
|
'.py',
|
|
'.rst',
|
|
'.sh',
|
|
'.txt',
|
|
'.xml',
|
|
'.yaml',
|
|
'.yml',
|
|
}
|
|
|
|
assume_binary = {
|
|
'.bin',
|
|
'.eot',
|
|
'.gz',
|
|
'.ico',
|
|
'.iso',
|
|
'.jpg',
|
|
'.otf',
|
|
'.p12',
|
|
'.png',
|
|
'.pyc',
|
|
'.rpm',
|
|
'.ttf',
|
|
'.woff',
|
|
'.woff2',
|
|
'.zip',
|
|
}
|
|
|
|
ext = os.path.splitext(path)[1]
|
|
|
|
if ext in assume_text:
|
|
return False
|
|
|
|
if ext in assume_binary:
|
|
return True
|
|
|
|
with open_binary_file(path) as path_fd:
|
|
return b'\0' in path_fd.read(4096)
|
|
|
|
|
|
def generate_name(length: int = 8) -> str:
|
|
"""Generate and return a random name."""
|
|
return ''.join(random.choice(string.ascii_letters + string.digits) for _idx in range(length))
|
|
|
|
|
|
def generate_password() -> str:
|
|
"""Generate and return random password."""
|
|
chars = [
|
|
string.ascii_letters,
|
|
string.digits,
|
|
string.ascii_letters,
|
|
string.digits,
|
|
'-',
|
|
] * 4
|
|
|
|
password = ''.join([random.choice(char) for char in chars[:-1]])
|
|
|
|
display.sensitive.add(password)
|
|
|
|
return password
|
|
|
|
|
|
class Display:
|
|
"""Manages color console output."""
|
|
|
|
clear = '\033[0m'
|
|
red = '\033[31m'
|
|
green = '\033[32m'
|
|
yellow = '\033[33m'
|
|
blue = '\033[34m'
|
|
purple = '\033[35m'
|
|
cyan = '\033[36m'
|
|
|
|
verbosity_colors = {
|
|
0: None,
|
|
1: green,
|
|
2: blue,
|
|
3: cyan,
|
|
}
|
|
|
|
def __init__(self) -> None:
|
|
self.verbosity = 0
|
|
self.color = sys.stdout.isatty()
|
|
self.warnings: list[str] = []
|
|
self.warnings_unique: set[str] = set()
|
|
self.fd = sys.stderr # default to stderr until config is initialized to avoid early messages going to stdout
|
|
self.rows = 0
|
|
self.columns = 0
|
|
self.truncate = 0
|
|
self.redact = True
|
|
self.sensitive: set[str] = set()
|
|
|
|
if os.isatty(0):
|
|
self.rows, self.columns = unpack('HHHH', fcntl.ioctl(0, TIOCGWINSZ, pack('HHHH', 0, 0, 0, 0)))[:2]
|
|
|
|
def __warning(self, message: str) -> None:
|
|
"""Internal implementation for displaying a warning message."""
|
|
self.print_message('WARNING: %s' % message, color=self.purple)
|
|
|
|
def review_warnings(self) -> None:
|
|
"""Review all warnings which previously occurred."""
|
|
if not self.warnings:
|
|
return
|
|
|
|
self.__warning('Reviewing previous %d warning(s):' % len(self.warnings))
|
|
|
|
for warning in self.warnings:
|
|
self.__warning(warning)
|
|
|
|
def warning(self, message: str, unique: bool = False, verbosity: int = 0) -> None:
|
|
"""Display a warning level message."""
|
|
if verbosity > self.verbosity:
|
|
return
|
|
|
|
if unique:
|
|
if message in self.warnings_unique:
|
|
return
|
|
|
|
self.warnings_unique.add(message)
|
|
|
|
self.__warning(message)
|
|
self.warnings.append(message)
|
|
|
|
def notice(self, message: str) -> None:
|
|
"""Display a notice level message."""
|
|
self.print_message('NOTICE: %s' % message, color=self.purple)
|
|
|
|
def error(self, message: str) -> None:
|
|
"""Display an error level message."""
|
|
self.print_message('ERROR: %s' % message, color=self.red)
|
|
|
|
def fatal(self, message: str) -> None:
|
|
"""Display a fatal level message."""
|
|
self.print_message('FATAL: %s' % message, color=self.red, stderr=True)
|
|
|
|
def info(self, message: str, verbosity: int = 0, truncate: bool = False) -> None:
|
|
"""Display an info level message."""
|
|
if self.verbosity >= verbosity:
|
|
color = self.verbosity_colors.get(verbosity, self.yellow)
|
|
self.print_message(message, color=color, truncate=truncate)
|
|
|
|
def print_message( # pylint: disable=locally-disabled, invalid-name
|
|
self,
|
|
message: str,
|
|
color: t.Optional[str] = None,
|
|
stderr: bool = False,
|
|
truncate: bool = False,
|
|
) -> None:
|
|
"""Display a message."""
|
|
if self.redact and self.sensitive:
|
|
for item in self.sensitive:
|
|
if not item:
|
|
continue
|
|
|
|
message = message.replace(item, '*' * len(item))
|
|
|
|
if truncate:
|
|
if len(message) > self.truncate > 5:
|
|
message = message[:self.truncate - 5] + ' ...'
|
|
|
|
if color and self.color:
|
|
# convert color resets in message to desired color
|
|
message = message.replace(self.clear, color)
|
|
message = '%s%s%s' % (color, message, self.clear)
|
|
|
|
fd = sys.stderr if stderr else self.fd
|
|
|
|
print(message, file=fd)
|
|
fd.flush()
|
|
|
|
|
|
class InternalError(Exception):
|
|
"""An unhandled internal error indicating a bug in the code."""
|
|
|
|
def __init__(self, message: str) -> None:
|
|
super().__init__(f'An internal error has occurred in ansible-test: {message}')
|
|
|
|
|
|
class ApplicationError(Exception):
|
|
"""General application error."""
|
|
|
|
|
|
class ApplicationWarning(Exception):
|
|
"""General application warning which interrupts normal program flow."""
|
|
|
|
|
|
class SubprocessError(ApplicationError):
|
|
"""Error resulting from failed subprocess execution."""
|
|
|
|
def __init__(
|
|
self,
|
|
cmd: list[str],
|
|
status: int = 0,
|
|
stdout: t.Optional[str] = None,
|
|
stderr: t.Optional[str] = None,
|
|
runtime: t.Optional[float] = None,
|
|
error_callback: t.Optional[c.Callable[[SubprocessError], None]] = None,
|
|
) -> None:
|
|
message = 'Command "%s" returned exit status %s.\n' % (shlex.join(cmd), status)
|
|
|
|
if stderr:
|
|
message += '>>> Standard Error\n'
|
|
message += '%s%s\n' % (stderr.strip(), Display.clear)
|
|
|
|
if stdout:
|
|
message += '>>> Standard Output\n'
|
|
message += '%s%s\n' % (stdout.strip(), Display.clear)
|
|
|
|
self.cmd = cmd
|
|
self.message = message
|
|
self.status = status
|
|
self.stdout = stdout
|
|
self.stderr = stderr
|
|
self.runtime = runtime
|
|
|
|
if error_callback:
|
|
error_callback(self)
|
|
|
|
self.message = self.message.strip()
|
|
|
|
super().__init__(self.message)
|
|
|
|
|
|
class MissingEnvironmentVariable(ApplicationError):
|
|
"""Error caused by missing environment variable."""
|
|
|
|
def __init__(self, name: str) -> None:
|
|
super().__init__('Missing environment variable: %s' % name)
|
|
|
|
self.name = name
|
|
|
|
|
|
class HostConnectionError(ApplicationError):
|
|
"""
|
|
Raised when the initial connection during host profile setup has failed and all retries have been exhausted.
|
|
Raised by provisioning code when one or more provisioning threads raise this exception.
|
|
Also raised when an SSH connection fails for the shell command.
|
|
"""
|
|
|
|
def __init__(self, message: str, callback: t.Callable[[], None] = None) -> None:
|
|
super().__init__(message)
|
|
|
|
self._callback = callback
|
|
|
|
def run_callback(self) -> None:
|
|
"""Run the error callback, if any."""
|
|
if self._callback:
|
|
self._callback()
|
|
|
|
|
|
def retry(func: t.Callable[..., TValue], ex_type: t.Type[BaseException] = SubprocessError, sleep: int = 10, attempts: int = 10, warn: bool = True) -> TValue:
|
|
"""Retry the specified function on failure."""
|
|
for dummy in range(1, attempts):
|
|
try:
|
|
return func()
|
|
except ex_type as ex:
|
|
if warn:
|
|
display.warning(str(ex))
|
|
|
|
time.sleep(sleep)
|
|
|
|
return func()
|
|
|
|
|
|
def parse_to_list_of_dict(pattern: str, value: str) -> list[dict[str, str]]:
|
|
"""Parse lines from the given value using the specified pattern and return the extracted list of key/value pair dictionaries."""
|
|
matched = []
|
|
unmatched = []
|
|
|
|
for line in value.splitlines():
|
|
match = re.search(pattern, line)
|
|
|
|
if match:
|
|
matched.append(match.groupdict())
|
|
else:
|
|
unmatched.append(line)
|
|
|
|
if unmatched:
|
|
raise Exception('Pattern "%s" did not match values:\n%s' % (pattern, '\n'.join(unmatched)))
|
|
|
|
return matched
|
|
|
|
|
|
def get_subclasses(class_type: t.Type[C]) -> list[t.Type[C]]:
|
|
"""Returns a list of types that are concrete subclasses of the given type."""
|
|
subclasses: set[t.Type[C]] = set()
|
|
queue: list[t.Type[C]] = [class_type]
|
|
|
|
while queue:
|
|
parent = queue.pop()
|
|
|
|
for child in parent.__subclasses__():
|
|
if child not in subclasses:
|
|
if not inspect.isabstract(child):
|
|
subclasses.add(child)
|
|
queue.append(child)
|
|
|
|
return sorted(subclasses, key=lambda sc: sc.__name__)
|
|
|
|
|
|
def is_subdir(candidate_path: str, path: str) -> bool:
|
|
"""Returns true if candidate_path is path or a subdirectory of path."""
|
|
if not path.endswith(os.path.sep):
|
|
path += os.path.sep
|
|
|
|
if not candidate_path.endswith(os.path.sep):
|
|
candidate_path += os.path.sep
|
|
|
|
return candidate_path.startswith(path)
|
|
|
|
|
|
def paths_to_dirs(paths: list[str]) -> list[str]:
|
|
"""Returns a list of directories extracted from the given list of paths."""
|
|
dir_names = set()
|
|
|
|
for path in paths:
|
|
while True:
|
|
path = os.path.dirname(path)
|
|
|
|
if not path or path == os.path.sep:
|
|
break
|
|
|
|
dir_names.add(path + os.path.sep)
|
|
|
|
return sorted(dir_names)
|
|
|
|
|
|
def str_to_version(version: str) -> tuple[int, ...]:
|
|
"""Return a version tuple from a version string."""
|
|
return tuple(int(n) for n in version.split('.'))
|
|
|
|
|
|
def version_to_str(version: tuple[int, ...]) -> str:
|
|
"""Return a version string from a version tuple."""
|
|
return '.'.join(str(n) for n in version)
|
|
|
|
|
|
def sorted_versions(versions: list[str]) -> list[str]:
|
|
"""Return a sorted copy of the given list of versions."""
|
|
return [version_to_str(version) for version in sorted(str_to_version(version) for version in versions)]
|
|
|
|
|
|
def import_plugins(directory: str, root: t.Optional[str] = None) -> None:
|
|
"""
|
|
Import plugins from the given directory relative to the given root.
|
|
If the root is not provided, the 'lib' directory for the test runner will be used.
|
|
"""
|
|
if root is None:
|
|
root = os.path.dirname(__file__)
|
|
|
|
path = os.path.join(root, directory)
|
|
package = __name__.rsplit('.', 1)[0]
|
|
prefix = '%s.%s.' % (package, directory.replace(os.path.sep, '.'))
|
|
|
|
for (_module_loader, name, _ispkg) in pkgutil.iter_modules([path], prefix=prefix):
|
|
module_path = os.path.join(root, name[len(package) + 1:].replace('.', os.path.sep) + '.py')
|
|
load_module(module_path, name)
|
|
|
|
|
|
def load_plugins(base_type: t.Type[C], database: dict[str, t.Type[C]]) -> None:
|
|
"""
|
|
Load plugins of the specified type and track them in the specified database.
|
|
Only plugins which have already been imported will be loaded.
|
|
"""
|
|
plugins: dict[str, t.Type[C]] = dict((sc.__module__.rsplit('.', 1)[1], sc) for sc in get_subclasses(base_type))
|
|
|
|
for plugin in plugins:
|
|
database[plugin] = plugins[plugin]
|
|
|
|
|
|
def load_module(path: str, name: str) -> None:
|
|
"""Load a Python module using the given name and path."""
|
|
if name in sys.modules:
|
|
return
|
|
|
|
spec = importlib.util.spec_from_file_location(name, path)
|
|
module = importlib.util.module_from_spec(spec)
|
|
sys.modules[name] = module
|
|
spec.loader.exec_module(module)
|
|
|
|
|
|
def sanitize_host_name(name: str) -> str:
|
|
"""Return a sanitized version of the given name, suitable for use as a hostname."""
|
|
return re.sub('[^A-Za-z0-9]+', '-', name)[:63].strip('-')
|
|
|
|
|
|
def get_generic_type(base_type: t.Type, generic_base_type: t.Type[TValue]) -> t.Optional[t.Type[TValue]]:
|
|
"""Return the generic type arg derived from the generic_base_type type that is associated with the base_type type, if any, otherwise return None."""
|
|
# noinspection PyUnresolvedReferences
|
|
type_arg = t.get_args(base_type.__orig_bases__[0])[0]
|
|
return None if isinstance(type_arg, generic_base_type) else type_arg
|
|
|
|
|
|
def get_type_associations(base_type: t.Type[TBase], generic_base_type: t.Type[TValue]) -> list[tuple[t.Type[TValue], t.Type[TBase]]]:
|
|
"""Create and return a list of tuples associating generic_base_type derived types with a corresponding base_type derived type."""
|
|
return [item for item in [(get_generic_type(sc_type, generic_base_type), sc_type) for sc_type in get_subclasses(base_type)] if item[1]]
|
|
|
|
|
|
def get_type_map(base_type: t.Type[TBase], generic_base_type: t.Type[TValue]) -> dict[t.Type[TValue], t.Type[TBase]]:
|
|
"""Create and return a mapping of generic_base_type derived types to base_type derived types."""
|
|
return {item[0]: item[1] for item in get_type_associations(base_type, generic_base_type)}
|
|
|
|
|
|
def verify_sys_executable(path: str) -> t.Optional[str]:
|
|
"""Verify that the given path references the current Python interpreter. If not, return the expected path, otherwise return None."""
|
|
if path == sys.executable:
|
|
return None
|
|
|
|
if os.path.realpath(path) == os.path.realpath(sys.executable):
|
|
return None
|
|
|
|
expected_executable = raw_command([path, '-c', 'import sys; print(sys.executable)'], capture=True)[0]
|
|
|
|
if expected_executable == sys.executable:
|
|
return None
|
|
|
|
return expected_executable
|
|
|
|
|
|
def type_guard(sequence: c.Sequence[t.Any], guard_type: t.Type[C]) -> TypeGuard[c.Sequence[C]]:
|
|
"""
|
|
Raises an exception if any item in the given sequence does not match the specified guard type.
|
|
Use with assert so that type checkers are aware of the type guard.
|
|
"""
|
|
invalid_types = set(type(item) for item in sequence if not isinstance(item, guard_type))
|
|
|
|
if not invalid_types:
|
|
return True
|
|
|
|
invalid_type_names = sorted(str(item) for item in invalid_types)
|
|
|
|
raise Exception(f'Sequence required to contain only {guard_type} includes: {", ".join(invalid_type_names)}')
|
|
|
|
|
|
display = Display() # pylint: disable=locally-disabled, invalid-name
|
|
|
|
_enable_vendoring()
|