From cd9827e6649b7f14513c2f201afc9e9d46998ddb Mon Sep 17 00:00:00 2001 From: Tobias Henkel Date: Sat, 19 Jan 2019 13:33:39 +0100 Subject: Manage ansible installations within zuul As a first step towards supporting multiple ansible versions we need tooling to manage ansible installations. This moves the installation of ansible from the requirements.txt into zuul. This is called as a setup hook to install the ansible versions into /lib/zuul/ansible. Further this tooling abstracts knowledge that the executor must know in order to actually run the correct version of ansible. The actual usage of multiple ansible versions will be done in follow-ups. For better maintainability the ansible plugins live in zuul/ansible/base where plugins can be kept in different versions if necessary. For each supported ansible version there is a specific folder that symlinks the according plugins. Change-Id: I5ce1385245c76818777aa34230786a9dbaf723e5 Depends-On: https://review.openstack.org/623927 --- zuul/ansible/base/callback/zuul_stream.py | 633 ++++++++++++++++++++++++++++++ 1 file changed, 633 insertions(+) create mode 100644 zuul/ansible/base/callback/zuul_stream.py (limited to 'zuul/ansible/base/callback/zuul_stream.py') diff --git a/zuul/ansible/base/callback/zuul_stream.py b/zuul/ansible/base/callback/zuul_stream.py new file mode 100644 index 000000000..966c9008c --- /dev/null +++ b/zuul/ansible/base/callback/zuul_stream.py @@ -0,0 +1,633 @@ +# Copyright 2017 Red Hat, Inc. +# +# Zuul is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# Zuul is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with Ansible. If not, see . + +# This is not needed in python3 - but it is needed in python2 because there +# is a json module in ansible.plugins.callback and python2 gets confused. +# Easy local testing with ansible-playbook is handy when hacking on zuul_stream +# so just put in the __future__ statement. +from __future__ import absolute_import + +import datetime +import logging +import logging.config +import json +import os +import socket +import threading +import time + +from ansible.plugins.callback import default +from zuul.ansible import paths + +from zuul.ansible import logconfig + +LOG_STREAM_PORT = 19885 + + +def zuul_filter_result(result): + """Remove keys from shell/command output. + + Zuul streams stdout into the log above, so including stdout and stderr + in the result dict that ansible displays in the logs is duplicate + noise. We keep stdout in the result dict so that other callback plugins + like ARA could also have access to it. But drop them here. + + Remove changed so that we don't show a bunch of "changed" titles + on successful shell tasks, since that doesn't make sense from a Zuul + POV. The super class treats missing "changed" key as False. + + Remove cmd because most of the script content where people want to + see the script run is run with -x. It's possible we may want to revist + this to be smarter about when we remove it - like, only remove it + if it has an embedded newline - so that for normal 'simple' uses + of cmd it'll echo what the command was for folks. + """ + + stdout = result.pop('stdout', '') + stdout_lines = result.pop('stdout_lines', []) + if not stdout_lines and stdout: + stdout_lines = stdout.split('\n') + + for key in ('changed', 'cmd', 'zuul_log_id', 'invocation', + 'stderr', 'stderr_lines'): + result.pop(key, None) + return stdout_lines + + +class CallbackModule(default.CallbackModule): + + ''' + This is the Zuul streaming callback. It's based on the default + callback plugin, but streams results from shell commands. + ''' + + CALLBACK_VERSION = 2.0 + CALLBACK_TYPE = 'stdout' + CALLBACK_NAME = 'zuul_stream' + + def __init__(self): + + super(CallbackModule, self).__init__() + self._task = None + self._daemon_running = False + self._play = None + self._streamers = [] + self._streamers_stop = False + self.configure_logger() + self._items_done = False + self._deferred_result = None + self._playbook_name = None + + def configure_logger(self): + # ansible appends timestamp, user and pid to the log lines emitted + # to the log file. We're doing other things though, so we don't want + # this. + logging_config = logconfig.load_job_config( + os.environ['ZUUL_JOB_LOG_CONFIG']) + + if self._display.verbosity > 2: + logging_config.setDebug() + + logging_config.apply() + + self._logger = logging.getLogger('zuul.executor.ansible') + + def _log(self, msg, ts=None, job=True, executor=False, debug=False): + msg = msg.rstrip() + if job: + now = ts or datetime.datetime.now() + self._logger.info("{now} | {msg}".format(now=now, msg=msg)) + if executor: + if debug: + self._display.vvv(msg) + else: + self._display.display(msg) + + def _read_log(self, host, ip, log_id, task_name, hosts): + self._log("[%s] Starting to log %s for task %s" + % (host, log_id, task_name), job=False, executor=True) + while True: + try: + s = socket.create_connection((ip, LOG_STREAM_PORT), 5) + # Disable the socket timeout after we have successfully + # connected to accomodate the fact that jobs may not be writing + # logs continously. Without this we can easily trip the 5 + # second timeout. + s.settimeout(None) + except socket.timeout: + self._log( + "Timeout exception waiting for the logger. " + "Please check connectivity to [%s:%s]" + % (ip, LOG_STREAM_PORT), executor=True) + self._log_streamline( + "localhost", + "Timeout exception waiting for the logger. " + "Please check connectivity to [%s:%s]" + % (ip, LOG_STREAM_PORT)) + return + except Exception: + self._log("[%s] Waiting on logger" % host, + executor=True, debug=True) + time.sleep(0.1) + continue + msg = "%s\n" % log_id + s.send(msg.encode("utf-8")) + buff = s.recv(4096) + buffering = True + while buffering: + if b'\n' in buff: + (line, buff) = buff.split(b'\n', 1) + # We can potentially get binary data here. In order to + # being able to handle that use the backslashreplace + # error handling method. This decodes unknown utf-8 + # code points to escape sequences which exactly represent + # the correct data without throwing a decoding exception. + done = self._log_streamline( + host, line.decode("utf-8", "backslashreplace")) + if done: + return + else: + more = s.recv(4096) + if not more: + buffering = False + else: + buff += more + if buff: + self._log_streamline( + host, buff.decode("utf-8", "backslashreplace")) + + def _log_streamline(self, host, line): + if "[Zuul] Task exit code" in line: + return True + elif self._streamers_stop and "[Zuul] Log not found" in line: + return True + elif "[Zuul] Log not found" in line: + # don't output this line + return False + else: + ts, ln = line.split(' | ', 1) + + self._log("%s | %s " % (host, ln), ts=ts) + return False + + def _log_module_failure(self, result, result_dict): + if 'module_stdout' in result_dict and result_dict['module_stdout']: + self._log_message( + result, status='MODULE FAILURE', + msg=result_dict['module_stdout']) + elif 'exception' in result_dict and result_dict['exception']: + self._log_message( + result, status='MODULE FAILURE', + msg=result_dict['exception']) + elif 'module_stderr' in result_dict: + self._log_message( + result, status='MODULE FAILURE', + msg=result_dict['module_stderr']) + + def v2_playbook_on_start(self, playbook): + self._playbook_name = os.path.splitext(playbook._file_name)[0] + + def v2_playbook_on_include(self, included_file): + for host in included_file._hosts: + self._log("{host} | included: {filename}".format( + host=host.name, + filename=included_file._filename)) + + def v2_playbook_on_play_start(self, play): + self._play = play + # Log an extra blank line to get space before each play + self._log("") + + # the name of a play defaults to the hosts string + name = play.get_name().strip() + msg = u"PLAY [{name}]".format(name=name) + + self._log(msg) + + def v2_playbook_on_task_start(self, task, is_conditional): + # Log an extra blank line to get space before each task + self._log("") + + self._task = task + + if self._play.strategy != 'free': + task_name = self._print_task_banner(task) + else: + task_name = task.get_name().strip() + + if task.action in ('command', 'shell'): + play_vars = self._play._variable_manager._hostvars + + hosts = self._get_task_hosts(task) + for host, inventory_hostname in hosts: + if host in ('localhost', '127.0.0.1'): + # Don't try to stream from localhost + continue + ip = play_vars[host].get( + 'ansible_host', play_vars[host].get( + 'ansible_inventory_host')) + if ip in ('localhost', '127.0.0.1'): + # Don't try to stream from localhost + continue + if task.loop: + # Don't try to stream from loops + continue + if play_vars[host].get('ansible_connection') in ('kubectl', ): + # Don't try to stream from kubectl connection + continue + + log_id = "%s-%s" % ( + task._uuid, paths._sanitize_filename(inventory_hostname)) + streamer = threading.Thread( + target=self._read_log, args=( + host, ip, log_id, task_name, hosts)) + streamer.daemon = True + streamer.start() + self._streamers.append(streamer) + + def v2_playbook_on_handler_task_start(self, task): + self.v2_playbook_on_task_start(task, False) + + def _stop_streamers(self): + self._streamers_stop = True + while True: + if not self._streamers: + break + streamer = self._streamers.pop() + streamer.join(30) + if streamer.is_alive(): + msg = "[Zuul] Log Stream did not terminate" + self._log(msg, job=True, executor=True) + self._streamers_stop = False + + def _process_result_for_localhost(self, result, is_task=True): + result_dict = dict(result._result) + localhost_names = ('localhost', '127.0.0.1', '::1') + is_localhost = False + task_host = result._host.get_name() + delegated_vars = result_dict.get('_ansible_delegated_vars', None) + if delegated_vars: + delegated_host = delegated_vars['ansible_host'] + if delegated_host in localhost_names: + is_localhost = True + elif result._task._variable_manager is None: + # Handle fact gathering which doens't have a variable manager + if task_host == 'localhost': + is_localhost = True + else: + task_hostvars = result._task._variable_manager._hostvars[task_host] + # Normally hosts in the inventory will have ansible_host + # or ansible_inventory host defined. The implied + # inventory record for 'localhost' will have neither, so + # default to that if none are supplied. + if task_hostvars.get('ansible_host', task_hostvars.get( + 'ansible_inventory_host', 'localhost')) in localhost_names: + is_localhost = True + + if not is_localhost and is_task: + self._stop_streamers() + if result._task.action in ('command', 'shell', + 'win_command', 'win_shell'): + stdout_lines = zuul_filter_result(result_dict) + # We don't have streaming for localhost and windows modules so get + # standard out after the fact. + if is_localhost or result._task.action in ( + 'win_command', 'win_shell'): + for line in stdout_lines: + hostname = self._get_hostname(result) + self._log("%s | %s " % (hostname, line)) + + def v2_runner_on_failed(self, result, ignore_errors=False): + result_dict = dict(result._result) + + self._handle_exception(result_dict) + + if result_dict.get('msg') == 'All items completed': + result_dict['status'] = 'ERROR' + self._deferred_result = result_dict + return + + self._process_result_for_localhost(result) + + if result._task.loop and 'results' in result_dict: + # items have their own events + pass + elif (result_dict.get('msg') == 'MODULE FAILURE'): + self._log_module_failure(result, result_dict) + else: + self._log_message( + result=result, status='ERROR', result_dict=result_dict) + if ignore_errors: + self._log_message(result, "Ignoring Errors", status="ERROR") + + def v2_runner_on_skipped(self, result): + if result._task.loop: + self._items_done = False + self._deferred_result = dict(result._result) + else: + reason = result._result.get('skip_reason') + if reason: + # No reason means it's an item, which we'll log differently + self._log_message(result, status='skipping', msg=reason) + + def v2_runner_item_on_skipped(self, result): + reason = result._result.get('skip_reason') + if reason: + self._log_message(result, status='skipping', msg=reason) + else: + self._log_message(result, status='skipping') + + if self._deferred_result: + self._process_deferred(result) + + def v2_runner_on_ok(self, result): + if (self._play.strategy == 'free' + and self._last_task_banner != result._task._uuid): + self._print_task_banner(result._task) + + result_dict = dict(result._result) + + self._clean_results(result_dict, result._task.action) + if '_zuul_nolog_return' in result_dict: + # We have a custom zuul module that doesn't want the parameters + # from its returned splatted to stdout. This is typically for + # modules that are collecting data to be displayed some other way. + for key in list(result_dict.keys()): + if key != 'changed': + result_dict.pop(key) + + if result_dict.get('changed', False): + status = 'changed' + else: + status = 'ok' + + if (result_dict.get('msg') == 'All items completed' + and not self._items_done): + result_dict['status'] = status + self._deferred_result = result_dict + return + + if not result._task.loop: + self._process_result_for_localhost(result) + else: + self._items_done = False + + self._handle_warnings(result_dict) + + if result._task.loop and 'results' in result_dict: + # items have their own events + pass + + elif (result_dict.get('msg') == 'MODULE FAILURE'): + self._log_module_failure(result, result_dict) + elif result._task.action == 'debug': + # this is a debug statement, handle it special + for key in [k for k in result_dict + if k.startswith('_ansible')]: + del result_dict[key] + if 'changed' in result_dict: + del result_dict['changed'] + keyname = next(iter(result_dict.keys())) + # If it has msg, that means it was like: + # + # debug: + # msg: Some debug text the user was looking for + # + # So we log it with self._log to get just the raw string the + # user provided. Note that msg may be a multi line block quote + # so we handle that here as well. + if keyname == 'msg': + msg_lines = result_dict['msg'].rstrip().split('\n') + for msg_line in msg_lines: + self._log(msg=msg_line) + else: + self._log_message( + msg=json.dumps(result_dict, indent=2, sort_keys=True), + status=status, result=result) + elif result._task.action not in ('command', 'shell'): + if 'msg' in result_dict: + self._log_message(msg=result_dict['msg'], + result=result, status=status) + else: + self._log_message( + result=result, + status=status) + elif 'results' in result_dict: + for res in result_dict['results']: + self._log_message( + result, + "Runtime: {delta}".format(**res)) + elif result_dict.get('msg') == 'All items completed': + self._log_message(result, result_dict['msg']) + else: + self._log_message( + result, + "Runtime: {delta}".format( + **result_dict)) + + def v2_runner_item_on_ok(self, result): + result_dict = dict(result._result) + self._process_result_for_localhost(result, is_task=False) + + if result_dict.get('changed', False): + status = 'changed' + else: + status = 'ok' + + if result_dict.get('msg') == 'MODULE FAILURE': + self._log_module_failure(result, result_dict) + elif result._task.action not in ('command', 'shell', + 'win_command', 'win_shell'): + if 'msg' in result_dict: + self._log_message( + result=result, msg=result_dict['msg'], status=status) + else: + self._log_message( + result=result, + msg=json.dumps(result_dict['item'], + indent=2, sort_keys=True), + status=status) + else: + stdout_lines = zuul_filter_result(result_dict) + for line in stdout_lines: + hostname = self._get_hostname(result) + self._log("%s | %s " % (hostname, line)) + + if isinstance(result_dict['item'], str): + self._log_message( + result, + "Item: {item} Runtime: {delta}".format(**result_dict)) + else: + self._log_message( + result, + "Item: Runtime: {delta}".format( + **result_dict)) + + if self._deferred_result: + self._process_deferred(result) + + def v2_runner_item_on_failed(self, result): + result_dict = dict(result._result) + self._process_result_for_localhost(result, is_task=False) + + if result_dict.get('msg') == 'MODULE FAILURE': + self._log_module_failure(result, result_dict) + elif result._task.action not in ('command', 'shell', + 'win_command', 'win_shell'): + self._log_message( + result=result, + msg="Item: {item}".format(item=result_dict['item']), + status='ERROR', + result_dict=result_dict) + else: + stdout_lines = zuul_filter_result(result_dict) + for line in stdout_lines: + hostname = self._get_hostname(result) + self._log("%s | %s " % (hostname, line)) + + # self._log("Result: %s" % (result_dict)) + self._log_message( + result, "Item: {item} Result: {rc}".format(**result_dict)) + + if self._deferred_result: + self._process_deferred(result) + + def v2_playbook_on_stats(self, stats): + # Add a spacer line before the stats so that there will be a line + # between the last task and the recap + self._log("") + + self._log("PLAY RECAP") + + hosts = sorted(stats.processed.keys()) + for host in hosts: + t = stats.summarize(host) + self._log( + "{host} |" + " ok: {ok}" + " changed: {changed}" + " unreachable: {unreachable}" + " failed: {failures}".format(host=host, **t)) + + # Add a spacer line after the stats so that there will be a line + # between each playbook + self._log("") + + def _process_deferred(self, result): + self._items_done = True + result_dict = self._deferred_result + self._deferred_result = None + status = result_dict.get('status') + + if status: + self._log_message(result, "All items complete", status=status) + + # Log an extra blank line to get space after each task + self._log("") + + def _print_task_banner(self, task): + + task_name = task.get_name().strip() + + if task.loop: + task_type = 'LOOP' + else: + task_type = 'TASK' + + # TODO(mordred) With the removal of printing task args, do we really + # want to keep doing this section? + task_args = task.args.copy() + is_shell = task_args.pop('_uses_shell', False) + if is_shell and task_name == 'command': + task_name = 'shell' + raw_params = task_args.pop('_raw_params', '').split('\n') + # If there's just a single line, go ahead and print it + if len(raw_params) == 1 and task_name in ('shell', 'command'): + task_name = '{name}: {command}'.format( + name=task_name, command=raw_params[0]) + + msg = "{task_type} [{task}]".format( + task_type=task_type, + task=task_name) + self._log(msg) + return task + + def _get_task_hosts(self, task): + result = [] + + # _restriction returns the parsed/compiled list of hosts after + # applying subsets/limits + hosts = self._play._variable_manager._inventory._restriction + for inventory_host in hosts: + # If this task has as delegate to, we don't care about the play + # hosts, we care about the task's delegate target. + if task.delegate_to: + host = task.delegate_to + else: + host = inventory_host + result.append((host, inventory_host)) + + return result + + def _dump_result_dict(self, result_dict): + result_dict = result_dict.copy() + for key in list(result_dict.keys()): + if key.startswith('_ansible'): + del result_dict[key] + zuul_filter_result(result_dict) + return result_dict + + def _log_message(self, result, msg=None, status="ok", result_dict=None): + hostname = self._get_hostname(result) + if result_dict: + result_dict = self._dump_result_dict(result_dict) + if result._task.no_log: + self._log("{host} | {msg}".format( + host=hostname, + msg="Output suppressed because no_log was given")) + return + if (not msg and result_dict + and set(result_dict.keys()) == set(['msg', 'failed'])): + msg = result_dict['msg'] + result_dict = None + if msg: + msg_lines = msg.rstrip().split('\n') + if len(msg_lines) > 1: + self._log("{host} | {status}:".format( + host=hostname, status=status)) + for msg_line in msg_lines: + self._log("{host} | {msg_line}".format( + host=hostname, msg_line=msg_line)) + else: + self._log("{host} | {status}: {msg}".format( + host=hostname, status=status, msg=msg)) + else: + self._log("{host} | {status}".format( + host=hostname, status=status, msg=msg)) + if result_dict: + result_string = json.dumps(result_dict, indent=2, sort_keys=True) + for line in result_string.split('\n'): + self._log("{host} | {line}".format(host=hostname, line=line)) + + def _get_hostname(self, result): + delegated_vars = result._result.get('_ansible_delegated_vars', None) + if delegated_vars: + return "{host} -> {delegated_host}".format( + host=result._host.get_name(), + delegated_host=delegated_vars['ansible_host']) + else: + return result._host.get_name() + + v2_runner_on_unreachable = v2_runner_on_failed -- cgit v1.2.1