summaryrefslogtreecommitdiff
path: root/pytests
diff options
context:
space:
mode:
authorJonathan Abrahams <jonathan@mongodb.com>2017-11-21 13:13:21 -0500
committerJonathan Abrahams <jonathan@mongodb.com>2017-11-21 13:13:21 -0500
commitdcaf6432b7219d57cc5cf0de56b61264b8155e5d (patch)
treece7fbf63fb60d2a7055b49be75993d9c0ac4e41e /pytests
parent470c30273c479c5ab139de0d38e016f95c589c49 (diff)
downloadmongo-dcaf6432b7219d57cc5cf0de56b61264b8155e5d.tar.gz
SERVER-31886 Add powercycle tasks to run on Windows
- Add capability to generate report.json in powertest.py - Add capability to generate and read a configuration YML file in powertest.py
Diffstat (limited to 'pytests')
-rwxr-xr-xpytests/powertest.py378
1 files changed, 274 insertions, 104 deletions
diff --git a/pytests/powertest.py b/pytests/powertest.py
index eec89171911..2fe5a5d8fa5 100755
--- a/pytests/powertest.py
+++ b/pytests/powertest.py
@@ -12,6 +12,7 @@ import collections
import copy
import datetime
import distutils.spawn
+import json
import importlib
import logging
import optparse
@@ -91,6 +92,10 @@ __version__ = "0.1"
LOGGER = logging.getLogger(__name__)
+_report_json_success = False
+_report_json = {}
+_report_json_file = ""
+
""" Client & server side powercycle test script.
This script can be run against any host which is reachable via ssh.
@@ -105,8 +110,43 @@ LOGGER = logging.getLogger(__name__)
def exit_handler():
- """Exit handler, deletes all named temporary files."""
- LOGGER.debug("Exit handler invoked, cleaning up temporary files")
+ """Exit handler actions:
+ - Generate report.json
+ - Kill spawned processes
+ - Delete all named temporary files
+ """
+ if _report_json:
+ LOGGER.debug("Exit handler: Updating report file %s", _report_json_file)
+ try:
+ test_start = _report_json["results"][0]["start"]
+ test_end = int(time.time())
+ test_time = test_end - test_start
+ if _report_json_success:
+ failures = 0
+ status = "pass"
+ exit_code = 0
+ else:
+ failures = 1
+ status = "fail"
+ exit_code = 1
+ _report_json["failures"] = failures
+ _report_json["results"][0]["status"] = status
+ _report_json["results"][0]["exit_code"] = exit_code
+ _report_json["results"][0]["end"] = test_end
+ _report_json["results"][0]["elapsed"] = test_time
+ with open(_report_json_file, "w") as jstream:
+ json.dump(_report_json, jstream)
+ LOGGER.debug("Exit handler: report file contents %s", _report_json)
+ except:
+ pass
+
+ LOGGER.debug("Exit handler: Killing processes")
+ try:
+ Processes.kill_all()
+ except:
+ pass
+
+ LOGGER.debug("Exit handler: Cleaning up temporary files")
try:
NamedTempFile.delete_all()
except:
@@ -115,7 +155,7 @@ def exit_handler():
def child_processes(parent_pid):
"""Returns a list of all child processes for a pid."""
- # The child processes cannot be obtained from the parent on Windows. See
+ # The child processes cannot be obtained from the parent on Windows from psutil. See
# https://stackoverflow.com/questions/30220732/python-psutil-not-showing-all-child-processes
child_procs = []
while psutil.pid_exists(parent_pid):
@@ -136,20 +176,19 @@ def kill_process(pid, kill_children=True):
try:
parent = psutil.Process(pid)
except psutil.NoSuchProcess:
- LOGGER.error("Could not kill process %d, as it no longer exists", pid)
+ LOGGER.warn("Could not kill process %d, as it no longer exists", pid)
return 0
- procs = []
+ procs = [parent]
if kill_children:
procs += child_processes(pid)
- procs.append(parent)
for proc in procs:
try:
- LOGGER.debug("Killing process %d", proc.pid)
+ LOGGER.debug("Killing process '%s' pid %d", proc.name(), proc.pid)
proc.kill()
except psutil.NoSuchProcess:
- LOGGER.error("Could not kill process %d, as it no longer exists", pid)
+ LOGGER.warn("Could not kill process %d, as it no longer exists", pid)
_, alive = psutil.wait_procs(procs, timeout=30, callback=None)
if alive:
@@ -161,10 +200,10 @@ def kill_process(pid, kill_children=True):
def kill_processes(procs, kill_children=True):
"""Kill a list of processes and optionally it's children."""
for proc in procs:
- LOGGER.debug("Killing parent process %d", proc.pid)
+ LOGGER.debug("Starting kill of parent process %d", proc.pid)
kill_process(proc.pid, kill_children=kill_children)
ret = proc.wait()
- LOGGER.debug("Kill of parent process %d has return code of %d", proc.pid, ret)
+ LOGGER.debug("Finished kill of parent process %d has return code of %d", proc.pid, ret)
def get_extension(filename):
@@ -203,7 +242,7 @@ def get_bin_dir(root_dir):
def create_temp_executable_file(cmds):
"""Creates an executable temporary file containing 'cmds'. Returns file name."""
- temp_file_name = NamedTempFile.create(suffix=".sh", dir="tmp")
+ temp_file_name = NamedTempFile.create(suffix=".sh", directory="tmp")
with NamedTempFile.get(temp_file_name) as temp_file:
temp_file.write(cmds)
os_st = os.stat(temp_file_name)
@@ -233,6 +272,7 @@ def start_cmd(cmd, use_file=False):
LOGGER.debug("Executing '%s'", cmd)
proc = subprocess.Popen(cmd, close_fds=True)
+ LOGGER.debug("Spawned process %s pid %d", psutil.Process(proc.pid).name(), proc.pid)
return proc
@@ -509,6 +549,31 @@ def call_remote_operation(local_ops, remote_python, script_name, client_args, op
return ret, output
+class Processes(object):
+ """Class to create and kill spawned processes."""
+
+ _PROC_LIST = []
+
+ @classmethod
+ def create(cls, cmds):
+ """Creates a spawned process."""
+ proc = start_cmd(cmds, use_file=True)
+ cls._PROC_LIST.append(proc)
+
+ @classmethod
+ def kill(cls, proc):
+ """Kills a spawned process and all it's children."""
+ kill_processes([proc], kill_children=True)
+ cls._PROC_LIST.remove(proc)
+
+ @classmethod
+ def kill_all(cls):
+ """Kill all spawned processes."""
+ procs = copy.copy(cls._PROC_LIST)
+ for proc in procs:
+ cls.kill(proc)
+
+
class NamedTempFile(object):
"""Class to control temporary files."""
@@ -516,12 +581,13 @@ class NamedTempFile(object):
_DIR_LIST = []
@classmethod
- def create(cls, dir=None, suffix=""):
+ def create(cls, directory=None, suffix=""):
"""Creates a temporary file, and optional directory, and returns the file name."""
- if dir and not os.path.isdir(dir):
- os.makedirs(dir)
- cls._DIR_LIST.append(dir)
- temp_file = tempfile.NamedTemporaryFile(suffix=suffix, dir=dir, delete=False)
+ if directory and not os.path.isdir(directory):
+ LOGGER.debug("Creating temporary directory %s", directory)
+ os.makedirs(directory)
+ cls._DIR_LIST.append(directory)
+ temp_file = tempfile.NamedTemporaryFile(suffix=suffix, dir=directory, delete=False)
cls._FILE_MAP[temp_file.name] = temp_file
return temp_file.name
@@ -537,32 +603,41 @@ class NamedTempFile(object):
"""Deletes temporary file. Raises an exception if the file is unknown."""
if name not in cls._FILE_MAP:
raise Exception("Unknown temporary file {}.".format(name))
+ if not os.path.exists(name):
+ LOGGER.debug("Temporary file %s no longer exists", name)
+ del cls._FILE_MAP[name]
+ return
try:
os.remove(name)
except (IOError, OSError) as err:
- LOGGER.warn("Unable to delete temporary file {} with error {}".format(name, err))
+ LOGGER.warn("Unable to delete temporary file %s with error %s", name, err)
if not os.path.exists(name):
del cls._FILE_MAP[name]
@classmethod
- def delete_dir(cls, dir):
+ def delete_dir(cls, directory):
"""Deletes temporary directory. Raises an exception if the directory is unknown."""
- if dir not in cls._DIR_LIST:
- raise Exception("Unknown temporary directory {}.".format(dir))
+ if directory not in cls._DIR_LIST:
+ raise Exception("Unknown temporary directory {}.".format(directory))
+ if not os.path.exists(directory):
+ LOGGER.debug("Temporary directory %s no longer exists", directory)
+ cls._DIR_LIST.remove(directory)
+ return
try:
- shutil.rmtree(dir)
+ shutil.rmtree(directory)
except (IOError, OSError) as err:
- LOGGER.warn("Unable to delete temporary directory {} with error {}".format(dir, err))
- if not os.path.exists(dir):
- cls._DIR_LIST.remove(dir)
+ LOGGER.warn(
+ "Unable to delete temporary directory %s with error %s", directory, err)
+ if not os.path.exists(directory):
+ cls._DIR_LIST.remove(directory)
@classmethod
def delete_all(cls):
"""Deletes all temporary files and directories."""
for name in list(cls._FILE_MAP):
cls.delete(name)
- for dir in cls._DIR_LIST:
- cls.delete_dir(dir)
+ for directory in cls._DIR_LIST:
+ cls.delete_dir(directory)
class ProcessControl(object):
@@ -750,8 +825,7 @@ class WindowsService(object):
win32serviceutil.QueryServiceStatus(serviceName=self.name))
if svc_state in self._states:
return self._states[svc_state]
- else:
- return "unknown"
+ return "unknown"
except pywintypes.error:
return "not installed"
@@ -982,12 +1056,20 @@ def remote_handler(options, operations):
port=options.port,
options=options.mongod_options)
+ mongo_client_opts = get_mongo_client_args(options, host="localhost", port=options.port)
+
# Perform the sequence of operations specified. If any operation fails
# then return immediately.
for operation in operations:
# This is the internal "crash" mechanism, which is executed on the remote host.
if operation == "crash_server":
ret, output = internal_crash(options.remote_sudo)
+ # An internal crash on Windows is not immediate
+ try:
+ LOGGER.info("Waiting after issuing internal crash!")
+ time.sleep(30)
+ except IOError:
+ pass
elif operation == "install_mongod":
ret, output = mongod.install(root_dir, options.tarball_url)
@@ -1023,7 +1105,7 @@ def remote_handler(options, operations):
return ret
LOGGER.info("Started mongod running on port %d pid %s",
options.port, mongod.get_pids())
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
LOGGER.info("Server buildinfo: %s", mongo.admin.command("buildinfo"))
LOGGER.info("Server serverStatus: %s", mongo.admin.command("serverStatus"))
if options.use_replica_set and options.repl_set:
@@ -1036,7 +1118,7 @@ def remote_handler(options, operations):
ret = wait_for_mongod_shutdown(options.db_path)
elif operation == "shutdown_mongod":
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
try:
mongo.admin.command("shutdown", force=True)
except pymongo.errors.AutoReconnect:
@@ -1048,26 +1130,26 @@ def remote_handler(options, operations):
LOGGER.info(output)
elif operation == "seed_docs":
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
ret = mongo_seed_docs(
mongo, options.db_name, options.collection_name, options.seed_doc_num)
elif operation == "validate_collections":
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
ret = mongo_validate_collections(mongo)
elif operation == "insert_canary":
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
ret = mongo_insert_canary(
mongo, options.db_name, options.collection_name, options.canary_doc)
elif operation == "validate_canary":
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
ret = mongo_validate_canary(
mongo, options.db_name, options.collection_name, options.canary_doc)
elif operation == "set_fcv":
- mongo = pymongo.MongoClient(host="localhost", port=options.port)
+ mongo = pymongo.MongoClient(**mongo_client_opts)
try:
ret = mongo.admin.command("setFeatureCompatibilityVersion", options.fcv_version)
ret = 0 if ret["ok"] == 1 else 1
@@ -1170,11 +1252,11 @@ def crash_server(options, crash_canary, canary_port, local_ops, script_name, cli
ec2 = aws_ec2.AwsEc2()
crash_func = ec2.control_instance
instance_id, _ = get_aws_crash_options(options.crash_options)
- crash_args = ["force-stop", instance_id, 240, True]
+ crash_args = ["force-stop", instance_id, 600, True]
else:
message = "Unsupported crash method '{}' provided".format(options.crash_method)
- LOGGER.error("Unsupported crash method '%s' provided", message)
+ LOGGER.error(message)
return 1, message
# Invoke the crash canary function, right before crashing the server.
@@ -1202,14 +1284,21 @@ def wait_for_mongod_shutdown(data_dir, timeout=120):
return 0
-def get_mongo_client_args(options):
+def get_mongo_client_args(options, host=None, port=None):
""" Returns keyword arg dict used in PyMongo client. """
- mongo_args = {}
+ # Set the serverSelectionTimeoutMS to 600 seconds
+ mongo_args = {"serverSelectionTimeoutMS": 600000}
+ # Set the serverSelectionTimeoutMS to 120 seconds
+ mongo_args["socketTimeoutMS"] = 120000
# Set the writeConcern
mongo_args = yaml.safe_load(options.write_concern)
# Set the readConcernLevel
if options.read_concern_level:
mongo_args["readConcernLevel"] = options.read_concern_level
+ if host:
+ mongo_args["host"] = host
+ if port:
+ mongo_args["port"] = port
return mongo_args
@@ -1218,10 +1307,10 @@ def mongo_shell(mongo_path, work_dir, host_port, mongo_cmds, retries=5, retry_sl
cmds = ("""
cd {};
echo {} | {} {}""".format(
- pipes.quote(work_dir),
- pipes.quote(mongo_cmds),
- pipes.quote(mongo_path),
- host_port))
+ pipes.quote(work_dir),
+ pipes.quote(mongo_cmds),
+ pipes.quote(mongo_path),
+ host_port))
attempt_num = 0
while True:
ret, output = execute_cmd(cmds, use_file=True)
@@ -1346,7 +1435,7 @@ def mongo_validate_canary(mongo, db_name, coll_name, doc):
def mongo_insert_canary(mongo, db_name, coll_name, doc):
""" Inserts a canary document with 'j' True. Returns 0 if successful. """
- LOGGER.info("Inserting canary document %s", doc)
+ LOGGER.info("Inserting canary document %s to DB %s Collection %s", doc, db_name, coll_name)
coll = mongo[db_name][coll_name].with_options(
write_concern=pymongo.write_concern.WriteConcern(j=True))
res = coll.insert_one(doc)
@@ -1400,17 +1489,21 @@ def resmoke_client(work_dir,
repeat_num,
pipes.quote(js_test),
log_output))
- ret, output, proc = None, None, None
+ ret, output = None, None
if no_wait:
- proc = start_cmd(cmds, use_file=True)
+ Processes.create(cmds)
else:
ret, output = execute_cmd(cmds, use_file=True)
- return ret, output, proc
+ return ret, output
def main():
""" Main program. """
+ global _report_json_success
+ global _report_json
+ global _report_json_file
+
atexit.register(exit_handler)
parser = optparse.OptionParser(usage="""
@@ -1482,6 +1575,18 @@ Examples:
action="store_true",
default=False)
+ test_options.add_option("--backupPathBefore",
+ dest="backup_path_before",
+ help="Path where the db_path is backed up before crash recovery,"
+ " defaults to '<rootDir>/data-beforerecovery'",
+ default=None)
+
+ test_options.add_option("--backupPathAfter",
+ dest="backup_path_after",
+ help="Path where the db_path is backed up after crash recovery,"
+ " defaults to '<rootDir>/data-afterrecovery'",
+ default=None)
+
validate_locations = ["local", "remote"]
test_options.add_option("--validate",
dest="validate_collections",
@@ -1651,17 +1756,6 @@ Examples:
help="Set the FeatureCompatibilityVersion of mongod.",
default=None)
- # Program options
- program_options.add_option("--remotePython",
- dest="remote_python",
- help="The python intepreter to use on the remote host"
- " [default: '%default']."
- " To be able to use a python virtual environment,"
- " which has already been provisioned on the remote"
- " host, specify something similar to this:"
- " 'source venv/bin/activate; python'",
- default="python")
-
# Client options
mongo_path = distutils.spawn.find_executable(
"mongo", os.getcwd() + os.pathsep + os.environ["PATH"])
@@ -1725,6 +1819,37 @@ Examples:
default=[])
# Program options
+ program_options.add_option("--configFile",
+ dest="config_file",
+ help="YAML configuration file of program options."
+ " Option values are mapped to command line option names."
+ " The command line option overrides any specified options"
+ " from this file.",
+ default=None)
+
+ program_options.add_option("--saveConfigOptions",
+ dest="save_config_options",
+ help="Save the program options to a YAML configuration file."
+ " If this options is specified the program only saves"
+ " the configuration file and exits.",
+ default=None)
+
+ program_options.add_option("--reportJsonFile",
+ dest="report_json_file",
+ help="Create or update the specified report file upon program"
+ " exit.",
+ default=None)
+
+ program_options.add_option("--remotePython",
+ dest="remote_python",
+ help="The python intepreter to use on the remote host"
+ " [default: '%default']."
+ " To be able to use a python virtual environment,"
+ " which has already been provisioned on the remote"
+ " host, specify something similar to this:"
+ " 'source venv/bin/activate; python'",
+ default="python")
+
program_options.add_option("--remoteSudo",
dest="remote_sudo",
help="Use sudo on the remote host for priveleged operations."
@@ -1763,18 +1888,6 @@ Examples:
action="store_true",
default=False)
- program_options.add_option("--backupPathBefore",
- dest="backup_path_before",
- help="Path where the db_path is backed up before crash recovery,"
- " defaults to '<rootDir>/data-beforerecovery/db'",
- default=None)
-
- program_options.add_option("--backupPathAfter",
- dest="backup_path_after",
- help="Path where the db_path is backed up after crash recovery,"
- " defaults to '<rootDir>/data-afterrecovery/db'",
- default=None)
-
program_options.add_option("--rsyncDest",
dest="rsync_dest",
help=optparse.SUPPRESS_HELP,
@@ -1797,12 +1910,60 @@ Examples:
LOGGER.info("powertest.py invocation: %s", " ".join(sys.argv))
+ # Command line options override the config file options.
+ config_options = None
+ if options.config_file:
+ with open(options.config_file) as ystream:
+ config_options = yaml.safe_load(ystream)
+ LOGGER.info("Loading config file %s with options %s", options.config_file, config_options)
+ # Load the options specified in the config_file
+ parser.set_defaults(**config_options)
+ options, args = parser.parse_args()
+ # Disable this option such that the remote side does not load a config_file.
+ options.config_file = None
+ config_options["config_file"] = None
+
+ if options.save_config_options:
+ # Disable this option such that the remote side does not save the config options.
+ save_config_options = options.save_config_options
+ options.save_config_options = None
+ save_options = {}
+ for opt_group in parser.option_groups:
+ for opt in opt_group.option_list:
+ if getattr(options, opt.dest) != opt.default:
+ save_options[opt.dest] = getattr(options, opt.dest)
+ LOGGER.info("Config options being saved %s", save_options)
+ with open(save_config_options, "w") as ystream:
+ yaml.safe_dump(save_options, ystream, default_flow_style=False)
+ sys.exit(0)
+
script_name = os.path.basename(__file__)
# Print script name and version.
if options.version:
print("{}:{}".format(script_name, __version__))
sys.exit(0)
+ if options.report_json_file:
+ _report_json_file = options.report_json_file
+ if _report_json_file and os.path.exists(_report_json_file):
+ with open(_report_json_file) as jstream:
+ _report_json = json.load(jstream)
+ else:
+ _report_json = {
+ "failures": 0,
+ "results": [
+ {"status": "fail",
+ "test_file": __name__,
+ "exit_code": 0,
+ "elapsed": 0,
+ "start": int(time.time()),
+ "end": int(time.time())}
+ ]
+ }
+ LOGGER.debug("Updating/creating report JSON %s", _report_json)
+ # Disable this option such that the remote side does not generate report.json
+ options.report_json_file = None
+
# Setup the crash options
if ((options.crash_method == "aws_ec2" or options.crash_method == "mpower") and
options.crash_options is None):
@@ -1851,6 +2012,12 @@ Examples:
if options.rsync_data:
rsync_cmd = "rsync_data"
+ backup_path_before = options.backup_path_before
+ if not backup_path_before:
+ backup_path_before = "{}/data-beforerecovery".format(options.root_dir)
+ backup_path_after = options.backup_path_after
+ if not backup_path_after:
+ backup_path_after = "{}/data-afterrecovery".format(options.root_dir)
else:
rsync_cmd = ""
rsync_opt = ""
@@ -1965,8 +2132,8 @@ Examples:
# remote host's invocation of this script.
elif isinstance(option_value, str) and re.search("\"|'| ", option_value):
option_value = "'{}'".format(option_value)
- # The tuple options need to be changed to a string.
- elif isinstance(option_value, tuple):
+ # The tuple, list or set options need to be changed to a string.
+ elif isinstance(option_value, (tuple, list, set)):
option_value = " ".join(map(str, option_value))
client_args = "{} {} {}".format(client_args, option.get_opt_string(), option_value)
@@ -2013,20 +2180,20 @@ Examples:
temp_client_files = []
+ validate_canary_local = False
if options.canary and loop_num > 1:
- canary_opt = "--docForCanary \"{}\"".format(canary_doc)
- validate_canary_cmd = "validate_canary" if options.canary else ""
+ if options.canary == "remote":
+ canary_opt = "--docForCanary \"{}\"".format(canary_doc)
+ validate_canary_cmd = "validate_canary" if options.canary else ""
+ else:
+ validate_canary_local = True
else:
canary_opt = ""
# Since rsync requires Posix style paths, we do not use os.path.join to
# construct the rsync destination directory.
if rsync_cmd:
- if options.backup_path_before:
- rsync_dest = options.backup_path_before
- else:
- rsync_dest = "{}/data-afterrecovery".format(options.root_dir)
- rsync_opt = " --rsyncDest {}".format(rsync_dest)
+ rsync_opt = "--rsyncDest {}".format(backup_path_before)
# Optionally, rsync the pre-recovery database.
# Start monogd on the secret port.
@@ -2060,20 +2227,30 @@ Examples:
if ret:
sys.exit(ret)
+ # Optionally validate canary document locally.
+ if validate_canary_local:
+ mongo = pymongo.MongoClient(
+ **get_mongo_client_args(options, host=mongod_host, port=secret_port))
+ ret = mongo_validate_canary(
+ mongo, options.db_name, options.collection_name, canary_doc)
+ LOGGER.info("Local canary validation: %d", ret)
+ if ret:
+ sys.exit(ret)
+
# Optionally, run local validation of collections.
if options.validate_collections == "local":
host_port = "{}:{}".format(mongod_host, secret_port)
- new_config_file = NamedTempFile.create(suffix=".yml", dir="tmp")
+ new_config_file = NamedTempFile.create(suffix=".yml", directory="tmp")
temp_client_files.append(new_config_file)
validation_test_data = {"skipValidationOnNamespaceNotFound": True}
new_resmoke_config(with_external_server, new_config_file, validation_test_data)
- ret, output, _ = resmoke_client(
+ ret, output = resmoke_client(
mongo_repo_root_dir,
mongo_path,
host_port,
"jstests/hooks/run_validate_collections.js",
new_config_file)
- LOGGER.info("Collection validation: %d %s", ret, output)
+ LOGGER.info("Local collection validation: %d %s", ret, output)
if ret:
sys.exit(ret)
@@ -2094,11 +2271,7 @@ Examples:
# Since rsync requires Posix style paths, we do not use os.path.join to
# construct the rsync destination directory.
if rsync_cmd:
- if options.backup_path_after:
- rsync_dest = options.backup_path_after
- else:
- rsync_dest = "{}/data-afterrecovery".format(options.root_dir)
- rsync_opt = " --rsyncDest {}".format(rsync_dest)
+ rsync_opt = "--rsyncDest {}".format(backup_path_after)
# Optionally, rsync the post-recovery database.
# Start monogd on the standard port.
@@ -2110,7 +2283,7 @@ Examples:
" {}"
" {}"
" start_mongod").format(
- rsync_opt, standard_port, use_replica_set, rsync_cmd)
+ rsync_opt, standard_port, use_replica_set, rsync_cmd)
ret, output = call_remote_operation(
local_ops,
options.remote_python,
@@ -2124,16 +2297,15 @@ Examples:
# Start CRUD clients
host_port = "{}:{}".format(mongod_host, standard_port)
- crud_procs = []
for i in xrange(options.num_crud_clients):
if options.config_crud_client == with_external_server:
- crud_config_file = NamedTempFile.create(suffix=".yml", dir="tmp")
+ crud_config_file = NamedTempFile.create(suffix=".yml", directory="tmp")
crud_test_data["collectionName"] = "{}-{}".format(options.collection_name, i)
new_resmoke_config(
with_external_server, crud_config_file, crud_test_data, eval_str)
else:
crud_config_file = options.config_crud_client
- _, _, proc = resmoke_client(
+ _, _ = resmoke_client(
work_dir=mongo_repo_root_dir,
mongo_path=mongo_path,
host_port=host_port,
@@ -2142,21 +2314,18 @@ Examples:
repeat_num=100,
no_wait=True,
log_file="crud_{}.log".format(i))
- crud_procs.append(proc)
- if crud_procs:
- LOGGER.info(
- "****Started %d CRUD client(s)****", options.num_crud_clients)
+ if options.num_crud_clients:
+ LOGGER.info("****Started %d CRUD client(s)****", options.num_crud_clients)
# Start FSM clients
- fsm_procs = []
for i in xrange(options.num_fsm_clients):
- fsm_config_file = NamedTempFile.create(suffix=".yml", dir="tmp")
+ fsm_config_file = NamedTempFile.create(suffix=".yml", directory="tmp")
fsm_test_data["dbNamePrefix"] = "fsm-{}".format(i)
# Do collection validation only for the first FSM client.
fsm_test_data["validateCollections"] = True if i == 0 else False
new_resmoke_config(with_external_server, fsm_config_file, fsm_test_data, eval_str)
- _, _, proc = resmoke_client(
+ _, _ = resmoke_client(
work_dir=mongo_repo_root_dir,
mongo_path=mongo_path,
host_port=host_port,
@@ -2165,9 +2334,8 @@ Examples:
repeat_num=100,
no_wait=True,
log_file="fsm_{}.log".format(i))
- fsm_procs.append(proc)
- if fsm_procs:
+ if options.num_fsm_clients:
LOGGER.info("****Started %d FSM client(s)****", options.num_fsm_clients)
# Crash the server. A pre-crash canary document is optionally written to the DB.
@@ -2175,8 +2343,8 @@ Examples:
if options.canary:
canary_doc = {"x": time.time()}
orig_canary_doc = copy.deepcopy(canary_doc)
- mongo_opts = get_mongo_client_args(options)
- mongo = pymongo.MongoClient(host=mongod_host, port=standard_port, **mongo_opts)
+ mongo = pymongo.MongoClient(
+ **get_mongo_client_args(options, host=mongod_host, port=standard_port))
crash_canary["function"] = mongo_insert_canary
crash_canary["args"] = [
mongo,
@@ -2193,7 +2361,7 @@ Examples:
time.sleep(10)
# Kill any running clients and cleanup temporary files.
- kill_processes(crud_procs + fsm_procs)
+ Processes.kill_all()
for temp_file in temp_client_files:
NamedTempFile.delete(temp_file)
@@ -2201,7 +2369,7 @@ Examples:
if options.crash_method == "aws_ec2":
ec2 = aws_ec2.AwsEc2()
ret, aws_status = ec2.control_instance(
- mode="start", image_id=instance_id, wait_time_secs=240, show_progress=True)
+ mode="start", image_id=instance_id, wait_time_secs=600, show_progress=True)
LOGGER.info("Start instance: %d %s****", ret, aws_status)
if ret:
raise Exception("Start instance failed: {}".format(aws_status))
@@ -2226,6 +2394,8 @@ Examples:
LOGGER.info("****Completed test loop %d test time %d seconds****", loop_num, test_time)
if loop_num == options.num_loops or test_time >= options.test_time:
break
+
+ _report_json_success = True
sys.exit(0)