1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
|
#!/usr/bin/env python3
"""Utility to validate resmoke tests runtime."""
import json
import sys
from collections import namedtuple
from datetime import datetime, timedelta
from statistics import mean
from typing import Dict, List
import click
import structlog
from buildscripts.resmokelib.testing.report import TestInfo, TestReport
from buildscripts.resmokelib.utils import get_task_name_without_suffix
from buildscripts.util.cmdutils import enable_logging
from evergreen import RetryingEvergreenApi, TestStats
LOGGER = structlog.get_logger("buildscripts.resmoke_tests_runtime_validate")
LOOK_BACK_NUM_DAYS = 20
BURN_IN_PREFIX = "burn_in:"
HISTORIC_MAX_MULTIPLIER = 1.5
IGNORE_LESS_THAN_SECS = 15
_TestData = namedtuple("TestData", ["test_file", "duration"])
def parse_resmoke_report(report_file: str) -> List[TestInfo]:
"""Get js tests info from resmoke report json."""
with open(report_file, "r") as fh:
report_data = json.load(fh)
test_report = TestReport.from_dict(report_data)
return [test_info for test_info in test_report.test_infos if "jstests" in test_info.test_file]
def get_historic_stats(evg_api_config: str, project_id: str, test_files: List[str], task_name: str,
build_variant: str) -> List[TestStats]:
"""Get historic test stats."""
evg_api = RetryingEvergreenApi.get_api(config_file=evg_api_config)
before_date = datetime.today()
after_date = before_date - timedelta(days=LOOK_BACK_NUM_DAYS)
base_task_name = get_task_name_without_suffix(task_name, build_variant).replace(
BURN_IN_PREFIX, "")
return evg_api.test_stats_by_project(project_id=project_id, after_date=after_date,
before_date=before_date, tests=test_files,
tasks=[base_task_name], variants=[build_variant])
def make_stats_map(stats: List[_TestData]) -> Dict[str, List[float]]:
"""Make test stats map."""
stats_map = {}
for stat in stats:
if stat.test_file in stats_map:
stats_map[stat.test_file].append(stat.duration)
else:
stats_map[stat.test_file] = [stat.duration]
return stats_map
@click.command()
@click.option("--resmoke-report-file", type=str, required=True,
help="Location of resmoke's report JSON file.")
@click.option("--evg-api-config", type=str, required=True,
help="Location of evergreen api configuration.")
@click.option("--project-id", type=str, required=True, help="Evergreen project id.")
@click.option("--build-variant", type=str, required=True, help="Evergreen build variant name.")
@click.option("--task-name", type=str, required=True, help="Evergreen task name.")
def main(resmoke_report_file: str, evg_api_config: str, project_id: str, build_variant: str,
task_name: str) -> None:
"""Compare resmoke tests runtime with historic stats."""
enable_logging(verbose=False)
current_test_infos = parse_resmoke_report(resmoke_report_file)
current_stats_map = make_stats_map([
_TestData(test_info.test_file, test_info.end_time - test_info.start_time)
for test_info in current_test_infos
])
historic_stats = get_historic_stats(evg_api_config, project_id, list(current_stats_map.keys()),
task_name, build_variant)
historic_stats_map = make_stats_map([
_TestData(test_stats.test_file, test_stats.avg_duration_pass)
for test_stats in historic_stats
])
failed = False
for test, stats in current_stats_map.items():
current_mean = mean(stats)
if current_mean < IGNORE_LESS_THAN_SECS:
continue
historic_test_stats = historic_stats_map.get(test)
if historic_test_stats:
historic_max = max(historic_test_stats)
target = historic_max * HISTORIC_MAX_MULTIPLIER
if current_mean > target:
LOGGER.error("Found long running test.", test_file=test,
current_mean_time=current_mean, maximum_expected_time=target,
historic_max_time=historic_max)
failed = True
LOGGER.info("Done comparing resmoke tests runtime with historic stats.")
if failed:
percent = int((HISTORIC_MAX_MULTIPLIER - 1) * 100)
LOGGER.error(
f"The test failed due to its runtime taking {percent}% more than the recent max"
" and can negatively contribute to the future patch build experience."
" Consider checking if there is an unexpected regression.")
LOGGER.error("If the test is being intentionally expanded, please split it up into separate"
" JS files that run as separate tests.")
LOGGER.error(
"If you believe the test has inherently large variability, please consider writing"
" a new test instead of modifying this one.")
LOGGER.error("For any other questions or concerns, please reach out to #server-testing.")
sys.exit(1)
if __name__ == '__main__':
main() # pylint: disable=no-value-for-parameter
|