##################################################### # A note on expansions # ##################################################### # Expansions usually appear in the form ${key|default} # If 'key' is found in the executor's map of currently known # expansions, the corresponding value is used. If the key can # not be found, the default is used. # # Arbitrary expansions can be specified in the YAML configuration # files in the following places: # - The 'expansions' field for buildvariants (branch file) # - The 'expansions' field for distros (distros file) # # A number of 'built-in' expansions are also available for use; these include: # - environment variables available on the host machine # - 'workdir' (references the executor's work directory). # - 'task_id' (references the task id of the task the executor is working on). # - 'build_variant' (references the executing task's buildvariant). # - 'config_root' (references the root directory for the executor's configuration artifacts). stepback: true command_type: system # Files that match an ignore-list pattern will not trigger a build, if they're the only modified # files in the patch. ignore: - ".*" - "!.clang-format" - "!.eslintrc.yml" - "*.md" - "*.rst" - "*.txt" - "/distsrc/**" - "/docs/**" - "/etc/*.yml" - "!/etc/evergreen.yml" - "README" ## Some variables for convenience: variables: # Used when the tests it runs depend only on mongod, mongos, the mongo shell and the tools. - &task_template name: template depends_on: - name: compile commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --help run_multiple_jobs: false max_jobs: 0 # If set in combination with run_multiple_jobs, will cap the number of jobs used. # Used for tests that invoke resmoke.py and require no additional setup. - &task_depending_on_all_template <<: *task_template depends_on: - name: compile_all - &benchmark_template name: benchmark_template depends_on: - name: compile_benchmarks commands: - func: "do benchmark setup" - func: "run tests" vars: resmoke_args: --help run_multiple_jobs: false - func: "send benchmark results" - &generate_compile_expansions command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose # We get the raw version string (r1.2.3-45-gabcdef) from git MONGO_VERSION=$(git describe) # If this is a patch build, we add the patch version id to the version string so we know # this build was a patch, and which evergreen task it came from if [ "${is_patch|}" = "true" ]; then MONGO_VERSION="$MONGO_VERSION-patch-${version_id}" fi # This script converts the generated version string into a sanitized version string for # use by scons and uploading artifacts as well as information about for the scons cache. ${activate_virtualenv} MONGO_VERSION=$MONGO_VERSION SCONS_CACHE_MODE=${scons_cache_mode|nolinked} USE_SCONS_CACHE=${use_scons_cache|false} $python buildscripts/generate_compile_expansions.py --out compile_expansions.yml - &apply_compile_expansions command: expansions.update params: file: src/compile_expansions.yml - &jepsen_config_vars jepsen_key_time_limit: --key-time-limit 15 jepsen_protocol_version: --protocol-version 1 jepsen_read_concern: "" jepsen_read_with_find_and_modify: "" jepsen_storage_engine: "" jepsen_test_name: "" # Empirically, we've had greater success in reproducing the issues found in MongoDB versions # 3.4.0-rc3 and 3.4.0-rc4 when running Jepsen with at least --time-limit=600. jepsen_time_limit: --time-limit 1200 jepsen_write_concern: "" # Template for running Jepsen tests - &run_jepsen_template name: run_jepsen_template depends_on: - name: compile commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars - &jstestfuzz_config_vars resmoke_args: --help # resmoke_args needs to be overridden to specify one of the jstestfuzz suites should_shuffle: false continue_on_failure: false - &run_jstestfuzz_tests func: "run tests" # Terminate the function when there has been no output to stdout for 30 minutes. E.g. when something is stuck in an infinite loop. # resmoke.py writes the test output to logkeeper and only writes to stdout when starting the next test. # resmoke.py not producing output on stdout means that the test is still running and presumably not going to finish. # Note that timeout_secs is different from exec_timeout_secs, which applies to a task and times out regardless of whether output has been written to stdout. timeout_secs: 1800 vars: <<: *jstestfuzz_config_vars # Used for tests that invoke 'resmoke.py --suites=jstestfuzz*'. - &jstestfuzz_template name: jstestfuzz_template exec_timeout_secs: 14400 # Time out the task if it runs for more than 4 hours. depends_on: - name: compile commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" - *run_jstestfuzz_tests # Templates used by powercycle - &powercycle_remote_credentials private_key_file: $(${posix_workdir})/powercycle.pem private_key_remote: ${powercycle_private_key} aws_key_remote: ${powercycle_aws_key} aws_secret_remote: ${powercycle_aws_secret} - &powercycle_ec2_instance aws_ec2_yml: aws_ec2.yml ec2_expire_hours: "24" ec2_monitor_files: proc.json system.json monitor_proc_file: proc.json monitor_system_file: system.json remote_dir: /log/powercycle secret_port: "20001" security_group_ids: ${powercycle_aws_security_group_id} security_groups: mci powercycle_testing subnet_id: ${powercycle_aws_subnet_id} ssh_identity: -i ${private_key_file} ssh_key_id: ${powercycle_ssh_key_id} standard_port: "20000" virtualenv_dir: venv_powercycle windows_crash_cmd: \"notmyfault/notmyfaultc64.exe -accepteula crash 1\" windows_crash_dl: https://download.sysinternals.com/files/NotMyFault.zip windows_crash_dir: notmyfault windows_crash_zip: notmyfault.zip - &powercycle_expansions params: updates: - key: backup_path_after value: ${remote_dir}/afterrecovery - key: backup_path_before value: ${remote_dir}/beforerecovery - key: backup_artifacts value: ${remote_dir}/afterrecovery* ${remote_dir}/beforerecovery* - key: db_path value: /data/db - key: log_path value: ${remote_dir}/mongod.log - key: exit_file value: powercycle_exit.yml - &powercycle_test ec2_artifacts: ${log_path} ${db_path} ${backup_artifacts} program_options: --exitYamlFile=${exit_file} --logLevel=info --backupPathBefore=${backup_path_before} --backupPathAfter=${backup_path_after} connection_options: --sshUserHost=${private_ip_address} --sshConnection=\"${ssh_identity} ${ssh_connection_options}\" test_options: --testLoops=15 --seedDocNum=10000 --rsync --rsyncExcludeFiles=diagnostic.data/metrics.interim* --validate=local --canary=local crash_options: --crashMethod=internal --crashOption=${windows_crash_cmd} --crashWaitTime=45 --jitterForCrashWaitTime=5 --instanceId=${instance_id} client_options: --numCrudClients=20 --numFsmClients=20 mongodb_options: --rootDir=${remote_dir}-${task_id} --mongodbBinDir=${remote_dir} mongod_options: --mongodUsablePorts ${standard_port} ${secret_port} --dbPath=${db_path} --logPath=${log_path} mongod_extra_options: --mongodOptions=\"--setParameter enableTestCommands=1\" ####################################### # Functions # ####################################### functions: "git get project" : &git_get_project command: git.get_project params: directory: ${git_project_directory|src} revisions: # for each module include revision as : ${_rev} enterprise: ${enterprise_rev} "fetch artifacts" : &fetch_artifacts command: s3.get params: aws_key: ${aws_key} aws_secret: ${aws_secret} remote_file: ${project}/${build_variant}/${revision}/artifacts/${build_id}.tgz bucket: mciuploads extract_to: src "fetch binaries" : &fetch_binaries command: s3.get params: aws_key: ${aws_key} aws_secret: ${aws_secret} remote_file: ${project}/${build_variant}/${revision}/binaries/mongo-${build_id}.${ext|tgz} bucket: mciuploads local_file: src/mongo-binaries.tgz "extract binaries" : &extract_binaries command: shell.exec params: working_dir: src script: | set -o errexit ${decompress} mongo-binaries.tgz cp mongodb*/bin/* . "check binary version" : &check_binary_version command: shell.exec params: working_dir: src script: | set -o errexit mongo_binary=$(find mongodb*/bin -name mongo${exe}) # There should only be one mongo shell if [ $(echo $mongo_binary | wc -w) -ne 1 ]; then echo "There is more than 1 extracted mongo binary: $mongo_binary" exit 1 fi ${activate_virtualenv} bin_ver=$($python -c "import yaml; print(yaml.safe_load(open('compile_expansions.yml'))['version']);" | tr -d '[ \r\n]') # Due to SERVER-23810, we cannot use $mongo_binary --quiet --nodb --eval "version();" mongo_ver=$($mongo_binary --version | perl -pe '/version v(.*)$/; $_ = $1;' | tr -d '[ \r\n]') # The versions must match if [ "$bin_ver" != "$mongo_ver" ]; then echo "The mongo version is $mongo_ver, expected version is $bin_ver" exit 1 fi "fetch benchmarks" : &fetch_benchmarks command: s3.get params: aws_key: ${aws_key} aws_secret: ${aws_secret} remote_file: ${project}/${build_variant}/${revision}/benchmarks/${build_id}.tgz bucket: mciuploads extract_to: src "get buildnumber" : &get_buildnumber command: keyval.inc params: key: "${build_variant}_v3.6" destination: "builder_num" "run diskstats": &run_diskstats command: shell.exec params: background: true system_log: true script: | set -o errexit set -o verbose # On Windows we can use typeperf.exe to dump performance counters. if [ "Windows_NT" = "$OS" ]; then typeperf -qx PhysicalDisk | grep Disk | grep -v _Total > disk_counters.txt typeperf -cf disk_counters.txt -si 5 -o mongo-diskstats # Linux: iostat -t option for timestamp. elif iostat -tdmx > /dev/null 2>&1; then iostat -tdmx 5 > mongo-diskstats # OSX: Simulate the iostat timestamp. elif iostat -d > /dev/null 2>&1; then iostat -d -w 5 | while IFS= read -r line; do printf '%s %s\n' "$(date +'%m/%d/%Y %H:%M:%S')" "$line" >> mongo-diskstats; done # Check if vmstat -t is available. elif vmstat -td > /dev/null 2>&1; then vmstat -td 5 > mongo-diskstats # Check if vmstat -T d is available. elif vmstat -T d > /dev/null 2>&1; then vmstat -T d 5 > mongo-diskstats else printf "Cannot collect mongo-diskstats on this platform\n" fi "collect system resource info": &collect_system_resource_info command: shell.exec params: working_dir: src background: true system_log: true script: | ${activate_virtualenv} $python buildscripts/collect_resource_info.py -o system_resource_info.json -i 5 # Run a monitor process as a background, system task to periodically # display how many threads interesting processes are using. "monitor process threads": &monitor_process_threads command: shell.exec params: background: true system_log: true script: | proc_list="(bsondump|java|lein|mongo|python|_test$|_test\.exe$)" if [ "Windows_NT" = "$OS" ]; then get_pids() { proc_pids=$(tasklist /fo:csv | awk -F'","' '{x=$1; gsub("\"","",x); print $2, x}' | grep -i -E $1 | cut -f1 -d ' '); } get_process_info() { proc_name=""; proc_info=$(wmic process where "ProcessId=\"$1\"" get "Name,ProcessId,ThreadCount" /format:csv 2> /dev/null | grep $1); if [ ! -z $proc_info ]; then proc_name=$(echo $proc_info | cut -f2 -d ','); proc_threads=$(echo $proc_info | cut -f4 -d ','); fi; } else get_pids() { proc_pids=$(pgrep $1); } get_process_info() { proc_name=$(ps -p $1 -o comm=); # /proc is available on Linux platforms if [ -f /proc/$1/status ]; then ${set_sudo} proc_threads=$($sudo grep Threads /proc/$1/status | sed "s/\s//g" | cut -f2 -d ":"); else proc_threads=$(ps -AM $1 | grep -vc PID); fi; } fi while [ 1 ] do get_pids $proc_list if [ ! -z "$proc_pids" ]; then printf "Running process/thread counter\n" printf "PROCESS\tPID\tTHREADS\n" fi for pid in $proc_pids do get_process_info $pid if [ ! -z "$proc_name" ]; then printf "$proc_name\t$pid\t$proc_threads\n" fi done sleep 60 done "set up credentials" : &set_up_credentials command: shell.exec params: working_dir: src silent: true script: | cat > mci.buildlogger < ${private_key_file|/dev/null} chmod 0600 ${private_key_file|/dev/null} fi if [ ! -d ~.aws ]; then mkdir -p ~/.aws fi # If ${aws_profile_remote} is not specified then the config & credentials are # stored in the 'default' profile. aws_profile="${aws_profile_remote|default}" # The profile in the config file is specified as [profile ], except # for [default], see http://boto3.readthedocs.io/en/latest/guide/configuration.html if [ $aws_profile = "default" ] ; then aws_profile_config="[default]" else aws_profile_config="[profile $aws_profile]" fi cat <> ~/.aws/config $aws_profile_config region = us-east-1 EOF # The profile in the credentials file is specified as []. cat <> ~/.aws/credentials [$aws_profile] aws_access_key_id = ${aws_key_remote} aws_secret_access_key = ${aws_secret_remote} EOF cat < ~/.boto [Boto] https_validate_certificates = False EOF "call BF Suggestion service": command: shell.exec params: working_dir: src shell: bash silent: true script: | report_file="report.json" # Check if the report file exists and has failures. if [ -f $report_file ] && grep -Eq "\"failures\": [1-9]" $report_file; then # Calling the BF Suggestion server endpoint to start feature extraction. payload="{\"task_id\": \"${task_id}\", \"execution\": ${execution}}" echo "Sending task info to the BF suggestion service" # The --user option is passed through stdin to avoid showing in process list. user_option="--user ${bfsuggestion_user}:${bfsuggestion_password}" curl --header "Content-Type: application/json" \ --data "$payload" \ --max-time 10 \ --silent \ --show-error \ --config - \ https://bfsuggestion.corp.mongodb.com/tasks <<< $user_option echo "Request to BF Suggestion service status: $?" fi "upload debugsymbols" : &upload_debugsymbols command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/mongo-debugsymbols.tgz remote_file: ${project}/${build_variant}/${revision}/debugsymbols/debugsymbols-${build_id}.${ext|tgz} bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} "fetch debugsymbols archive" : &fetch_debugsymbols_archive command: s3.get params: aws_key: ${aws_key} aws_secret: ${aws_secret} remote_file: ${project}/${build_variant}/${revision}/debugsymbols/debugsymbols-${build_id}.${ext|tgz} bucket: mciuploads local_file: src/mongo-debugsymbols.tgz "extract debugsymbols" : &extract_debugsymbols command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose # Debug symbols are not created for every variant if [ ! -f mongo-debugsymbols.tgz ]; then exit fi ${decompress} mongo-debugsymbols.tgz files="mongo mongod mongos" file_exts="debug dSYM pdb" for file_ext in $file_exts do for file in $files do mv mongodb*/$file.$file_ext . 2>/dev/null || true done rm -r mongodb*/*.$file_ext 2>/dev/null || true done rm mongo-debugsymbols.tgz 2>/dev/null || true "fetch and build OpenSSL" : command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose if [ "${build_openssl|}" = "true" ]; then ${activate_virtualenv} bash buildscripts/fetch_and_build_openssl.sh "$python" "${openssl_make_flags|}" "${openssl_config_flags|}" fi "build new tools" : command: shell.exec params: working_dir: src/src/mongo/gotools/src/github.com/mongodb/mongo-tools script: | set -o verbose set -o errexit # make sure newlines in the scripts are handled correctly by windows if [ "Windows_NT" = "$OS" ]; then set -o igncr fi; # set_goenv provides set_goenv(), print_ldflags() and print_tags() used below . ./set_goenv.sh GOROOT="" set_goenv || exit env | grep ^GO go version build_tools="bsondump mongostat mongofiles mongoexport mongoimport mongorestore mongodump mongotop" if [ "${build_mongoreplay}" = "true" ]; then build_tools="$build_tools mongoreplay" fi for i in $build_tools; do go build $(buildflags) -ldflags "$(print_ldflags)" ${args} -tags "$(print_tags ${tooltags})" -o "../../../../../../mongo-tools/$i${exe|}" $i/main/$i.go "../../../../../../mongo-tools/$i${exe|}" --version file "../../../../../../mongo-tools/$i${exe|}" done mkdir -p ../../../../../../mongo-tools/distsrc cp THIRD-PARTY-NOTICES ../../../../../../mongo-tools/distsrc/THIRD-PARTY-NOTICES.gotools # This function should only be called from patch-build-only tasks. "get added and modified patch files": command: shell.exec params: working_dir: src shell: bash script: | set -o verbose set -o errexit git diff HEAD --name-only --line-prefix="${workdir}/src/" --diff-filter=d >> modified_and_created_patch_files.txt if [ -d src/mongo/db/modules/enterprise ]; then pushd src/mongo/db/modules/enterprise git diff HEAD --name-only --line-prefix="${workdir}/src/src/mongo/db/modules/enterprise/" --diff-filter=d >> ~1/modified_and_created_patch_files.txt popd fi "send benchmark results" : command: json.send params: name: perf file: src/perf.json "do setup" : - *fetch_artifacts - *fetch_binaries - *extract_binaries - *check_binary_version - *get_buildnumber - *set_up_credentials - *run_diskstats - *monitor_process_threads - *collect_system_resource_info "do benchmark setup" : - *git_get_project - *get_buildnumber - *set_up_credentials - *fetch_benchmarks "set up virtualenv" : command: shell.exec type: test params: script: | # exit immediately if virtualenv is not found set -o errexit set -o verbose python_loc=$(which ${python|/opt/mongodbtoolchain/v2/bin/python2}) python3_loc=$(which ${python|/opt/mongodbtoolchain/v2/bin/python3}) if [ "Windows_NT" = "$OS" ]; then python_loc=$(cygpath -w $python_loc) python3_loc=$(cygpath -w c:/python/Python36/python.exe) fi # Set up virtualenv in ${workdir} virtualenv --python "$python_loc" --system-site-packages "${workdir}/venv" # Add virtualenv for python3 in ${workdir} virtualenv --python "$python3_loc" --system-site-packages "${workdir}/venv_3" "run tests" : - command: expansions.update params: updates: - key: aws_key_remote value: ${mongodatafiles_aws_key} - key: aws_profile_remote value: mongodata_aws - key: aws_secret_remote value: ${mongodatafiles_aws_secret} - *set_up_remote_credentials - command: shell.exec type: test params: working_dir: src shell: bash script: | set -o errexit set -o verbose # Set the TMPDIR environment variable to be a directory in the task's working # directory so that temporary files created by processes spawned by resmoke.py get # cleaned up after the task completes. This also ensures the spawned processes # aren't impacted by limited space in the mount point for the /tmp directory. export TMPDIR="${workdir}/tmp" mkdir -p $TMPDIR if [ ${disable_unit_tests|false} = "false" ]; then # activate the virtualenv if it has been set up ${activate_virtualenv} pip install boto3==1.5.27 if [ -f /proc/self/coredump_filter ]; then # Set the shell process (and its children processes) to dump ELF headers (bit 4), # anonymous shared mappings (bit 1), and anonymous private mappings (bit 0). echo 0x13 > /proc/self/coredump_filter if [ -f /sbin/sysctl ]; then # Check that the core pattern is set explicitly on our distro image instead # of being the OS's default value. This ensures that coredump names are consistent # across distros and can be picked up by Evergreen. core_pattern=$(/sbin/sysctl -n "kernel.core_pattern") if [ "$core_pattern" = "dump_%e.%p.core" ]; then echo "Enabling coredumps" ulimit -c unlimited fi fi fi extra_args="" if [ ${run_multiple_jobs|false} = true ]; then processor_architecture=$(uname -m) num_jobs_available=${num_jobs_available|1} # Reduce the number of available jobs by half when running any sharding*, replica_sets* # and select jsCore passthrough tasks on Windows and ARM to avoid overwhelming test hosts. if [ "Windows_NT" = "$OS" ] || [ "aarch64" = $processor_architecture ]; then case "${task_name}" in replica_sets_initsync_jscore_passthrough \ |replica_sets_initsync_static_jscore_passthrough \ |replica_sets_resync_static_jscore_passthrough \ |replica_sets* \ |sharding* \ ) num_jobs_available=$((${num_jobs_available|2} / 2)) echo "Reducing jobs from ${num_jobs_available|1} to $num_jobs_available" ;; esac fi # Reduce the number of available jobs by two-thirds when running the # secondary_reads_passthrough task on the Enterprise RHEL 6.2 InMemory build variant. if [[ "${build_variant}" = "enterprise-rhel-62-64-bit-inmem" \ && "${task_name}" = "secondary_reads_passthrough" ]]; then num_jobs_available=$((${num_jobs_available|3} / 3)) echo "Reducing jobs from ${num_jobs_available|1} to $num_jobs_available" fi if [ ${max_jobs|0} -gt 0 ] && [ ${max_jobs|0} -lt $num_jobs_available ]; then extra_args="$extra_args --jobs=${max_jobs}" else extra_args="$extra_args --jobs=$num_jobs_available" fi fi if [ ${should_shuffle|true} = true ]; then extra_args="$extra_args --shuffle" fi if [ ${continue_on_failure|true} = true ]; then extra_args="$extra_args --continueOnFailure" fi # Default storageEngineCacheSizeGB to 1. Override on individual test config if needed. # resmoke will assign to the appropriate parameter on storage engines that support it. set +o errexit echo "${resmoke_args}" | grep -q storageEngineCacheSizeGB if [ $? -eq 1 ]; then extra_args="$extra_args --storageEngineCacheSizeGB=1" fi set -o errexit path_value="$PATH" if [ ${variant_path_suffix} ]; then path_value="$path_value:${variant_path_suffix}" fi if [ ${task_path_suffix} ]; then path_value="$path_value:${task_path_suffix}" fi if [ "${is_patch|}" = "true" ]; then extra_args="$extra_args --tagFile=etc/test_lifecycle.yml --patchBuild" else extra_args="$extra_args --tagFile=etc/test_retrial.yml" fi # The "resmoke_wrapper" expansion is used by the 'burn_in_tests' task to wrap the resmoke.py # invocation. It doesn't set any environment variables and should therefore come last in # this list of expansions. set +o errexit PATH="$path_value" \ AWS_PROFILE=${aws_profile_remote} \ ${gcov_environment} \ ${lang_environment} \ ${san_options} \ ${san_symbolizer} \ ${snmp_config_path} \ ${resmoke_wrapper} \ $python buildscripts/evergreen_run_tests.py \ ${resmoke_args} \ $extra_args \ ${test_flags} \ --log=buildlogger \ --staggerJobs=on \ --taskId=${task_id} \ --taskName=${task_name} \ --executionNumber=${execution} \ --projectName=${project} \ --variantName=${build_variant} \ --distroId=${distro_id} \ --gitRevision=${revision} \ --archiveFile=archive.json \ --reportFile=report.json \ --perfReportFile=perf.json resmoke_exit_code=$? set -o errexit # 74 is exit code for IOError on POSIX systems, which is raised when the machine is # shutting down. # # 75 is exit code resmoke.py uses when the log output would be incomplete due to failing # to communicate with logkeeper. if [[ $resmoke_exit_code = 74 || $resmoke_exit_code = 75 ]]; then echo $resmoke_exit_code > run_tests_infrastructure_failure exit 0 fi exit $resmoke_exit_code fi # end if ${disable_unit_tests} # The existence of the "run_tests_infrastructure_failure" file indicates this failure isn't # directly actionable. We use type=setup rather than type=system or type=test for this command # because we don't intend for any human to look at this failure. - command: shell.exec type: setup params: working_dir: src script: | set -o verbose if [ -f run_tests_infrastructure_failure ]; then exit $(cat run_tests_infrastructure_failure) fi "scons compile" : command: shell.exec type: test params: working_dir: src script: | set -o errexit set -o verbose rm -rf ${install_directory|/data/mongo-install-directory} extra_args="" if [ "${targets}" = "all" ] && [ -n "${num_scons_compile_all_jobs_available|}" ]; then echo "Changing SCons to run with --jobs=${num_scons_compile_all_jobs_available|}" extra_args="$extra_args --jobs=${num_scons_compile_all_jobs_available|}" fi ${activate_virtualenv} ${compile_env|} $python ./buildscripts/scons.py ${compile_flags|} ${task_compile_flags|} ${scons_cache_args|} $extra_args ${targets} ${additional_targets|} MONGO_VERSION=${version} "generate compile expansions" : command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose # We get the raw version string (r1.2.3-45-gabcdef) from git MONGO_VERSION=$(git describe) # If this is a patch build, we add the patch version id to the version string so we know # this build was a patch, and which evergreen task it came from if [ "${is_patch|}" = "true" ]; then MONGO_VERSION="$MONGO_VERSION-patch-${version_id}" fi # This script converts the generated version string into a sanitized version string for # use by scons and uploading artifacts as well as information about for the scons cache. ${activate_virtualenv} MONGO_VERSION=$MONGO_VERSION SCONS_CACHE_MODE=${scons_cache_mode|nolinked} USE_SCONS_CACHE=${use_scons_cache|false} $python buildscripts/generate_compile_expansions.py --out compile_expansions.yml "apply compile expansions" : command: expansions.update params: file: src/compile_expansions.yml "do jepsen setup" : - command: shell.exec params: working_dir: src script: | set -o errexit # Build libfaketime. A version of libfaketime at least as new as v0.9.6-9-g75896bd is # required to use the FAKETIME_NO_CACHE and FAKETIME_TIMESTAMP_FILE environment variables. # Additionally, a version of libfaketime containing the changes mentioned in SERVER-29336 # is required to avoid needing to use libfaketimeMT.so.1 and serializing all calls to # fake_clock_gettime() with a mutex. git clone --branch=for-jepsen --depth=1 git@github.com:10gen/libfaketime.git cd libfaketime branch=$(git symbolic-ref --short HEAD) commit=$(git show -s --pretty=format:"%h - %an, %ar: %s") echo "Git branch: $branch, commit: $commit" make PREFIX=$(pwd)/build/ LIBDIRNAME='.' install - command: shell.exec params: working_dir: src script: | set -o errexit git clone --branch=jepsen-mongodb-master --depth=1 git@github.com:10gen/jepsen.git jepsen-mongodb cd jepsen-mongodb branch=$(git symbolic-ref --short HEAD) commit=$(git show -s --pretty=format:"%h - %an, %ar: %s") echo "Git branch: $branch, commit: $commit" lein install - command: shell.exec params: working_dir: src script: | set -o errexit ${activate_virtualenv} $python -c 'import socket; num_nodes = 5; print("\n".join(["%s:%d" % (socket.gethostname(), port) for port in range(20000, 20000 + num_nodes)]))' > nodes.txt "run jepsen test" : - command: shell.exec type: test timeout_secs: 2700 # Timeout test if there is no output for more than 45 minutes. params: working_dir: src/jepsen-mongodb script: | set -o verbose # Set the TMPDIR environment variable to be a directory in the task's working # directory so that temporary files created by processes spawned by jepsen get # cleaned up after the task completes. This also ensures the spawned processes # aren't impacted by limited space in the mount point for the /tmp directory. # We also need to set the _JAVA_OPTIONS environment variable so that lein will # recognize this as the default temp directory. export TMPDIR="${workdir}/tmp" mkdir -p $TMPDIR export _JAVA_OPTIONS=-Djava.io.tmpdir=$TMPDIR start_time=$(date +%s) lein run test --test ${jepsen_test_name} \ --mongodb-dir ../ \ --working-dir ${workdir}/src/jepsen-workdir \ --clock-skew faketime \ --libfaketime-path ${workdir}/src/libfaketime/build/libfaketime.so.1 \ --virtualization none \ --nodes-file ../nodes.txt \ ${jepsen_key_time_limit} \ ${jepsen_protocol_version} \ ${jepsen_read_concern} \ ${jepsen_read_with_find_and_modify} \ ${jepsen_storage_engine} \ ${jepsen_time_limit} \ ${jepsen_write_concern} \ 2>&1 \ | tee jepsen_${task_name}_${execution}.log end_time=$(date +%s) elapsed_secs=$((end_time-start_time)) # Since we cannot use PIPESTATUS to get the exit code from the "lein run ..." pipe in dash shell, # we will check the output for success, failure or set up error. Note that 'grep' returns with exit code # 0 if it finds a match, and exit code 1 if no match is found. grep -q "Everything looks good" jepsen_${task_name}_${execution}.log grep_exit_code=$? if [ $grep_exit_code -eq 0 ]; then status='"pass"' failures=0 final_exit_code=0 else grep -q "Analysis invalid" jepsen_${task_name}_${execution}.log grep_exit_code=$? if [ $grep_exit_code -eq 0 ]; then status='"fail"' failures=1 final_exit_code=1 else # If the failure is due to setup, then this is considered a system failure. echo $grep_exit_code > jepsen_system_failure_${task_name}_${execution} exit 0 fi fi # Create report.json echo "{\"failures\": $failures, \"results\": [{\"status\": $status, \"exit_code\": $final_exit_code, \"test_file\": \"${task_name}\", \"start\": $start_time, \"end\": $end_time, \"elapsed\": $elapsed_secs}]}" > ../report.json exit $final_exit_code - command: shell.exec params: working_dir: src/jepsen-mongodb script: | set -o verbose # Jepsen system failure if file exists. if [ -f jepsen_system_failure_${task_name}_${execution} ]; then exit $(cat jepsen_system_failure_${task_name}_${execution}) fi "setup jstestfuzz": - command: shell.exec params: working_dir: src shell: bash script: | set -o errexit set -o verbose ${add_nodejs_to_path} git clone git@github.com:10gen/jstestfuzz.git pushd jstestfuzz npm install npm run prepare popd "lint fuzzer sanity patch": - command: shell.exec type: test params: working_dir: src shell: bash script: | set -eo pipefail set -o verbose ${add_nodejs_to_path} # Run parse-jsfiles on 50 files at a time with 32 processes in parallel. # Grep returns 1 if it fails to find a match. (grep "\.js$" modified_and_created_patch_files.txt || true) | xargs -P 32 -L 50 npm run --prefix jstestfuzz parse-jsfiles -- "lint fuzzer sanity all": - command: shell.exec type: test params: working_dir: src shell: bash script: | set -eo pipefail set -o verbose ${add_nodejs_to_path} # Run parse-jsfiles on 50 files at a time with 32 processes in parallel. find "$PWD/jstests" "$PWD/src/mongo/db/modules/enterprise" -name "*.js" -print | xargs -P 32 -L 50 npm run --prefix jstestfuzz parse-jsfiles -- "run jstestfuzz": - command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose cp mongodb*/bin/mongod . git clone --depth 1 git@github.com:10gen/mongo-enterprise-modules.git jstests/enterprise_tests git clone --depth 1 git@github.com:10gen/QA.git jstests/qa_tests - command: shell.exec params: working_dir: src/jstestfuzz script: | set -o errexit set -o verbose ${add_nodejs_to_path} npm run ${npm_command|jstestfuzz} -- ${jstestfuzz_vars} --branch ${branch_name} - command: archive.targz_pack params: target: "jstests.tgz" source_dir: "src/jstestfuzz" include: - "out/*.js" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: jstests.tgz remote_file: ${project}/${build_variant}/${revision}/jstestfuzz/${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Generated Tests - Execution ${execution} "run idl tests" : - command: shell.exec type: test params: working_dir: src script: | set -o verbose set -o errexit ${activate_virtualenv} $python buildscripts/idl/run_tests.py "run powercycle test" : - command: shell.exec type: test params: working_dir: src shell: bash script: | set -o verbose set -o errexit ${activate_virtualenv} ${set_sudo} if [ ! -z $sudo ]; then remote_sudo="--remoteSudo" fi # The virtaulenv bin_dir is different for Linux and Windows bin_dir=$(find $VIRTUAL_ENV -name activate | sed -e "s,$VIRTUAL_ENV,,;s,activate,,;s,/,,g") cmds="source ${virtualenv_dir|venv}/$bin_dir/activate" cmds="$cmds; python -u" # The remote python operates in a virtualenv remote_python="--remotePython=\"$cmds\"" # Initialize report.json. The report will be finalized by powertest.py. report_json="report.json" start_time=$(date +%s) status="\"fail\"" failures=0 exit_code=1 end_time=$start_time elapsed_secs=0 echo "{\"failures\": $failures, \"results\": [{\"status\": $status, \"exit_code\": $exit_code, \"test_file\": \"${task_name}\", \"start\": $start_time, \"end\": $end_time, \"elapsed\": $elapsed_secs}]}" > $report_json generate_report_json="--reportJsonFile=$report_json" # Windows task overrides: # - Execute 10 test loops # - Cap the maximum number of clients to 10 each if [ "Windows_NT" = "$OS" ]; then test_override=--testLoops=10 max_clients=10 for client in --numCrudClients --numFsmClients do override=$(echo ${client_options} | awk "BEGIN {FS=\" |=\"} {for(j=1;j<=NF;j++) if (\$j~/^$client/) {min=(\$(j+1) < $max_clients) ? \$(j+1) : $max_clients; printf \"%s=%d\", \$j,min}}") client_override="$client_override $override" done fi # Set an exit trap so we can save the real exit status (see SERVER-34033). trap 'echo $? > error_exit.txt; exit 0' EXIT config_file=powertest.yml eval $python pytests/powertest.py \ "--saveConfigOptions=$config_file \ ${connection_options} \ ${program_options} \ $generate_report_json \ $remote_sudo \ $remote_python \ ${test_options} \ $test_override \ ${crash_options} \ ${client_options} \ $client_override \ ${mongodb_options} \ ${mongod_options} \ ${mongod_extra_options}" set +o errexit $python -u pytests/powertest.py --configFile=$config_file - command: expansions.update params: ignore_missing_file: true file: src/${exit_file} - command: shell.exec params: working_dir: src shell: bash script: | # Trigger a system failure if powertest.py failed due to ssh access. if [ -n "${ec2_ssh_failure}" ]; then echo "ec2_ssh_failure detected - $(cat ${exit_file})" exit ${exit_code} fi - command: shell.exec params: working_dir: src shell: bash script: | if [ ! -f report.json ]; then exit 0 fi grep -q "pass" report.json pass=$? # On test success, we only archive mongod.log. if [ $pass -eq 0 ]; then echo "ec2_artifacts: ${log_path}" > ec2_artifacts.yml fi - command: expansions.update params: ignore_missing_file: true file: src/ec2_artifacts.yml - command: shell.exec type: test params: shell: bash script: | # Test exits from here with specified exit_code. if [ -n "${exit_code}" ]; then # Python program saved exit_code exit_code=${exit_code} elif [ -f error_exit.txt ]; then # Bash trap exit_code exit_code=$(cat error_exit.txt) else exit_code=0 fi echo "Exiting powercycle with code $exit_code" exit $exit_code "do multiversion setup" : command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose ${activate_virtualenv} # The Windows build variants are running python 2.7.3 and require TLS 1.2 from pyOpenSSL pip install 'pyOpenSSL ; sys_platform == "win32" or sys_platform == "cygwin"' rm -rf /data/install /data/multiversion $python buildscripts/setup_multiversion_mongodb.py \ --installDir /data/install \ --linkDir /data/multiversion \ --edition ${multiversion_edition|base} \ --platform ${multiversion_platform|linux} \ --architecture ${multiversion_architecture|x86_64} \ --useLatest 3.0 3.2 3.4 3.6.14 "do snmp setup" : command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose mkdir -p snmpconf cp -f src/mongo/db/modules/enterprise/docs/mongod.conf.master snmpconf/mongod.conf "do watchdog setup" : command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose bash src/mongo/db/modules/enterprise/jstests/watchdog/charybdefs_setup.sh "cleanup environment" : command: shell.exec params: script: | set -o verbose rm -rf src /data/db/* mongo-diskstats* mongo-*.tgz ~/.aws ~/.boto venv "kill processes" : command: shell.exec params: silent: true script: | process_kill_list="(^cl\.exe$|bsondump|java|lein|lldb|mongo|python|_test$|_test\.exe$)" # Exclude Evergreen agent processes and other system daemons process_exclude_list="(main|tuned|evergreen)" if [ "Windows_NT" = "$OS" ]; then # Get the list of Windows tasks (tasklist list format): # - Transpose the Image Name and PID # - The first column has the process ID # - The second column (and beyond) has task name # - Grep for the task names of interest while ignoring any names that are in the exclude list processes=$(tasklist /fo:csv | awk -F'","' '{x=$1; gsub("\"","",x); print $2, x}' | grep -i -E "$process_kill_list" | grep -i -v -E "$process_exclude_list") # Kill the Windows process by process ID with force (/f) kill_process () { pid=$(echo $1 | cut -f1 -d ' '); echo "Killing process $1"; taskkill /pid "$pid" /f; } else # Get the list of Unix tasks (pgrep full & long): # - Grep for the task names of interest while ignoring any names that are in the exclude list # - The first column has the process ID # - The second column (and beyond) has task name # There are 2 "styles" of grep, figure out which one works grep=grep echo "a" | $grep -qE "(a)" > /dev/null 2>&1 if [ $? -ne 0 ]; then # Extended grep on Solaris is in the following location grep=/usr/xpg4/bin/grep if [ ! -f $grep ]; then echo "Extended $grep does not exist" return 1 fi fi # There are 2 "styles" of pgrep, figure out which one works. # Due to https://bugs.launchpad.net/ubuntu/+source/procps/+bug/1501916 # we cannot rely on the return status ($?) to detect if the option is supported. pgrep -f --list-full ".*" 2>&1 | $grep -qE "(illegal|invalid|unrecognized) option" if [ $? -ne 0 ]; then pgrep_list=$(pgrep -f --list-full "$process_kill_list") else pgrep_list=$(pgrep -f -l "$process_kill_list") fi # Since a process name might have a CR or LF in it, we need to delete any lines from # pgrep which do not start with space(s) and 1 digit and trim any leading spaces. processes=$(echo "$pgrep_list" | $grep -ivE "$process_exclude_list" | sed -e '/^ *[0-9]/!d; s/^ *//; s/[[:cntrl:]]//g;') # Kill the Unix process ID with signal KILL (9) kill_process () { pid=$(echo $1 | cut -f1 -d ' '); echo "Killing process $1"; kill -9 $pid; } fi # Since a full process name can have spaces, the IFS (internal field separator) # should not include a space, just a LF & CR IFS=$(printf "\n\r") for process in $processes do kill_process "$process" done "run kitchen": command: shell.exec type: test params: shell: bash working_dir: src/buildscripts/package_test script: | set -o errexit export KITCHEN_ARTIFACTS_URL="https://s3.amazonaws.com/mciuploads/${project}/${build_variant}/${revision}/artifacts/${build_id}-packages.tgz" export KITCHEN_SECURITY_GROUP="${kitchen_security_group}" export KITCHEN_SSH_KEY_ID="${kitchen_ssh_key_id}" export KITCHEN_SUBNET="${kitchen_subnet}" export KITCHEN_VPC="${kitchen_vpc}" # set expiration tag 2 hours in the future, since no test should take this long ${activate_virtualenv} export KITCHEN_EXPIRE="$($python -c 'import datetime; print((datetime.datetime.utcnow() + datetime.timedelta(hours=2)).strftime("%Y-%m-%d %H:%M:%S"))')" for i in {1..3} do if ! kitchen verify "${packager_distro}"; then verified="false" kitchen destroy "${packager_distro}" || true sleep 30 else verified="true" break fi done kitchen destroy "${packager_distro}" || true test "$verified" = "true" "fetch test_lifecycle.yml": - command: shell.exec type: test params: working_dir: src script: | set -o verbose ${activate_virtualenv} $python buildscripts/fetch_test_lifecycle.py \ --metadataRepo git@github.com:mongodb/mongo-test-metadata.git \ --lifecycleFile etc/test_lifecycle.yml \ --referencesFile references.yml \ --destinationFile etc/test_lifecycle.yml \ --revision ${revision} \ ${project} exit_code=$? if [ ${fail_task_on_error|false} = true ]; then exit $exit_code else exit 0 fi "copy ec2 monitor files": ©_ec2_monitor_files command: shell.exec params: background: true system_log: true working_dir: src silent: false script: | while [ 1 ] do # Tar/zip monitor files on remote host. if [ -z "${ec2_monitor_files}" ] || [ -z "${instance_id}" ]; then exit 0 fi # Ensure we use the latest private_ip_address, as it could change if the EC2 instance # has been stopped and started. ${activate_virtualenv} # Specify '--mode start' to ensure the remote instance is running. monitor_ec2_yml=monitor_ec2.yml $python buildscripts/aws_ec2.py --imageId ${instance_id} --mode start --yamlFile $monitor_ec2_yml echo "AMI EC2 instance ${instance_id} status: $(cat $monitor_ec2_yml)" private_ip_address=$($python buildscripts/yaml_key_value.py --yamlFile $monitor_ec2_yml --yamlKey private_ip_address) if [ -z "$private_ip_address" ]; then echo "Cannot determine the IP address for the remote monitor." continue fi cmd="${tar|tar} czf ec2_monitor_files.tgz ${ec2_monitor_files}" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" ssh_connection_options="$ssh_connection_options -o ConnectionAttempts=3" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@$private_ip_address \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmd" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@$private_ip_address \ --operation "copy_from" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --file ec2_monitor_files.tgz sleep 30 done "set up EC2 instance": &set_up_ec2_instance - command: shell.exec params: working_dir: src script: | set -o errexit ${activate_virtualenv} pip install -r buildscripts/requirements.txt pip install -r pytests/requirements.txt if [ ! -z "${subnet_id}" ]; then subnet_id="-n ${subnet_id}" fi for security_group_id in ${security_group_ids} do security_group_ids="$security_group_ids -g $security_group_id" done if [ -z "${security_group_ids}" ]; then for security_group in ${security_groups} do security_groups="$security_groups -s $security_group" done fi if [ -n "${ec2_expire_hours}" ]; then expire_hours="-e ${ec2_expire_hours}" # Since Windows hosts are expensive to keep running we'll expire it after 3 hours. if [ "Windows_NT" = "$OS" ]; then expire_hours="-e 3" fi fi # Clone another instance of this host in EC2. buildscripts/launch_evergreen_ec2_instance.sh \ $expire_hours \ -k ${ssh_key_id} \ $security_groups \ $security_group_ids \ $subnet_id \ -t "AMI Evergreen ${task_id}" \ -y ${aws_ec2_yml} - command: expansions.update params: file: src/${aws_ec2_yml} - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} # Copy mount_drives.sh script to remote host. ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --operation "copy_to" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --file buildscripts/mount_drives.sh - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} # Mount /data on the attached drive(s), more than 1 indicates a RAID set. ${set_sudo} script_opts="-d '${data_device_names}'" if [ ! -z "${raid_data_device_name}" ]; then script_opts="$script_opts -r ${raid_data_device_name}" fi if [ ! -z "${fstype}" ]; then script_opts="$script_opts -t ${fstype}" fi if [ ! -z "${fs_options}" ]; then script_opts="$script_opts -o '${fs_options}'" fi # Mount /log on the attached drive. if [ ! -z "${log_device_name}" ]; then script_opts="$script_opts -l '${log_device_name}'" log="/log" fi group=$(id -Gn $USER | cut -f1 -d ' ') || true user_group="$USER:$group" script_opts="$script_opts -u $user_group" data_db=/data/db cmds="$sudo bash mount_drives.sh $script_opts; mount; ls -ld $data_db $log; df" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} # Create remote_dir, if specified as expansion macro and is not '.' (pwd). if [[ -z "${remote_dir|}" || ${remote_dir} == "." ]]; then exit 0 fi ${set_sudo} ssh_connection_options="${ssh_identity} ${ssh_connection_options}" group=$(id -Gn $USER | cut -f1 -d ' ') || true user_group="$USER:$group" set_permission="chmod 777 ${remote_dir}" if [ "Windows_NT" = "$OS" ]; then set_permission="setfacl -s user::rwx,group::rwx,other::rwx ${remote_dir}" fi cmds="$sudo mkdir -p ${remote_dir}; $sudo chown $user_group ${remote_dir}; $set_permission; ls -ld ${remote_dir}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} # Copy buildscripts, pytests and mongoDB executables to the remote host. file_param="--file buildscripts --file pytests" mongo_executables="mongo mongod mongos" for executable in $mongo_executables do file_param="$file_param --file $executable${exe}" done ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --operation "copy_to" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ $file_param \ --remoteDir ${remote_dir} - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} # Set up virtualenv on remote. cmds="python_loc=\$(which \${python|/opt/mongodbtoolchain/v2/bin/python2})" cmds="$cmds; remote_dir=${remote_dir|.}" cmds="$cmds; if [ \"Windows_NT\" = \"$OS\" ]; then python_loc=\$(cygpath -w \$python_loc); remote_dir=\$(cygpath -w \$remote_dir); fi" cmds="$cmds; virtualenv --python \$python_loc --system-site-packages ${virtualenv_dir|venv}" cmds="$cmds; activate=\$(find ${virtualenv_dir|venv} -name 'activate')" cmds="$cmds; . \$activate" cmds="$cmds; pip install -r \$remote_dir/pytests/requirements.txt" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" - command: shell.exec params: shell: bash working_dir: src script: | if [ "Windows_NT" = "$OS" ]; then exit 0 fi # Enable core dumps on non-Windows remote hosts. # The core pattern must specify a director, since mongod --fork will chdir("/") # and cannot generate a core dump there (see SERVER-21635). # We need to reboot the host for the core limits to take effect. ${set_sudo} core_pattern=${remote_dir}/dump_%e.%p.core sysctl_conf=/etc/sysctl.conf cmds="ulimit -a" cmds="$cmds; echo \"$USER - core unlimited\" | $sudo tee -a /etc/security/limits.conf" cmds="$cmds; if [ -f $sysctl_conf ]" cmds="$cmds; then grep ^kernel.core_pattern $sysctl_conf" cmds="$cmds; if [ \$? -eq 0 ]" cmds="$cmds; then $sudo sed -i \"s,kernel.core_pattern=.*,kernel.core_pattern=$core_pattern,\" $sysctl_conf" cmds="$cmds; else echo \"kernel.core_pattern=$core_pattern\" | $sudo tee -a $sysctl_conf" cmds="$cmds; fi" cmds="$cmds; else echo Cannot change the core pattern and no core dumps will be generated." cmds="$cmds; fi" # The following line for restarting the machine is based on # https://unix.stackexchange.com/a/349558 in order to ensure the ssh client gets a # response from the remote machine before it restarts. cmds="$cmds; nohup $sudo reboot &>/dev/null & exit" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" ${activate_virtualenv} $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" - command: shell.exec params: shell: bash working_dir: src script: | if [ "Windows_NT" = "$OS" ]; then exit 0 fi # Always exit successfully, as this is just informational. trap 'echo "Trapped exit code $?, exiting with 0"; exit 0' EXIT # Print the ulimit & kernel.core_pattern cmds="uptime" cmds="$cmds; ulimit -a" cmds="$cmds; if [ -f /sbin/sysctl ]" cmds="$cmds; then /sbin/sysctl kernel.core_pattern" cmds="$cmds; fi" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" ${activate_virtualenv} $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|3} \ --commands "$cmds" - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} ${set_sudo} # Set up curator to collect system & process stats on remote. if [ "Windows_NT" = "$OS" ]; then variant=windows-64 else variant=ubuntu1604 fi # Download stable version of curator curator_hash=117d1a65256ff78b6d15ab79a1c7088443b936d0 curator_url="https://s3.amazonaws.com/boxes.10gen.com/build/curator/curator-dist-$variant-$curator_hash.tar.gz" cmds="curl -s $curator_url | tar -xzv" if [ "Windows_NT" = "$OS" ]; then # Since curator runs as SYSTEM user, ensure the output files can be accessed. cmds="$cmds; touch ${monitor_system_file}; chmod 777 ${monitor_system_file}" cmds="$cmds; cygrunsrv --install curator_sys --path curator --chdir \$HOME --args 'stat system --file ${monitor_system_file}'" cmds="$cmds; touch ${monitor_proc_file}; chmod 777 ${monitor_proc_file}" cmds="$cmds; cygrunsrv --install curator_proc --path curator --chdir \$HOME --args 'stat process-all --file ${monitor_proc_file}'" cmds="$cmds; cygrunsrv --start curator_sys" cmds="$cmds; cygrunsrv --start curator_proc" else cmds="$cmds; cmd=\"@reboot cd \$HOME && ${sudo} ./curator stat system >> ${monitor_system_file}\"" cmds="$cmds; (crontab -l ; echo \"\$cmd\") | crontab -" cmds="$cmds; cmd=\"@reboot cd \$HOME && ${sudo} ./curator stat process-all >> ${monitor_proc_file}\"" cmds="$cmds; (crontab -l ; echo \"\$cmd\") | crontab -" cmds="$cmds; crontab -l" cmds="$cmds; { ${sudo} \$HOME/curator stat system --file ${monitor_system_file} > /dev/null 2>&1 & ${sudo} \$HOME/curator stat process-all --file ${monitor_proc_file} > /dev/null 2>&1 & } & disown" fi ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" - command: shell.exec params: shell: bash working_dir: src script: | set -o errexit ${activate_virtualenv} ${set_sudo} # Many systems have the firewall disabled, by default. In case the firewall is # enabled we add rules for the mongod ports on the remote. # RHEL 7 firewall rules if [ ! -z "$(which firewall-cmd 2> /dev/null)" ]; then cmds="$sudo firewall-cmd --permanent --zone=public --add-port=ssh/tcp" cmds="$cmds; $sudo firewall-cmd --permanent --zone=public --add-port=${standard_port}/tcp" cmds="$cmds; $sudo firewall-cmd --permanent --zone=public --add-port=${secret_port}/tcp" cmds="$cmds; $sudo firewall-cmd --reload" cmds="$cmds; $sudo firewall-cmd --list-all" # ArchLinux, Debian, RHEL 6 firewall rules elif [ ! -z "$($sudo iptables --list 2> /dev/null)" ]; then cmds="$sudo iptables -I INPUT 1 -p tcp --dport ssh -j ACCEPT" cmds="$cmds; $sudo iptables -I INPUT 1 -p tcp --dport ${standard_port} -j ACCEPT" cmds="$cmds; $sudo iptables -I INPUT 1 -p tcp --dport ${secret_port} -j ACCEPT" if [ -d /etc/iptables ]; then rules_file=/etc/iptables/iptables.rules elif [ -f /etc/sysconfig/iptables ]; then rules_file=/etc/sysconfig/iptables else rules_file=/etc/iptables.up.rules fi cmds="$cmds; $sudo iptables-save | $sudo tee $rules_file" cmds="$cmds; $sudo iptables --list-rules" elif [ ! -z "$($sudo service iptables status 2> /dev/null)" ]; then cmds="$sudo iptables -I INPUT 1 -p tcp --dport ssh -j ACCEPT" cmds="$cmds; $sudo iptables -I INPUT 1 -p tcp --dport ${standard_port} -j ACCEPT" cmds="$cmds; $sudo iptables -I INPUT 1 -p tcp --dport ${secret_port} -j ACCEPT" cmds="$cmds; $sudo service iptables save" cmds="$cmds; $sudo service iptables status" # Ubuntu firewall rules elif [ ! -z "$($sudo ufw status 2> /dev/null)" ]; then cmds="$sudo ufw allow ssh/tcp" cmds="$cmds; $sudo ufw allow ${standard_port}/tcp" cmds="$cmds; $sudo ufw allow ${secret_port}/tcp" cmds="$cmds; $sudo ufw reload" cmds="$cmds; $sudo ufw status" # SuSE firewall rules # TODO: Add firewall rules using SuSEfirewall2 elif [ ! -z "$($sudo /sbin/SuSEfirewall2 help 2> /dev/null)" ]; then cmds="$sudo /sbin/SuSEfirewall2 stop" cmds="$cmds; $sudo /sbin/SuSEfirewall2 off" # Windows firewall rules elif [ ! -z "$(netsh advfirewall show store 2> /dev/null)" ]; then add_rule="netsh advfirewall firewall add rule" cmds="$add_rule name='MongoDB port ${standard_port} in' dir=in action=allow protocol=TCP localport=${standard_port}" cmds="$cmds; $add_rule name='MongoDB port ${standard_port} out' dir=in action=allow protocol=TCP localport=${standard_port}" cmds="$cmds; $add_rule name='MongoDB port ${secret_port} in' dir=in action=allow protocol=TCP localport=${secret_port}" cmds="$cmds; $add_rule name='MongoDB port ${secret_port} out' dir=in action=allow protocol=TCP localport=${secret_port}" cmds="$cmds; netsh advfirewall firewall show rule name=all | grep -A 13 'MongoDB'" else echo "Firewall not active or unknown firewall command on this platform" exit 0 fi set -o errexit if [ ! -z "$cmds" ]; then ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" fi - command: shell.exec params: shell: bash working_dir: src script: | ${activate_virtualenv} set -o errexit if [[ "Windows_NT" != "$OS" || -z "${windows_crash_zip}" ]]; then exit 0 fi # Install NotMyFault, used to crash Windows. cmds="curl -s -o ${windows_crash_zip} ${windows_crash_dl}" cmds="$cmds; unzip -q ${windows_crash_zip} -d ${windows_crash_dir}" cmds="$cmds; chmod +x ${windows_crash_dir}/*.exe" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmds" - *copy_ec2_monitor_files ### Determine & set remote EC2 IP address ### "get EC2 address": &get_ec2_address command: shell.exec params: shell: bash working_dir: src script: | if [ -z "${instance_id}" ]; then exit 0 fi # Ensure we use the latest private_ip_address, as it could change if the EC2 instance # has been stopped and started. ${activate_virtualenv} # Specify '--mode start' to ensure the remote instance is running. now=$(date +'%Y%m%d%H%M%S') aws_ec2_status_yml=aws_ec2_status.yml $python buildscripts/aws_ec2.py \ --imageId ${instance_id} \ --mode start \ --yamlFile $aws_ec2_status_yml \ --consoleOutputFile ec2_console_$now.log \ --consoleScreenshotFile ec2_console_screen_shot_$now.jpg private_ip_address=$($python buildscripts/yaml_key_value.py --yamlFile $aws_ec2_status_yml --yamlKey private_ip_address) echo "private_ip_address: $private_ip_address" > private_ip_address.yml "update EC2 address": &update_ec2_address command: expansions.update params: file: src/private_ip_address.yml ### Process & archive remote EC2 artifacts ### "tar EC2 artifacts": &tar_ec2_artifacts command: shell.exec params: shell: bash working_dir: src script: | ${activate_virtualenv} # Tar/zip artifacts on remote host. if [[ -z "${ec2_artifacts}" || -n "${ec2_ssh_failure}" ]]; then exit 0 fi cmd="${tar|tar} czf ec2_artifacts.tgz ${ec2_artifacts}" ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --commands "$cmd" "copy EC2 artifacts": ©_ec2_artifacts command: shell.exec params: shell: bash working_dir: src script: | ${activate_virtualenv} # Copy remote artifacts. if [[ -z "${ec2_artifacts}" || -n "${ec2_ssh_failure}" ]]; then exit 0 fi ssh_connection_options="${ssh_identity} ${ssh_connection_options}" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --operation "copy_from" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries|0} \ --file ec2_artifacts.tgz "cleanup EC2 instance": &cleanup_ec2_instance command: shell.exec params: shell: bash working_dir: src script: | # We do not terminate the EC2 instance if there was an ec2_ssh_failure. if [[ -z "${ec2_artifacts}" || -n "${ec2_ssh_failure}" ]]; then exit 0 fi ${activate_virtualenv} echo "Terminating $instance_id" aws_ec2=$($python buildscripts/aws_ec2.py --imageId ${instance_id} --mode terminate) echo "Terminated AMI EC2 instance: $aws_ec2" "gather remote mongo coredumps": &gather_remote_mongo_coredumps command: shell.exec params: shell: bash working_dir: "src" script: | if [[ ! -f ${aws_ec2_yml|""} || -n "${ec2_ssh_failure}" ]]; then exit 0 fi ssh_connection_options="${ssh_identity} ${ssh_connection_options}" remote_dir=${remote_dir|.} # Find all core files and move to $remote_dir cmds="core_files=\$(/usr/bin/find -H . \( -name '*.core' -o -name '*.mdmp' \) 2> /dev/null)" cmds="$cmds; if [ -z \"\$core_files\" ]; then exit 0; fi" cmds="$cmds; echo Found remote core files \$core_files, moving to \$(pwd)" cmds="$cmds; for core_file in \$core_files" cmds="$cmds; do base_name=\$(echo \$core_file | sed 's/.*\///')" cmds="$cmds; if [ ! -f \$base_name ]; then mv \$core_file .; fi" cmds="$cmds; done" ${activate_virtualenv} $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries} \ --commands "$cmds" \ --commandDir $remote_dir "copy remote mongo coredumps": ©_remote_mongo_coredumps command: shell.exec params: shell: bash working_dir: "src" script: | if [[ ! -f ${aws_ec2_yml|""} || -n "${ec2_ssh_failure}" ]]; then exit 0 fi ssh_connection_options="${ssh_identity} ${ssh_connection_options}" remote_dir=${remote_dir|.} ${activate_virtualenv} $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --operation "copy_from" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries} \ --file "$remote_dir/*.core" \ --file "$remote_dir/*.mdmp" # Since both type of core files do not exist on the same host, this command # will always return non-zero. As the core file retrieval is optional, we # always exit successfully. exit 0 "archive remote EC2 artifacts": &archive_remote_ec2_artifacts command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/ec2_artifacts.tgz remote_file: ${project}/${build_variant}/${revision}/remote_ec2/remote_ec2_artifacts-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Remote EC2 Artifacts - Execution ${execution} optional: true "archive remote EC2 monitor files": &archive_remote_ec2_monitor_files command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/ec2_monitor_files.tgz remote_file: ${project}/${build_variant}/${revision}/remote_ec2/remote_ec2_monitor-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Remote EC2 Monitor - Execution ${execution} optional: true "gather EC2 console artifacts": &gather_ec2_console_artifacts command: shell.exec params: working_dir: src script: | ec2_console_files=$(ls ec2_console* 2> /dev/null) if [ -n "$ec2_console_files" ]; then ${tar|tar} czf ec2_console_files.tgz $ec2_console_files fi "archive EC2 console artifacts": &archive_ec2_console_artifacts command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/ec2_console_files.tgz remote_file: ${project}/${build_variant}/${revision}/ec2/ec2_console-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: EC2 Console files - Execution ${execution} optional: true "save ec2 task artifacts": - *get_ec2_address - *update_ec2_address - *tar_ec2_artifacts - *copy_ec2_artifacts - *gather_remote_mongo_coredumps - *copy_remote_mongo_coredumps - *gather_ec2_console_artifacts - *archive_ec2_console_artifacts - *cleanup_ec2_instance - *archive_remote_ec2_artifacts - *archive_remote_ec2_monitor_files ### Process & archive local client logs ### "tar local client logs": &tar_local_client_logs command: shell.exec params: working_dir: src script: | client_logs=$(ls crud*.log fsm*.log 2> /dev/null) if [ ! -z "$client_logs" ]; then ${tar|tar} czf client-logs.tgz $client_logs fi "archive local client logs": &archive_local_client_logs command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/client-logs.tgz remote_file: ${project}/${build_variant}/${revision}/client_logs/mongo-client-logs-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Client logs - Execution ${execution} optional: true "save local client logs": - *tar_local_client_logs - *archive_local_client_logs pre: - func: "kill processes" - func: "cleanup environment" - func: "set up virtualenv" - command: expansions.update params: updates: - key: activate_virtualenv value: | # check if virtualenv is set up if [ -d "${workdir}/venv" ]; then if [ "Windows_NT" = "$OS" ]; then # Need to quote the path on Windows to preserve the separator. . "${workdir}/venv/Scripts/activate" 2> /tmp/activate_error.log else . ${workdir}/venv/bin/activate 2> /tmp/activate_error.log fi if [ $? -ne 0 ]; then echo "Failed to activate virtualenv: $(cat /tmp/activate_error.log)" fi python=python else python=${python|/opt/mongodbtoolchain/v2/bin/python2} fi echo "python set to $(which $python)" - key: activate_virtualenv_3 value: | # check if virtualenv for python3 is set up if [ -d "${workdir}/venv_3" ]; then if [ "Windows_NT" = "$OS" ]; then # Need to quote the path on Windows to preserve the separator. . "${workdir}/venv_3/Scripts/activate" 2> /tmp/activate_error.log else . ${workdir}/venv_3/bin/activate 2> /tmp/activate_error.log fi if [ $? -ne 0 ]; then echo "Failed to activate virtualenv: $(cat /tmp/activate_error.log)" fi python=python else if [ "Windows_NT" = "$OS" ]; then python=/cygdrive/c/python/Python36/python else python=${python3|/opt/mongodbtoolchain/v2/bin/python3} fi fi echo "python set to $(which $python)" - key: add_nodejs_to_path value: | # Add node and npm binaries to PATH if [ "Windows_NT" = "$OS" ]; then # An "npm" directory might not have been created in %APPDATA% by the Windows installer. # Work around the issue by specifying a different %APPDATA% path. # See: https://github.com/nodejs/node-v0.x-archive/issues/8141 export APPDATA=${workdir}/npm-app-data export PATH="$PATH:/cygdrive/c/Program Files (x86)/nodejs" # Windows location # TODO: this is to work around BUILD-8652 cd "$(pwd -P | sed 's,cygdrive/c/,cygdrive/z/,')" else export PATH="$PATH:/opt/node/bin" fi - key: posix_workdir value: eval 'if [ "Windows_NT" = "$OS" ]; then echo $(cygpath -u "${workdir}"); else echo ${workdir}; fi' # For ssh disable the options GSSAPIAuthentication, CheckHostIP, StrictHostKeyChecking # & UserKnownHostsFile, since these are local connections from one AWS instance to another. - key: ssh_connection_options value: -o GSSAPIAuthentication=no -o CheckHostIP=no -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o ConnectTimeout=30 -o ConnectionAttempts=20 - key: ssh_retries value: "10" - key: set_sudo value: | set -o > /tmp/settings.log set +o errexit grep errexit /tmp/settings.log | grep on errexit_on=$? # Set errexit "off". set +o errexit sudo= # Use sudo, if it is supported. sudo date > /dev/null 2>&1 if [ $? -eq 0 ]; then sudo=sudo fi # Set errexit "on", if previously enabled. if [ $errexit_on -eq 0 ]; then set -o errexit fi - command: shell.exec params: system_log: true script: | ulimit -a # Clear the dmesg ring buffer. The "post" phase will check dmesg for OOM messages. ${set_sudo} $sudo dmesg -c > /dev/null 2>&1 if [ $? -eq 0 ]; then echo "Cleared the dmesg ring buffer" else echo "Could not clear the dmesg ring buffer" fi post: - command: attach.results params: file_location: src/report.json - command: attach.artifacts params: ignore_artifacts_for_spawn: false files: - src/archive.json - func: "save ec2 task artifacts" - func: "call BF Suggestion service" - func: "kill processes" - func: "save local client logs" # Print out any Out of Memory killed process messages. - command: shell.exec params: system_log: true working_dir: src # Temporary files created in src will be cleaned up in "pre". script: | ${set_sudo} # Use dmesg -T option, if supported, to display timestamps. dmesg=dmesg $sudo dmesg -T > /dev/null 2>&1 if [ $? -eq 0 ]; then dmesg="dmesg -T" fi $sudo $dmesg 2> /dev/null > dmesg.txt if [ $? -ne 0 ]; then echo "Cannot check for OOM (Out of memory) killed processes on this platform" exit 0 fi egrep -i '(Out of memory|OOM[- ]killer|Killed process)' dmesg.txt > oom.txt if [ -s oom.txt ]; then echo "OOM (Out of memory) killed processes detected" cat oom.txt else echo "No OOM (Out of memory) killed processes detected" fi # Cleanup after the watchdog FUSE testing - command: shell.exec params: working_dir: src script: | if [ -d /data/thrift ]; then rm -rf /data/thrift fi if [ -d /data/charybdefs ]; then rm -rf /data/charybdefs fi # Attach scons config log - command: s3.put params: optional: true aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/build/scons/config.log remote_file: ${project}/${build_variant}/${revision}/artifacts/config-${build_id}.log bucket: mciuploads permissions: public-read content_type: text/plain display_name: config.log # Process and save coverage data. - command: shell.exec params: working_dir: src script: | set +o errexit if [ -d "./build" ]; then file_list=$(find ./build -type f -name "*.gcda") if [ -n "$file_list" ]; then for gcda_file in $file_list; do echo "Processing file $gcda_file" /opt/mongodbtoolchain/v2/bin/gcov -i $gcda_file base_name=$(echo $gcda_file | rev | cut -f1 -d '/' | cut -f2 -d '.' | rev) gcov_file=$base_name.gcda.gcov if [ -f "$gcov_file" ]; then # Add a prefix to the intermediate file, since it does not have a unique name. # Convert the '/' to '#' in the file path. file_prefix=$(echo $gcda_file | sed -e 's,^\./,,' | rev | cut -f2- -d '/' | rev | tr -s '/' '#') new_gcov_file=$file_prefix#$base_name.gcda.gcov if [ ! -f $new_gcov_file ]; then echo "Renaming gcov intermediate file $gcov_file to $new_gcov_file" mv $gcov_file $new_gcov_file else # We treat this as a fatal condition and remove all of the coverage files. echo "Not renaming $gcov_file as $new_gcov_file already exists!" rm -f *.gcda.gcov exit 1 fi fi done fi fi - command: archive.targz_pack params: target: "src/gcov-intermediate-files.tgz" source_dir: "src" include: - "*.gcda.gcov" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: "src/gcov-intermediate-files.tgz" remote_file: ${project}/${build_variant}/${revision}/gcov/gcov-intermediate-files-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: gcov intermediate files - Execution ${execution} optional: true - command: archive.targz_pack params: target: "src/jepsen-mongod-logs.tgz" source_dir: "${workdir}/src/jepsen-workdir" include: - "./**.log" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/jepsen-mongod-logs.tgz remote_file: ${project}/${build_variant}/${revision}/jepsen/jepsen-mongod-logs-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Jepsen mongod Logs - ${execution} optional: true # Gather and archive the Jepsen results. - command: archive.targz_pack params: target: "src/jepsen-results.tgz" source_dir: "src/jepsen-mongodb/store" include: - "./**" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/jepsen-results.tgz remote_file: ${project}/${build_variant}/${revision}/jepsen/jepsen-results-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Jepsen Test Results - ${execution} optional: true # Gather and archive mongo coredumps. - command: shell.exec params: working_dir: "src" script: | # Find all core files and move to src core_files=$(/usr/bin/find -H .. \( -name "*.core" -o -name "*.mdmp" \) 2> /dev/null) for core_file in $core_files do base_name=$(echo $core_file | sed "s/.*\///") # Move file if it does not already exist if [ ! -f $base_name ]; then mv $core_file . fi done - command: archive.targz_pack params: target: "mongo-coredumps.tgz" source_dir: "src" include: - "./**.core" - "./**.mdmp" # Windows: minidumps - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: mongo-coredumps.tgz remote_file: ${project}/${build_variant}/${revision}/coredumps/mongo-coredumps-${build_id}-${task_name}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Core Dumps - Execution ${execution} optional: true # Gather and archive failed unit test binaries. - command: shell.exec params: working_dir: "src" script: | mkdir unittest_binaries # Find all core files core_files=$(/usr/bin/find -H . \( -name "dump_*.core" -o -name "*.mdmp" \) 2> /dev/null) for core_file in $core_files do # A core file name does not always have the executable name that generated it. # See http://stackoverflow.com/questions/34801353/core-dump-filename-gets-thread-name-instead-of-executable-name-with-core-pattern # On platforms with GDB, we get the binary name from core file gdb=/opt/mongodbtoolchain/gdb/bin/gdb if [ -f $gdb ]; then binary_file=$($gdb -batch --quiet -ex "core $core_file" 2> /dev/null | grep "Core was generated" | cut -f2 -d "\`" | cut -f1 -d "'" | cut -f1 -d " ") binary_file_locations=$binary_file else # Find the base file name from the core file name, note it may be truncated. # Remove leading 'dump_' and trailing '..core' or '..mdmp' binary_file=$(echo $core_file | sed "s/.*\///;s/dump_//;s/\..*\.core//;s/\..*\.mdmp//") # Locate the binary file. Since the base file name might be truncated, the find # may return more than 1 file. binary_file_locations=$(/usr/bin/find -H . -name "$binary_file*${exe}" 2> /dev/null) fi if [ -z "$binary_file_locations" ]; then echo "Cannot locate the unittest binary file ($binary_file) that generated the core file $core_file" fi for binary_file_location in $binary_file_locations do new_binary_file=unittest_binaries/$(echo $binary_file_location | sed "s/.*\///") if [ ! -f $new_binary_file ]; then mv $binary_file_location $new_binary_file fi # On Windows if a .pdb symbol file exists, include it in the archive. pdb_file=$(echo $binary_file_location | sed "s/\.exe/.pdb/") if [ -f $pdb_file ]; then new_pdb_file=unittest_binaries/$(echo $pdb_file | sed "s/.*\///") mv $pdb_file $new_pdb_file fi done done - command: archive.targz_pack params: target: "mongo-unittests.tgz" source_dir: "src/unittest_binaries" include: - "./*_test${exe}" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: mongo-unittests.tgz remote_file: ${project}/${build_variant}/${revision}/unittests/mongo-unittests-${build_id}-${task_name}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Unit tests - Execution ${execution} optional: true # Gather and archive hang_analyzer debugger files. - command: archive.targz_pack params: target: "src/mongo-hanganalyzer.tgz" source_dir: "src" include: - "./debugger*.*" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/mongo-hanganalyzer.tgz remote_file: ${project}/${build_variant}/${revision}/hanganalyzer/mongo-hanganalyzer-${build_id}-${task_name}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Hang Analyzer Output - Execution ${execution} optional: true # Gather and archive disk statistics. - command: archive.targz_pack params: target: "diskstats.tgz" source_dir: "./" include: - "./mongo-diskstats*" - "./mongo-diskstats*.csv" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: diskstats.tgz remote_file: ${project}/${build_variant}/${revision}/diskstats/mongo-diskstats-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Disk Stats - Execution ${execution} optional: true # Gather and archive system resource information. - command: archive.targz_pack params: target: "system-resource-info.tgz" source_dir: src include: - "./system_resource_info*" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: system-resource-info.tgz remote_file: ${project}/${build_variant}/${revision}/systemresourceinfo/mongo-system-resource-info-${task_id}-${execution}.tgz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: System Resource Info - Execution ${execution} optional: true # Cleanup steps. - command: shell.exec params: working_dir: src script: | # removes files from the (local) scons cache when it's over a # threshold, to the $prune_ratio percentage. Ideally override # these default values in the distro config in evergreen. ${activate_virtualenv} if [ -d "${scons_cache_path}" ]; then $python buildscripts/scons_cache_prune.py --cache-dir '${scons_cache_path}' --cache-size ${scons_cache_size|200} --prune-ratio ${scons_prune_ratio|0.8} fi - func: "cleanup environment" timeout: - func: "fetch debugsymbols archive" - func: "extract debugsymbols" - func: "get EC2 address" - func: "update EC2 address" - command: shell.exec params: working_dir: src script: | set -o verbose ${activate_virtualenv} hang_analyzer_option="-o file -o stdout -p ${hang_analyzer_processes|dbtest,java,mongo,mongod,mongos,python,_test} -g bsondump,mongodump,mongoexport,mongofiles,mongoimport,mongoreplay,mongorestore,mongostat,mongotop" if [ ${hang_analyzer_dump_core|true} = true ]; then hang_analyzer_option="-c $hang_analyzer_option" fi echo "Calling the hang analyzer: PATH=\"/opt/mongodbtoolchain/gdb/bin:$PATH\" $python buildscripts/hang_analyzer.py $hang_analyzer_option" PATH="/opt/mongodbtoolchain/gdb/bin:$PATH" $python buildscripts/hang_analyzer.py $hang_analyzer_option # Call hang_analyzer.py script for tasks that are running remote mongo processes # Any task which uses remote processes will have previously loaded this file to set these # expansion macros. if [ -n "${private_ip_address}" ]; then core_ext=core if [ "Windows_NT" = "$OS" ]; then core_ext=mdmp fi ssh_connection_options="${ssh_identity} ${ssh_connection_options}" # buildscripts must be installed in ${remote_dir} on the remote host. remote_dir=${remote_dir|.} # Copy mongoDB debug symbols to the remote host. debug_files=$(ls *.debug *.dSYM *.pdb 2> /dev/null) for debug_file in $debug_files do file_param="$file_param --file $debug_file" done if [ ! -z "$file_param" ]; then $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --operation "copy_to" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries} \ $file_param \ --remoteDir $remote_dir fi cmds="PATH=\"/opt/mongodbtoolchain/gdb/bin:\$PATH\" $python buildscripts/hang_analyzer.py $hang_analyzer_option" $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries} \ --commands "$cmds" \ --commandDir $remote_dir $python buildscripts/remote_operations.py \ --verbose \ --userHost $USER@${private_ip_address} \ --operation "copy_from" \ --sshConnectionOptions "$ssh_connection_options" \ --retries ${ssh_retries} \ --file "$remote_dir/debugger*.*" \ --file "$remote_dir/*.$core_ext" fi ####################################### # Tasks # ####################################### ## The test_lifecycle_excluded_tasks are a list of tasks names which include: ## - Non jstests, i.e., compile, unittests ## - Non standard jstests, i.e., jstestfuzz ## Note that the task name supports glob patterns. test_lifecycle_excluded_tasks: - burn_in_tests - compile* - dbtest* - idl_tests - integration* - jepsen* - jstestfuzz* - lint - mongos* - package - push - rollback_fuzzer* - unittests* - update_fuzzer* tasks: ## compile - build all scons targets except unittests ## - name: compile depends_on: [] commands: - command: manifest.load - func: "git get project" - func: "get buildnumber" - func: "set up credentials" - func: "fetch and build OpenSSL" - func: "build new tools" # noop if ${newtools} is not "true" - func: "generate compile expansions" # Then we load the generated version data into the agent so we can use it in task definitions - func: "apply compile expansions" - func: "scons compile" vars: targets: core tools dbtest integration_tests dist dist-debugsymbols distsrc-${ext|tgz} ${msi_target|} task_compile_flags: >- --use-new-tools --build-mongoreplay="${build_mongoreplay}" --detect-odr-violations - command: shell.exec type: test params: working_dir: src script: | set -o errexit set -o verbose mv mongodb-src-*.${ext|tgz} distsrc.${ext|tgz} mv mongodb-*-debugsymbols.${ext|tgz} mongo-debugsymbols.tgz || true mv mongodb-*.${ext|tgz} mongodb-binaries.tgz # If strip is on the path (everywhere except windows) then we should strip the test binaries # before tarring them up if [ -x ${strip_path|/usr/bin/strip} ]; then cat build/integration_tests.txt | xargs ${strip_command|/usr/bin/strip} ${strip_command|/usr/bin/strip} dbtest ${strip_command|/usr/bin/strip} mongobridge fi - command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose ${activate_virtualenv} if [ "${has_packages|}" = "true" ] ; then cd buildscripts $python ${packager_script} --prefix `pwd`/.. --distros ${packager_distro} --tarball `pwd`/../mongodb-binaries.tgz -s ${version} -m HEAD -a ${packager_arch} cd .. fi # Create separate shell archive mkdir -p shell-archive/build cd shell-archive ${platform_decompress|tar xzvf} ../mongodb-binaries.tgz find . -mindepth 3 ! -name "mongo${exe}" -type f -exec rm {} \; # delete bin/* except bin/mongo $python ../buildscripts/make_archive.py -o mongodb-shell.${ext|tgz} $(find mongodb-* -type f) cd .. # Test lifecycle is temporarily disabled to reduce load on Evergreen API. # - func: "fetch test_lifecycle.yml" # vars: # # Do not fail the task if we fail to fetch the test_lifecycle.yml file # fail_task_on_error: false - command: archive.targz_pack params: target: "artifacts.tgz" source_dir: "src" include: - "src/mongo/db/modules/enterprise/jstests/**" - "compile_expansions.yml" - "src/mongo/db/modules/subscription/jstests/**" - "src/mongo/db/modules/enterprise/docs/**" - "*.exe" - "jstests/**" - "pytests/**" - "./test*" - "./dbtest*" - "./mongobridge*" - "buildscripts/**" - "*Example" - "*Test" - "./**.pdb" - "./**.msi" - "./etc/*san.suppressions" - "./etc/repo_config.yaml" - "./etc/test_lifecycle.yml" - "./etc/test_retrial.yml" - "./build/integration_tests/**" - "./build/integration_tests.txt" - "./build/**.gcno" - "repo/**" - "src/mongo/util/options_parser/test_config_files/**" - "library_dependency_graph.json" - "src/third_party/JSON-Schema-Test-Suite/tests/draft4/**" exclude_files: - "*_test.pdb" - command: archive.targz_pack params: target: "packages.tgz" source_dir: "src" include: - "repo/**" - "./**.msi" - command: s3.put params: optional: true aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: packages.tgz remote_file: ${project}/${build_variant}/${revision}/artifacts/${build_id}-packages.tgz bucket: mciuploads permissions: public-read content_type: application/tar display_name: Packages - func: "upload debugsymbols" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/mongodb-binaries.tgz remote_file: ${project}/${build_variant}/${revision}/binaries/mongo-${build_id}.${ext|tgz} bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Binaries - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/shell-archive/mongodb-shell.${ext|tgz} remote_file: ${project}/${build_variant}/${revision}/binaries/mongo-shell-${build_id}.${ext|tgz} bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Shell - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: artifacts.tgz remote_file: ${project}/${build_variant}/${revision}/artifacts/${build_id}.tgz bucket: mciuploads permissions: public-read content_type: application/tar display_name: Artifacts - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/distsrc.${ext|tgz} remote_file: ${project}/${build_variant}/${revision}/sources/mongo-src-${build_id}.${ext|tgz} bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Source tarball # We only need to upload the source tarball from one of the build variants # because it should be the same everywhere, so just use linux-64/windows-64-2k8. build_variants: [ linux-64, windows-64-2k8-ssl ] ## compile_all - build all scons targets including unittests ## - name: compile_all commands: - command: manifest.load - func: "git get project" - func: "get buildnumber" - func: "set up credentials" - func: "fetch and build OpenSSL" - func: "build new tools" # noop if ${newtools} is not "true" - func: "generate compile expansions" # Then we load the generated version data into the agent so we can use it in task definitions. - func: "apply compile expansions" - func: "scons compile" vars: targets: all task_compile_flags: >- --use-new-tools --build-mongoreplay="${build_mongoreplay}" --detect-odr-violations - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/library_dependency_graph.json remote_file: ${project}/${build_variant}/${revision}/library_dependency_graph.${build_id}.json bucket: mciuploads permissions: public-read content_type: application/json display_name: Library Dependency Graph (library_dependency_graph.json) build_variants: [enterprise-ubuntu1604-64] # Run the C++ unittests as part of compile_all. The compiled binaries are automatically # installed into the top-level directory by SCons. # noop if ${disable_unit_tests} is "true" - func: "run tests" vars: resmoke_args: --suites=unittests run_multiple_jobs: true - name: compile_benchmarks depends_on: [] commands: - command: manifest.load - func: "git get project" - func: "get buildnumber" - func: "set up credentials" - func: "fetch and build OpenSSL" - func: "generate compile expansions" # Then we load the generated version data into the agent so we can use it in task definitions - func: "apply compile expansions" - func: "scons compile" vars: targets: benchmarks - command: archive.targz_pack params: target: "benchmarks.tgz" source_dir: "src" include: - "./build/benchmarks.txt" - "./build/**_bm" - "./build/**_bm.gcno" - "./build/**_bm.exe" - "./build/**_bm.pdb" - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: benchmarks.tgz remote_file: ${project}/${build_variant}/${revision}/benchmarks/${build_id}.tgz bucket: mciuploads permissions: public-read content_type: application/tar display_name: Benchmarks ## lint ## - name: lint depends_on: [] commands: - command: manifest.load - func: "git get project" - command: shell.exec type: test params: working_dir: src script: | set -o errexit set -o verbose # Since mypy requires python3, we need to activate the venv_3 ${activate_virtualenv_3} pip install -r buildscripts/requirements.txt updated_mypy=$(PATH=$PATH:/opt/mongodbtoolchain/v2/bin which mypy) deactivate ### ${activate_virtualenv} pip install -r buildscripts/requirements.txt MYPY=$updated_mypy ${compile_env|} $python ./buildscripts/scons.py ${compile_flags|} --stack-size=1024 lint - <<: *task_template name: burn_in_tests depends_on: - name: compile commands: - func: "git get project" # The repository is cloned in a directory distinct from src for the modified test detection # because the extraction of the artifacts performed in the 'do setup' causes # 'git diff --name-only' to see all tests as modified on Windows (git 1.9.5). See SERVER-30634. vars: git_project_directory: burn_in_tests_clonedir - func: "do setup" - command: shell.exec params: working_dir: burn_in_tests_clonedir script: | set -o errexit set -o verbose # If this is a scheduled build, we check for changes against the last scheduled commit. if [ "${is_patch}" != "true" ]; then burn_in_args="--checkEvergreen" fi # Copy the dbtest executable from the src dir because burn_in_tests.py calls it to get the # list of dbtest suites. cp ../src/dbtest${exe} . # Capture a list of new and modified tests. ${activate_virtualenv} $python buildscripts/burn_in_tests.py --branch=${branch_name} --buildVariant=${build_variant} --testListOutfile=jstests/new_tests.json --noExec $burn_in_args # Copy the results to the src dir. cp jstests/new_tests.json ../src/jstests/new_tests.json - func: "do multiversion setup" - func: "run tests" vars: task_path_suffix: /data/multiversion resmoke_wrapper: $python buildscripts/burn_in_tests.py --testListFile=jstests/new_tests.json resmoke_args: --repeat=2 run_multiple_jobs: true - <<: *benchmark_template name: benchmarks_orphaned commands: - func: "do benchmark setup" - func: "run tests" vars: resmoke_args: --suites=benchmarks run_multiple_jobs: false - func: "send benchmark results" - <<: *benchmark_template name: benchmarks_sharding commands: - func: "do benchmark setup" - func: "run tests" vars: resmoke_args: --suites=benchmarks_sharding run_multiple_jobs: false - func: "send benchmark results" - <<: *run_jepsen_template name: jepsen_register_findAndModify_WT commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_read_with_find_and_modify: --read-with-find-and-modify jepsen_storage_engine: --storage-engine wiredTiger jepsen_test_name: register - <<: *run_jepsen_template name: jepsen_register_linearizableRead commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_read_concern: --read-concern linearizable jepsen_storage_engine: --storage-engine mmapv1 jepsen_test_name: register # Running Jepsen when using the MMAPv1 storage engine for longer periods of time can lead to # out of memory issues and task timeouts during the analysis phase. jepsen_time_limit: --time-limit 300 - <<: *run_jepsen_template name: jepsen_register_linearizableRead_WT commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_read_concern: --read-concern linearizable jepsen_storage_engine: --storage-engine wiredTiger jepsen_test_name: register - <<: *run_jepsen_template name: jepsen_set_linearizableRead commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_read_concern: --read-concern linearizable jepsen_storage_engine: --storage-engine mmapv1 jepsen_test_name: set # Running Jepsen when using the MMAPv1 storage engine for longer periods of time can lead to # out of memory issues and task timeouts during the analysis phase. jepsen_time_limit: --time-limit 300 - <<: *run_jepsen_template name: jepsen_set_linearizableRead_WT commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_read_concern: --read-concern linearizable jepsen_storage_engine: --storage-engine wiredTiger jepsen_test_name: set - <<: *run_jepsen_template # We only run this test on wiredTiger since MMAPv1 doesn't support majority read concern. name: jepsen_read-concern-majority_WT commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_storage_engine: --storage-engine wiredTiger jepsen_test_name: read-concern-majority - <<: *run_jepsen_template # We only run this test on wiredTiger since MMAPv1 doesn't support majority read concern. name: jepsen_read-concern-majority_w1_WT commands: - func: "do setup" - func: "do jepsen setup" - func: "run jepsen test" vars: <<: *jepsen_config_vars jepsen_storage_engine: --storage-engine wiredTiger jepsen_test_name: read-concern-majority jepsen_write_concern: --write-concern w1 ## jstestfuzz standalone update generational fuzzer mmapv1 ## - <<: *jstestfuzz_template name: update_fuzzer commands: - func: "do setup" - func: "do multiversion setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: npm_command: update-fuzzer - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars task_path_suffix: /data/multiversion resmoke_args: --suites=generational_fuzzer --storageEngine=mmapv1 ## jstestfuzz replication update generational fuzzer mmapv1 ## - <<: *jstestfuzz_template name: update_fuzzer_replication commands: - func: "do setup" - func: "do multiversion setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: npm_command: update-fuzzer - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars task_path_suffix: /data/multiversion resmoke_args: --suites=generational_fuzzer_replication --storageEngine=mmapv1 ## jstestfuzz standalone update generational fuzzer WT ## - <<: *jstestfuzz_template name: update_fuzzer_WT commands: - func: "do setup" - func: "do multiversion setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: npm_command: update-fuzzer - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars task_path_suffix: /data/multiversion resmoke_args: --suites=generational_fuzzer --storageEngine=wiredTiger ## jstestfuzz replication update generational fuzzer WT ## - <<: *jstestfuzz_template name: update_fuzzer_replication_WT commands: - func: "do setup" - func: "do multiversion setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: npm_command: update-fuzzer - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars task_path_suffix: /data/multiversion resmoke_args: --suites=generational_fuzzer_replication --storageEngine=wiredTiger ## rollback generational fuzzer mmapv1 ## - <<: *jstestfuzz_template name: rollback_fuzzer commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 15 npm_command: rollback-fuzzer - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=rollback_fuzzer --storageEngine=mmapv1 ## rollback generational fuzzer WT ## - <<: *jstestfuzz_template name: rollback_fuzzer_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 15 npm_command: rollback-fuzzer - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=rollback_fuzzer --storageEngine=wiredTiger ## jstestfuzz mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz --storageEngine=mmapv1 ## jstestfuzz WT ## - <<: *jstestfuzz_template name: jstestfuzz_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz --storageEngine=wiredTiger ## jstestfuzz concurrent mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz --storageEngine=mmapv1 --numClientsPerFixture=10 ## jstestfuzz concurrent WT ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz concurrent replica set mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_replication commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication --storageEngine=mmapv1 --numClientsPerFixture=10 ## jstestfuzz concurrent replica set WT ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_replication_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz concurrent replica set with logical session mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_replication_session commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_session --storageEngine=mmapv1 --numClientsPerFixture=10 ## jstestfuzz concurrent replica set with logical session WT ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_replication_session_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_session --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded --storageEngine=mmapv1 --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster WT ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster causal consistency mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded_causal_consistency commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_causal_consistency --storageEngine=mmapv1 --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster causal consistency WT ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded_causal_consistency_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_causal_consistency --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster continuous stepdown ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded_continuous_stepdown commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 25 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_continuous_stepdown --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster with logical session mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded_session commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_session --storageEngine=mmapv1 --numClientsPerFixture=10 ## jstestfuzz concurrent sharded cluster with logical session WT ## - <<: *jstestfuzz_template name: jstestfuzz_concurrent_sharded_session_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_session --storageEngine=wiredTiger --numClientsPerFixture=10 ## jstestfuzz replica set mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_replication commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication --storageEngine=mmapv1 ## jstestfuzz replica set WT ## - <<: *jstestfuzz_template name: jstestfuzz_replication_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication --storageEngine=wiredTiger ## jstestfuzz initial sync replica set mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_replication_initsync commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_initsync --storageEngine=mmapv1 ## jstestfuzz initial sync replica set WT ## - <<: *jstestfuzz_template name: jstestfuzz_replication_initsync_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_initsync --storageEngine=wiredTiger ## jstestfuzz initial sync replica set mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_replication_resync commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_resync --storageEngine=mmapv1 ## jstestfuzz initial sync replica set WT ## - <<: *jstestfuzz_template name: jstestfuzz_replication_resync_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 75 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_resync --storageEngine=wiredTiger ## jstestfuzz replica set with logical session mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_replication_session commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_session --storageEngine=mmapv1 ## jstestfuzz replica set with logical session WT ## - <<: *jstestfuzz_template name: jstestfuzz_replication_session_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_replication_session --storageEngine=wiredTiger ## jstestfuzz sharded cluster mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_sharded commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded --storageEngine=mmapv1 ## jstestfuzz sharded cluster WT ## - <<: *jstestfuzz_template name: jstestfuzz_sharded_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded --storageEngine=wiredTiger ## jstestfuzz sharded cluster causal consistency mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_sharded_causal_consistency commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_causal_consistency --storageEngine=mmapv1 ## jstestfuzz sharded cluster causal consistency WT ## - <<: *jstestfuzz_template name: jstestfuzz_sharded_causal_consistency_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_causal_consistency --storageEngine=wiredTiger ## jstestfuzz sharded cluster continuous stepdown ## - <<: *jstestfuzz_template name: jstestfuzz_sharded_continuous_stepdown commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --numGeneratedFiles 50 --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_continuous_stepdown --storageEngine=wiredTiger ## jstestfuzz sharded cluster with logical session mmapv1 ## - <<: *jstestfuzz_template name: jstestfuzz_sharded_session commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_session --storageEngine=mmapv1 ## jstestfuzz sharded cluster with logical session WT ## - <<: *jstestfuzz_template name: jstestfuzz_sharded_session_WT commands: - func: "do setup" - func: "setup jstestfuzz" - func: "run jstestfuzz" vars: jstestfuzz_vars: --jsTestsDir ../jstests - <<: *run_jstestfuzz_tests vars: <<: *jstestfuzz_config_vars resmoke_args: --suites=jstestfuzz_sharded_session --storageEngine=wiredTiger # Check that the mutational fuzzer can parse JS files modified in a patch build. - name: lint_fuzzer_sanity_patch patch_only: true commands: - command: manifest.load - func: "git get project" - func: "get added and modified patch files" - func: "setup jstestfuzz" - func: "lint fuzzer sanity patch" # Check that the mutational fuzzer can parse all JS filess. - name: lint_fuzzer_sanity_all commands: - command: manifest.load - func: "git get project" - func: "setup jstestfuzz" - func: "lint fuzzer sanity all" ## integration test suites ## - <<: *task_template name: aggregation commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: aggregation_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: aggregation_WT_ese depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation_ese --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: aggregation_auth depends_on: - name: aggregation_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation_auth run_multiple_jobs: true - <<: *task_template name: aggregation_facet_unwind_passthrough_WT depends_on: - name: aggregation_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation_facet_unwind_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: aggregation_read_concern_majority_passthrough_WT depends_on: - name: aggregation_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation_read_concern_majority_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: aggregation_sharded_collections_passthrough_WT depends_on: - name: aggregation_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=aggregation_sharded_collections_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: audit commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=audit --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: audit_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=audit --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: auth commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=auth --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: auth_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=auth --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: auth_audit commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=auth_audit --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: auth_audit_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=auth_audit --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: change_streams_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=change_streams --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: change_streams_mongos_sessions_passthrough_WT depends_on: - name: change_streams_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=change_streams_mongos_sessions_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: change_streams_mongos_passthrough_WT depends_on: - name: change_streams_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=change_streams_mongos_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: change_streams_secondary_reads_WT depends_on: - name: change_streams_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=change_streams_secondary_reads --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: change_streams_sharded_collections_passthrough_WT depends_on: - name: change_streams_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=change_streams_sharded_collections_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: dbtest commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=dbtest --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: dbtest_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=dbtest --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: disk commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=disk_mmapv1 run_multiple_jobs: false - <<: *task_template name: disk_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=disk_wiredtiger --excludeWithAnyTags=requires_mmapv1 run_multiple_jobs: false - <<: *task_template name: durability commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=durability run_multiple_jobs: true - <<: *task_template name: dur_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=dur_jscore_passthrough run_multiple_jobs: true - <<: *task_template name: ese_WT commands: - func: "do setup" - command: shell.exec type: test params: working_dir: src script: | set -o errexit set -o verbose ${activate_virtualenv} $python -m pip install -r src/mongo/db/modules/enterprise/jstests/encryptdb/libs/requirements.txt - func: "run tests" vars: resmoke_args: --suites=ese --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: failpoints commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=failpoints run_multiple_jobs: true - <<: *task_template name: failpoints_auth commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=failpoints_auth run_multiple_jobs: true - <<: *task_template name: gle_auth commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth --shellWriteMode=legacy --shellReadMode=legacy --storageEngine=mmapv1 --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: gle_auth_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth --shellWriteMode=legacy --shellReadMode=legacy --storageEngine=wiredTiger --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: gle_auth_write_cmd commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth --shellWriteMode=commands --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: gle_auth_write_cmd_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth --shellWriteMode=commands --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: gle_auth_basics_passthrough commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth_basics_passthrough --shellWriteMode=legacy --shellReadMode=legacy --storageEngine=mmapv1 --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: gle_auth_basics_passthrough_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth_basics_passthrough --shellWriteMode=legacy --shellReadMode=legacy --storageEngine=wiredTiger --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: gle_auth_basics_passthrough_write_cmd commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth_basics_passthrough --shellWriteMode=commands --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: gle_auth_basics_passthrough_write_cmd_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=gle_auth_basics_passthrough --shellWriteMode=commands --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: integration_tests_standalone commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=integration_tests_standalone run_multiple_jobs: true - <<: *task_template name: integration_tests_replset commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=integration_tests_replset run_multiple_jobs: true - <<: *task_template name: integration_tests_sharded commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=integration_tests_sharded run_multiple_jobs: true - <<: *task_template name: external_auth_WT commands: - func: "do setup" - command: shell.exec type: test params: working_dir: src script: | set -o errexit set -o verbose ${activate_virtualenv} $python -m pip install -r src/mongo/db/modules/enterprise/jstests/external_auth/lib/external-auth-requirements.txt - func: "run tests" vars: resmoke_args: --suites=external_auth --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_gle_auth_basics_passthrough depends_on: - name: gle_auth_basics_passthrough commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_gle_auth_basics_passthrough --shellWriteMode=legacy --shellReadMode=legacy --storageEngine=mmapv1 --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: sharding_gle_auth_basics_passthrough_WT depends_on: - name: gle_auth_basics_passthrough_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_gle_auth_basics_passthrough --shellWriteMode=legacy --shellReadMode=legacy --storageEngine=wiredTiger --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: sharding_gle_auth_basics_passthrough_write_cmd depends_on: - name: gle_auth_basics_passthrough_write_cmd commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_gle_auth_basics_passthrough --shellWriteMode=commands --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: sharding_gle_auth_basics_passthrough_write_cmd_WT depends_on: - name: gle_auth_basics_passthrough_write_cmd_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_gle_auth_basics_passthrough --shellWriteMode=commands --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: jsCore_inMem commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core --storageEngine=inMemory run_multiple_jobs: true - <<: *task_template name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: jsCore_WT_ese depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core_ese --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: jsCore_compatibility depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core --shellReadMode=legacy --shellWriteMode=compatibility --storageEngine=mmapv1 --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: jsCore_compatibility_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core --shellReadMode=legacy --shellWriteMode=compatibility --storageEngine=wiredTiger --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: jsCore_auth depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core_auth run_multiple_jobs: true - <<: *task_template name: jsCore_minimum_batch_size_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core_minimum_batch_size --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: jsCore_op_query depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core_op_query --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: jsCore_op_query_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=core_op_query --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: master_slave_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=master_slave_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: master_slave_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=master_slave_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: replica_sets_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_initsync_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_initsync_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: replica_sets_initsync_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_initsync_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_initsync_static_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_initsync_static_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: replica_sets_initsync_static_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_initsync_static_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_resync_static_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_resync_static_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: replica_sets_resync_static_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_resync_static_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_kill_primary_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_kill_primary_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_kill_secondaries_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_kill_secondaries_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_jscore_fcv34_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_jscore_fcv34_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: replica_sets_jscore_fcv34_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_jscore_fcv34_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: mongosTest commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=mongos_test run_multiple_jobs: true - <<: *task_template name: mmap commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=mmap run_multiple_jobs: true - <<: *task_template name: multiversion commands: - func: "do setup" - func: "do multiversion setup" - func: "run tests" vars: task_path_suffix: /data/multiversion resmoke_args: --suites=multiversion --storageEngine=mmapv1 --excludeWithAnyTags=requires_wiredtiger run_multiple_jobs: true - <<: *task_template name: multiversion_auth commands: - func: "do setup" - func: "do multiversion setup" - func: "run tests" vars: task_path_suffix: /data/multiversion resmoke_args: --suites=multiversion_auth --storageEngine=mmapv1 --excludeWithAnyTags=requires_wiredtiger run_multiple_jobs: true - <<: *task_template name: multiversion_WT commands: - func: "do setup" - func: "do multiversion setup" - func: "run tests" vars: task_path_suffix: /data/multiversion resmoke_args: --suites=multiversion --storageEngine=wiredTiger --excludeWithAnyTags=requires_mmapv1 run_multiple_jobs: true - <<: *task_template name: multiversion_multistorage_engine commands: - func: "do setup" - func: "do multiversion setup" - func: "run tests" vars: task_path_suffix: /data/multiversion resmoke_args: --suites=multiversion_multistorage_engine run_multiple_jobs: true - <<: *task_template name: noPassthrough commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=no_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: noPassthrough_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=no_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: noPassthroughWithMongod commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=no_passthrough_with_mongod --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: noPassthroughWithMongod_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=no_passthrough_with_mongod --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: bulk_gle_passthrough commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=bulk_gle_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: bulk_gle_passthrough_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=bulk_gle_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: slow1 commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=slow1 --storageEngine=mmapv1 run_multiple_jobs: false - <<: *task_template name: slow1_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=slow1 --storageEngine=wiredTiger run_multiple_jobs: false - <<: *task_template name: serial_run commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=serial_run --storageEngine=mmapv1 run_multiple_jobs: false - <<: *task_template name: serial_run_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=serial_run --storageEngine=wiredTiger run_multiple_jobs: false - <<: *task_template name: sharded_collections_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharded_collections_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: sharded_collections_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharded_collections_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: sharding_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: secondary_reads_passthrough depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=secondary_reads_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_jscore_op_query_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_jscore_op_query_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_jscore_passthrough_wire_ops_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_jscore_passthrough --storageEngine=wiredTiger --shellReadMode=legacy --shellWriteMode=compatibility --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: parallel depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=parallel --storageEngine=mmapv1 - <<: *task_template name: parallel_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=parallel --storageEngine=wiredTiger - <<: *task_template name: parallel_compatibility depends_on: - name: jsCore_compatibility commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=parallel --shellReadMode=legacy --shellWriteMode=compatibility --storageEngine=mmapv1 --excludeWithAnyTags=requires_find_command - <<: *task_template name: parallel_compatibility_WT depends_on: - name: jsCore_compatibility_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=parallel --shellReadMode=legacy --shellWriteMode=compatibility --storageEngine=wiredTiger --excludeWithAnyTags=requires_find_command - <<: *task_template name: concurrency commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout vars: resmoke_args: --suites=concurrency --storageEngine=mmapv1 - <<: *task_template name: concurrency_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout vars: resmoke_args: --suites=concurrency --storageEngine=wiredTiger - <<: *task_template name: concurrency_replication commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout vars: resmoke_args: --suites=concurrency_replication --storageEngine=mmapv1 - <<: *task_template name: concurrency_replication_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout vars: resmoke_args: --suites=concurrency_replication --storageEngine=wiredTiger - <<: *task_template name: concurrency_sharded exec_timeout_secs: 43200 # 12 hour timeout for the task overall commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded --storageEngine=mmapv1 - <<: *task_template name: concurrency_sharded_WT exec_timeout_secs: 43200 # 12 hour timeout for the task overall commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded --storageEngine=wiredTiger - <<: *task_template name: concurrency_sharded_causal_consistency commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded_causal_consistency --storageEngine=mmapv1 - <<: *task_template name: concurrency_sharded_causal_consistency_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded_causal_consistency --storageEngine=wiredTiger - <<: *task_template name: concurrency_sharded_causal_consistency_and_balancer commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded_causal_consistency_and_balancer --storageEngine=mmapv1 - <<: *task_template name: concurrency_sharded_causal_consistency_and_balancer_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded_causal_consistency_and_balancer --storageEngine=wiredTiger - <<: *task_template name: concurrency_sharded_with_stepdowns_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded_with_stepdowns --storageEngine=wiredTiger - <<: *task_template name: concurrency_sharded_with_stepdowns_and_balancer_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout for each test vars: resmoke_args: --suites=concurrency_sharded_with_stepdowns_and_balancer --storageEngine=wiredTiger - <<: *task_template name: concurrency_simultaneous commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout vars: resmoke_args: --suites=concurrency jstests/concurrency/fsm_all_simultaneous.js --storageEngine=mmapv1 - <<: *task_template name: concurrency_simultaneous_WT commands: - func: "do setup" - func: "run tests" timeout_secs: 21600 # 6 hour timeout vars: resmoke_args: --suites=concurrency jstests/concurrency/fsm_all_simultaneous.js --storageEngine=wiredTiger - <<: *task_template name: read_concern_linearizable_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_concern_linearizable_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: read_concern_linearizable_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_concern_linearizable_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: read_concern_majority_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_concern_majority_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: write_concern_majority_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=write_concern_majority_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: replica_sets_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_WT_ese commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_ese --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: replica_sets_auth commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_auth run_multiple_jobs: true - <<: *task_template name: replica_sets_pv0 commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_pv0 run_multiple_jobs: true - <<: *task_template name: replica_sets_rollback_refetch_no_uuid commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=replica_sets_rollback_refetch_no_uuid run_multiple_jobs: true - <<: *task_template name: master_slave commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=master_slave --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: master_slave_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=master_slave --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: master_slave_auth depends_on: - name: master_slave_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=master_slave_auth run_multiple_jobs: true - <<: *task_template name: sasl commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sasl run_multiple_jobs: true - <<: *task_template name: sharding commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: sharding_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_csrs_continuous_config_stepdown_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_continuous_config_stepdown --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_WT_ese commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_ese --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_op_query_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding --shellReadMode=legacy --storageEngine=wiredTiger --excludeWithAnyTags=requires_find_command run_multiple_jobs: true - <<: *task_template name: sharding_auth commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_auth run_multiple_jobs: true - <<: *task_template name: sharding_auth_audit_WT depends_on: - name: sharding_auth commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharding_auth_audit --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharding_last_stable_mongos_and_mixed_shards commands: - func: "do setup" - func: "do multiversion setup" - func: "run tests" vars: task_path_suffix: /data/multiversion resmoke_args: --suites=sharding_last_stable_mongos_and_mixed_shards run_multiple_jobs: true - <<: *task_template name: snmp commands: - func: "do setup" - func: "do snmp setup" - func: "run tests" vars: snmp_config_path: SNMPCONFPATH=snmpconf resmoke_args: --suites=snmp --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: snmp_WT commands: - func: "do setup" - func: "do snmp setup" - func: "run tests" vars: snmp_config_path: SNMPCONFPATH=snmpconf resmoke_args: --suites=snmp --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: ssl commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=ssl run_multiple_jobs: true - <<: *task_template name: sslSpecial commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=ssl_special run_multiple_jobs: true - <<: *task_template name: tool commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=tool --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: tool_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=tool --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: jsCore_decimal commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=decimal --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: jsCore_decimal_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=decimal --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: read_only commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_only --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: read_only_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_only --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: read_only_sharded commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_only_sharded --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: read_only_sharded_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=read_only_sharded --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: session_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=session_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: causally_consistent_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=causally_consistent_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: causally_consistent_jscore_passthrough_auth_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=causally_consistent_jscore_passthrough_auth --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: sharded_causally_consistent_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=sharded_causally_consistent_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: retryable_writes_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=retryable_writes_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_default_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_default_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_default_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_default_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_100ms_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_100ms_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_100ms_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_100ms_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_1sec_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_1sec_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_1sec_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_1sec_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_10sec_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_10sec_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_replication_10sec_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_replication_10sec_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_default_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_default_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_default_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_default_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_100ms_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_100ms_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_100ms_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_100ms_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_1sec_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_1sec_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_1sec_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_1sec_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_10sec_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_10sec_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_sharding_10sec_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_sharding_10sec_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_default_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_default_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_default_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_default_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_100ms_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_100ms_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_100ms_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_100ms_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_1sec_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_1sec_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_1sec_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_1sec_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_10sec_refresh_jscore_passthrough depends_on: - name: jsCore commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_10sec_refresh_jscore_passthrough --storageEngine=mmapv1 run_multiple_jobs: true - <<: *task_template name: logical_session_cache_standalone_10sec_refresh_jscore_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=logical_session_cache_standalone_10sec_refresh_jscore_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: retryable_writes_jscore_stepdown_passthrough_WT depends_on: - name: jsCore_WT commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=retryable_writes_jscore_stepdown_passthrough --storageEngine=wiredTiger run_multiple_jobs: true - <<: *task_template name: watchdog commands: - func: "do setup" - func: "do watchdog setup" - func: "run tests" vars: resmoke_args: --suites=watchdog --storageEngine=mmapv1 run_multiple_jobs: false - <<: *task_template name: watchdog_WT commands: - func: "do setup" - func: "do watchdog setup" - func: "run tests" vars: resmoke_args: --suites=watchdog --storageEngine=wiredTiger run_multiple_jobs: false - <<: *task_template name: watchdog_inmemory commands: - func: "do setup" - func: "do watchdog setup" - func: "run tests" vars: resmoke_args: --suites=watchdog --storageEngine=inMemory run_multiple_jobs: false - <<: *task_template name: jsonSchema commands: - func: "do setup" - func: "run tests" vars: resmoke_args: --suites=json_schema --storageEngine=wiredTiger run_multiple_jobs: false - name: powercycle_WT exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test mongod_extra_options: --mongodOptions=\"--setParameter enableTestCommands=1 --storageEngine wiredTiger\" - name: powercycle_fcv3.4_WT exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test client_options: --numCrudClients=20 --numFsmClients=20 --fsmWorkloadBlacklistFiles=toggle_feature_compatibility.js mongod_options: --mongodUsablePorts ${standard_port} ${secret_port} --dbPath=${db_path} --logPath=${log_path} --fcv=3.4 mongod_extra_options: --mongodOptions=\"--setParameter enableTestCommands=1 --storageEngine wiredTiger\" - name: powercycle_kill_mongod exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test crash_options: --crashMethod=kill --crashWaitTime=45 --jitterForCrashWaitTime=5 --instanceId=${instance_id} mongod_extra_options: --mongodOptions=\"--setParameter enableTestCommands=1 --storageEngine wiredTiger\" - name: powercycle_replication_WT exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test mongod_extra_options: --replSet=powercycle --mongodOptions=\"--setParameter enableTestCommands=1 --storageEngine wiredTiger\" - name: powercycle_replication_smalloplog_WT exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test mongod_extra_options: --replSet=powercycle --mongodOptions=\"--setParameter enableTestCommands=1 --oplogSize 20 --storageEngine wiredTiger\" - name: powercycle_syncdelay_WT exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test mongod_extra_options: --mongodOptions=\"--setParameter enableTestCommands=1 --syncdelay 10 --storageEngine wiredTiger\" - name: powercycle_write_concern_majority_WT exec_timeout_secs: 7200 # 2 hour timeout for the task overall depends_on: - name: compile commands: - func: "do setup" - func: "set up remote credentials" vars: <<: *powercycle_remote_credentials - func: "set up EC2 instance" vars: <<: *powercycle_ec2_instance - command: expansions.update <<: *powercycle_expansions - func: "run powercycle test" timeout_secs: 1800 # 30 minute timeout for no output vars: <<: *powercycle_test client_options: "--numCrudClients=20 --numFsmClients=20 --writeConcern='{\"w\": \"majority\"}'" mongod_extra_options: --mongodOptions \"--setParameter enableTestCommands=1 --storageEngine wiredTiger\" - name: idl_tests depends_on: - name: compile commands: - func: "do setup" - func: "run idl tests" - name: package depends_on: - name: compile commands: - func: "fetch artifacts" - func: "set up remote credentials" vars: private_key_file: ~/.ssh/kitchen.pem private_key_remote: ${kitchen_private_key} aws_key_remote: ${kitchen_aws_key} aws_secret_remote: ${kitchen_aws_secret} - func: "run kitchen" - name: push patchable: false depends_on: - name: "*" stepback: false commands: - func: "fetch artifacts" - func: "fetch binaries" - command: s3.get params: aws_key: ${aws_key} aws_secret: ${aws_secret} remote_file: ${project}/${build_variant}/${revision}/binaries/mongo-shell-${build_id}.${ext|tgz} bucket: mciuploads local_file: src/mongo-shell.tgz - command: s3.get params: aws_key: ${aws_key} aws_secret: ${aws_secret} remote_file: ${project}/${build_variant}/${revision}/sources/mongo-src-${build_id}.${ext|tgz} bucket: mciuploads local_file: src/distsrc.${ext|tgz} build_variants: [ linux-64, windows-64-2k8-ssl ] - *apply_compile_expansions - func: "fetch debugsymbols archive" - func: "set up remote credentials" vars: aws_key_remote: ${repo_aws_key} aws_secret_remote: ${repo_aws_secret} - command: shell.exec params: working_dir: src silent: true script: | set -o errexit cat < notary_env.sh export NOTARY_TOKEN=${signing_auth_token_36} export BARQUE_USERNAME=${barque_user} export BARQUE_API_KEY=${barque_api_key} EOF echo "${signing_auth_token_36}" > signing_auth_token - command: shell.exec params: working_dir: src script: | . ./notary_env.sh set -o errexit set -o verbose mv mongo-binaries.tgz mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz} mv mongo-shell.tgz mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz} mv mongo-debugsymbols.tgz mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz} || true mv distsrc.${ext|tgz} mongodb-src-${src_suffix}.${ext|tar.gz} || true /usr/bin/find build/ -type f | grep msi$ | xargs -I original_filename cp original_filename mongodb-win32-${push_arch}-${suffix}.msi || true /usr/local/bin/notary-client.py --key-name "server-3.6" --auth-token-file ${workdir}/src/signing_auth_token --comment "Evergreen Automatic Signing ${revision} - ${build_variant} - ${branch_name}" --notary-url http://notary-service.build.10gen.cc:5000 --skip-missing mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz} mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz} mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz} mongodb-win32-${push_arch}-${suffix}.msi mongodb-src-${src_suffix}.${ext|tar.gz} if [ "${has_packages|}" = "true" ]; then CURATOR_RELEASE=${curator_release|"latest"} curl -L -O http://boxes.10gen.com/build/curator/curator-dist-rhel70-$CURATOR_RELEASE.tar.gz tar -zxvf curator-dist-rhel70-$CURATOR_RELEASE.tar.gz ./curator repo submit --service ${barque_url} --config ./etc/repo_config.yaml --distro ${packager_distro} --edition ${repo_edition} --version ${version} --arch ${packager_arch} --packages https://s3.amazonaws.com/mciuploads/${project}/${build_variant}/${revision}/artifacts/${build_id}-packages.tgz fi # Put the binaries tarball/zipfile - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz} aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz} # Put the shell tarball/zipfile - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz} aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz} # Put the source tarball - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-src-${src_suffix}.${ext|tar.gz} aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz} build_variants: [ linux-64, windows-64-2k8-ssl ] # Put the debug symbols - command: s3.put params: aws_secret: ${aws_secret} aws_key: ${aws_key} permissions: public-read local_file: src/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz} bucket: build-push-testing content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz} optional: true # Put the binaries tarball signature - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sig aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sig # Put the shell tarball signature - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sig aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sig # Put the source tarball signature - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-src-${src_suffix}.${ext|tar.gz}.sig aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.sig build_variants: [ linux-64, windows-64-2k8-ssl ] # Put the debug symbols signature - command: s3.put params: aws_secret: ${aws_secret} aws_key: ${aws_key} permissions: public-read local_file: src/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.sig bucket: build-push-testing content_type: ${content_type|application/x-gzip} remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.sig optional: true # Put the signed MSI file - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} permissions: public-read build_variants: ["enterprise-windows-64-2k8", "windows-64-2k8-ssl"] local_file: src/mongodb-win32-${push_arch}-${suffix}-signed.msi bucket: build-push-testing content_type: application/x-msi remote_file: ${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi # Put the binaries tarball sha1 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha1 aws_key: ${aws_key} permissions: public-read bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha1 # Put the shell tarball sha1 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha1 aws_key: ${aws_key} permissions: public-read bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha1 # Put the source tarball sha1 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-src-${src_suffix}.${ext|tar.gz}.sha1 aws_key: ${aws_key} permissions: public-read bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.sha1 build_variants: [ linux-64, windows-64-2k8-ssl ] # Put the debug symbols sha1 - command: s3.put params: aws_secret: ${aws_secret} aws_key: ${aws_key} permissions: public-read local_file: src/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.sha1 bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.sha1 optional: true # Push the signed MSI sha1 - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} permissions: public-read build_variants: ["enterprise-windows-64-2k8", "windows-64-2k8-ssl"] local_file: src/mongodb-win32-${push_arch}-${suffix}-signed.msi.sha1 bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi.sha1 # Put the binaries tarball sha256 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha256 permissions: public-read aws_key: ${aws_key} bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha256 # Put the shell tarball sha256 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha256 permissions: public-read aws_key: ${aws_key} bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha256 # Put the source tarball sha256 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-src-${src_suffix}.${ext|tar.gz}.sha256 permissions: public-read aws_key: ${aws_key} bucket: build-push-testing content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.sha256 build_variants: [ linux-64, windows-64-2k8-ssl ] # Put the debug symbols sha256 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.sha256 aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.sha256 optional: true # Put the signed MSI sha256 - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} build_variants: ["enterprise-windows-64-2k8", "windows-64-2k8-ssl"] local_file: src/mongodb-win32-${push_arch}-${suffix}-signed.msi.sha256 bucket: build-push-testing permissions: public-read remote_file: ${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi.sha256 content_type: text/plain # Put the binaries tarball md5 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.md5 aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.md5 # Put the shell tarball md5 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.md5 aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.md5 # Put the source tarball md5 - command: s3.put params: aws_secret: ${aws_secret} local_file: src/mongodb-src-${src_suffix}.${ext|tar.gz}.md5 aws_key: ${aws_key} bucket: build-push-testing permissions: public-read content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.md5 build_variants: [ linux-64, windows-64-2k8-ssl ] # Put the debug symbols md5 - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.md5 bucket: build-push-testing content_type: text/plain permissions: public-read remote_file: ${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.md5 optional: true # Put the signed MSI md5 - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} build_variants: ["enterprise-windows-64-2k8", "windows-64-2k8-ssl"] local_file: src/mongodb-win32-${push_arch}-${suffix}-signed.msi.md5 bucket: build-push-testing permissions: public-read content_type: text/plain remote_file: ${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi.md5 - command: s3Copy.copy params: aws_key: ${aws_key} aws_secret: ${aws_secret} s3_copy_files: #Binaries - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}', 'bucket': '${push_bucket}'}} #Shell - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}', 'bucket': '${push_bucket}'}} #Source tarball - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}', 'bucket': 'build-push-testing'}, 'destination': {'path': 'src/mongodb-src-${src_suffix}.${ext|tar.gz}', 'bucket': '${push_bucket}'}, 'build_variants': [ 'linux-64', 'windows-64-2k8-ssl' ]} #MSI (Windows only) - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-win32-${push_arch}-${suffix}-signed.msi', 'bucket': '${push_bucket}'}, 'build_variants': [ 'enterprise-windows-64-2k8', 'windows-64-2k8-ssl' ]} #Binaries Signature - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sig', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sig', 'bucket': '${push_bucket}'}} #Shell Signature - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sig', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sig', 'bucket': '${push_bucket}'}} #Source tarball signature - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.sig', 'bucket': 'build-push-testing'}, 'destination': {'path': 'src/mongodb-src-${src_suffix}.${ext|tar.gz}.sig', 'bucket': '${push_bucket}'}, 'build_variants': [ 'linux-64', 'windows-64-2k8-ssl' ]} #SHA1 for binaries - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha1', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha1', 'bucket': '${push_bucket}'}} #SHA1 for shell - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha1', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha1', 'bucket': '${push_bucket}'}} #SHA1 for source tarball - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.sha1', 'bucket': 'build-push-testing'}, 'destination': {'path': 'src/mongodb-src-${src_suffix}.${ext|tar.gz}.sha1', 'bucket': '${push_bucket}'}, 'build_variants': [ 'linux-64', 'windows-64-2k8-ssl' ]} #SHA1 for MSI - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi.sha1', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-win32-${push_arch}-${suffix}-signed.msi.sha1', 'bucket': '${push_bucket}'}, 'build_variants': ['enterprise-windows-64-2k8', 'windows-64-2k8-ssl']} #SHA256 for binaries - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha256', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha256', 'bucket': '${push_bucket}'}} #SHA256 for shell - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.sha256', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.sha256', 'bucket': '${push_bucket}'}} #SHA256 for source tarball - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.sha256', 'bucket': 'build-push-testing'}, 'destination': {'path': 'src/mongodb-src-${src_suffix}.${ext|tar.gz}.sha256', 'bucket': '${push_bucket}'}, 'build_variants': [ 'linux-64', 'windows-64-2k8-ssl' ]} #SHA256 for MSI files - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi.sha256', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-win32-${push_arch}-${suffix}-signed.msi.sha256', 'bucket': '${push_bucket}'}, 'build_variants': ['enterprise-windows-64-2k8', 'windows-64-2k8-ssl']} #MD5 for binaries - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.md5', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-${suffix}.${ext|tgz}.md5', 'bucket': '${push_bucket}'}} #MD5 for shell - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-shell-${push_name}-${push_arch}-${suffix}-${task_id}.${ext|tgz}.md5', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-shell-${push_name}-${push_arch}-${suffix}.${ext|tgz}.md5', 'bucket': '${push_bucket}'}} #MD5 for source tarball - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-src-${src_suffix}-${task_id}.${ext|tar.gz}.md5', 'bucket': 'build-push-testing'}, 'destination': {'path': 'src/mongodb-src-${src_suffix}.${ext|tar.gz}.md5', 'bucket': '${push_bucket}'}, 'build_variants': [ 'linux-64', 'windows-64-2k8-ssl' ]} #MD5 for MSIs - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-win32-${push_arch}-${suffix}-${task_id}-signed.msi.md5', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-win32-${push_arch}-${suffix}-signed.msi.md5', 'bucket': '${push_bucket}'}, 'build_variants': ['enterprise-windows-64-2k8', 'windows-64-2k8-ssl'], } # Debug symbols are not created for all variants and the copy is optional. - command: s3Copy.copy params: aws_key: ${aws_key} aws_secret: ${aws_secret} optional: true s3_copy_files: #Debug Symbols - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}', 'bucket': '${push_bucket}'}} #Debug Symbols Signature - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.sig', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.sig', 'bucket': '${push_bucket}'}} #SHA1 for debug symbols - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.sha1', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.sha1', 'bucket': '${push_bucket}'}} #SHA256 for debugsymbols - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.sha256', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.sha256', 'bucket': '${push_bucket}'}} #MD5 for debugsymbols - {'source': {'path': '${push_path}-STAGE/${push_name}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}-${task_id}.${ext|tgz}.md5', 'bucket': 'build-push-testing'}, 'destination': {'path': '${push_path}/mongodb-${push_name}-${push_arch}-debugsymbols-${suffix}.${ext|tgz}.md5', 'bucket': '${push_bucket}'}} - name: fetch_test_lifecycle depends_on: [] commands: - func: "git get project" - func: "fetch test_lifecycle.yml" vars: # This task is meant to fail if there is an error while fetching test_lifecycle.yml since # the compile task won't fail. fail_task_on_error: true - name: update_test_lifecycle exec_timeout_secs: 21600 # 6 hour timeout for the task overall depends_on: [] commands: - func: "git get project" - command: shell.exec timeout_secs: 14400 # Timeout if there is no output for 4 hours type: test params: working_dir: src silent: true script: | set -o errexit # Set up the git ssh private key mkdir -p ~/.ssh echo -n "${testlifecycle_ssh_key}" > ~/.ssh/test_lifecycle.pem chmod 0600 ~/.ssh/test_lifecycle.pem export GIT_SSH_COMMAND="ssh -i ~/.ssh/test_lifecycle.pem" # Create the jira credentials configuration file cat > .jira.yml <- --ssl MONGO_DISTMOD=rhel62 -j$(grep -c ^processor /proc/cpuinfo) --release --variables-files=etc/scons/mongodbtoolchain_gcc.vars multiversion_platform: rhel62 multiversion_edition: enterprise num_jobs_available: $(grep -c ^processor /proc/cpuinfo) has_packages: true packager_script: packager-enterprise.py packager_arch: x86_64 packager_distro: rhel62 repo_edition: enterprise use_scons_cache: true tooltags: "ssl sasl" build_mongoreplay: true tasks: - name: compile - name: aggregation - name: audit - name: auth - name: auth_audit - name: bulk_gle_passthrough - name: concurrency - name: concurrency_replication - name: concurrency_sharded - name: concurrency_sharded_causal_consistency - name: concurrency_sharded_causal_consistency_and_balancer - name: concurrency_simultaneous - name: dbtest - name: disk - name: dur_jscore_passthrough - name: durability - name: gle_auth - name: gle_auth_basics_passthrough - name: gle_auth_basics_passthrough_write_cmd - name: gle_auth_write_cmd - name: jsCore - name: jsCore_compatibility - name: jsCore_decimal - name: jsCore_op_query - name: jstestfuzz - name: jstestfuzz_concurrent - name: jstestfuzz_concurrent_replication - name: jstestfuzz_concurrent_replication_session - name: jstestfuzz_concurrent_sharded - name: jstestfuzz_concurrent_sharded_causal_consistency - name: jstestfuzz_concurrent_sharded_continuous_stepdown - name: jstestfuzz_concurrent_sharded_session - name: jstestfuzz_replication - name: jstestfuzz_replication_initsync - name: jstestfuzz_replication_resync - name: jstestfuzz_replication_session - name: jstestfuzz_sharded - name: jstestfuzz_sharded_causal_consistency - name: jstestfuzz_sharded_continuous_stepdown - name: jstestfuzz_sharded_session - name: logical_session_cache_replication_1sec_refresh_jscore_passthrough - name: logical_session_cache_sharding_1sec_refresh_jscore_passthrough - name: logical_session_cache_standalone_1sec_refresh_jscore_passthrough - name: master_slave - name: master_slave_jscore_passthrough - name: mmap - name: multiversion - name: multiversion_auth - name: noPassthrough - name: noPassthroughWithMongod - name: parallel - name: parallel_compatibility - name: read_concern_linearizable_passthrough distros: - rhel62-large - name: read_only - name: read_only_sharded - name: replica_sets distros: - rhel62-large - name: replica_sets_initsync_jscore_passthrough distros: - rhel62-large - name: replica_sets_initsync_static_jscore_passthrough distros: - rhel62-large - name: replica_sets_jscore_passthrough distros: - rhel62-large - name: replica_sets_jscore_fcv34_passthrough distros: - rhel62-large - name: replica_sets_resync_static_jscore_passthrough distros: - rhel62-large - name: rollback_fuzzer - name: serial_run - name: sharded_collections_jscore_passthrough - name: sharding distros: - rhel62-large - name: sharding_gle_auth_basics_passthrough - name: sharding_gle_auth_basics_passthrough_write_cmd - name: sharding_jscore_passthrough - name: slow1 - name: snmp - name: tool - name: update_fuzzer - name: update_fuzzer_replication - name: enterprise-windows-64-2k8-mmapv1 display_name: Enterprise Windows 2008R2 MMAPv1 modules: - enterprise run_on: - windows-64-vs2015-small batchtime: 1440 # 1 day expansions: test_flags: --storageEngine=mmapv1 --excludeWithAnyTags=requires_wiredtiger,requires_document_locking,requires_majority_read_concern platform_decompress: unzip exe: .exe multiversion_platform: windows multiversion_edition: enterprise msi_target: msi content_type: application/zip compile_flags: >- --release --ssl MONGO_DISTMOD=windows-64 CPPPATH="c:/openssl/include c:/sasl/include c:/snmp/include c:/curl/include" LIBPATH="c:/openssl/lib c:/sasl/lib c:/snmp/lib c:/curl/lib" -j$(( $(grep -c ^processor /proc/cpuinfo) / 2 )) --dynamic-windows --win-version-min=ws08r2 num_scons_compile_all_jobs_available: $(( $(grep -c ^processor /proc/cpuinfo) / 4 )) python: python num_jobs_available: $(grep -c ^processor /proc/cpuinfo) ext: zip use_scons_cache: true tooltags: "ssl sasl" build_mongoreplay: false tasks: - name: compile - name: aggregation - name: audit - name: auth - name: auth_audit - name: bulk_gle_passthrough - name: concurrency distros: - windows-64-vs2015-large - name: concurrency_replication distros: - windows-64-vs2015-large - name: concurrency_sharded distros: - windows-64-vs2015-large - name: concurrency_sharded_causal_consistency distros: - windows-64-vs2015-large - name: concurrency_sharded_causal_consistency_and_balancer distros: - windows-64-vs2015-large - name: concurrency_simultaneous distros: - windows-64-vs2015-large - name: dbtest - name: disk - name: dur_jscore_passthrough - name: durability - name: gle_auth - name: gle_auth_basics_passthrough - name: gle_auth_basics_passthrough_write_cmd - name: gle_auth_write_cmd - name: jsCore - name: jsCore_compatibility - name: jsCore_decimal - name: jsCore_op_query - name: jstestfuzz - name: jstestfuzz_concurrent - name: jstestfuzz_concurrent_replication - name: jstestfuzz_concurrent_replication_session - name: jstestfuzz_concurrent_sharded - name: jstestfuzz_concurrent_sharded_causal_consistency - name: jstestfuzz_concurrent_sharded_continuous_stepdown - name: jstestfuzz_concurrent_sharded_session - name: jstestfuzz_replication - name: jstestfuzz_replication_initsync - name: jstestfuzz_replication_resync - name: jstestfuzz_replication_session - name: jstestfuzz_sharded - name: jstestfuzz_sharded_causal_consistency - name: jstestfuzz_sharded_continuous_stepdown - name: jstestfuzz_sharded_session - name: logical_session_cache_replication_1sec_refresh_jscore_passthrough - name: logical_session_cache_sharding_1sec_refresh_jscore_passthrough - name: logical_session_cache_standalone_1sec_refresh_jscore_passthrough - name: master_slave - name: master_slave_jscore_passthrough - name: mmap - name: noPassthrough - name: noPassthroughWithMongod - name: parallel - name: parallel_compatibility - name: read_concern_linearizable_passthrough distros: - windows-64-vs2015-large - name: read_only - name: read_only_sharded - name: replica_sets distros: - windows-64-vs2015-large - name: replica_sets_initsync_jscore_passthrough distros: - windows-64-vs2015-large - name: replica_sets_initsync_static_jscore_passthrough distros: - windows-64-vs2015-large - name: replica_sets_jscore_passthrough distros: - windows-64-vs2015-large - name: replica_sets_jscore_fcv34_passthrough distros: - windows-64-vs2015-large - name: replica_sets_resync_static_jscore_passthrough distros: - windows-64-vs2015-large - name: rollback_fuzzer - name: serial_run - name: sharded_collections_jscore_passthrough - name: sharding distros: - windows-64-vs2015-large - name: sharding_gle_auth_basics_passthrough - name: sharding_gle_auth_basics_passthrough_write_cmd - name: sharding_jscore_passthrough - name: slow1 - name: snmp - name: tool - name: update_fuzzer - name: update_fuzzer_replication - name: enterprise-osx-1010-mmapv1 display_name: Enterprise OS X 10.10 MMAPv1 modules: - enterprise run_on: - macos-1012 batchtime: 1440 # 1 day expansions: test_flags: --storageEngine=mmapv1 --excludeWithAnyTags=requires_wiredtiger,requires_ssl_mongo_tools,requires_document_locking,requires_majority_read_concern compile_env: DEVELOPER_DIR=/Applications/Xcode8.3.app compile_flags: >- --ssl -j$(sysctl -n hw.logicalcpu) --release --libc++ CCFLAGS="-mmacosx-version-min=10.10" LINKFLAGS="-mmacosx-version-min=10.10" LIBPATH="$PWD/openssl_install_dir/lib" CPPPATH="$PWD/openssl_install_dir/include" multiversion_platform: osx multiversion_edition: enterprise num_jobs_available: 1 tooltags: "ssl sasl" build_mongoreplay: true build_openssl: true openssl_config_flags: "-mmacosx-version-min=10.10" openssl_make_flags: "-j$(sysctl -n hw.logicalcpu)" tasks: - name: compile - name: aggregation - name: audit - name: auth - name: auth_audit - name: bulk_gle_passthrough - name: concurrency - name: concurrency_replication - name: concurrency_sharded - name: concurrency_sharded_causal_consistency - name: concurrency_sharded_causal_consistency_and_balancer - name: concurrency_simultaneous - name: dbtest - name: disk - name: dur_jscore_passthrough - name: durability - name: gle_auth - name: gle_auth_basics_passthrough - name: gle_auth_basics_passthrough_write_cmd - name: gle_auth_write_cmd - name: jsCore - name: jsCore_compatibility - name: jsCore_decimal - name: jsCore_op_query - name: jstestfuzz - name: jstestfuzz_concurrent - name: jstestfuzz_concurrent_replication - name: jstestfuzz_concurrent_replication_session - name: jstestfuzz_concurrent_sharded - name: jstestfuzz_concurrent_sharded_causal_consistency - name: jstestfuzz_concurrent_sharded_continuous_stepdown - name: jstestfuzz_concurrent_sharded_session - name: jstestfuzz_replication - name: jstestfuzz_replication_initsync - name: jstestfuzz_replication_resync - name: jstestfuzz_replication_session - name: jstestfuzz_sharded - name: jstestfuzz_sharded_causal_consistency - name: jstestfuzz_sharded_continuous_stepdown - name: jstestfuzz_sharded_session - name: logical_session_cache_replication_1sec_refresh_jscore_passthrough - name: logical_session_cache_sharding_1sec_refresh_jscore_passthrough - name: logical_session_cache_standalone_1sec_refresh_jscore_passthrough - name: master_slave - name: master_slave_jscore_passthrough - name: mmap - name: noPassthrough - name: noPassthroughWithMongod - name: parallel - name: parallel_compatibility - name: read_concern_linearizable_passthrough - name: read_only - name: read_only_sharded - name: replica_sets - name: replica_sets_initsync_jscore_passthrough - name: replica_sets_initsync_static_jscore_passthrough - name: replica_sets_jscore_passthrough - name: replica_sets_jscore_fcv34_passthrough - name: replica_sets_resync_static_jscore_passthrough - name: rollback_fuzzer - name: serial_run - name: sharded_collections_jscore_passthrough - name: sharding_gle_auth_basics_passthrough - name: sharding_gle_auth_basics_passthrough_write_cmd - name: sharding_jscore_passthrough - name: slow1 - name: tool - name: update_fuzzer - name: linux-64-duroff-mmapv1 display_name: Linux (No Journal) MMAPv1 run_on: - rhel62-small batchtime: 1440 # 1 day expansions: push_path: linux push_bucket: downloads.mongodb.org push_name: linux-duroff push_arch: x86_64 compile_flags: -j$(grep -c ^processor /proc/cpuinfo) --variables-files=etc/scons/mongodbtoolchain_gcc.vars num_jobs_available: $(grep -c ^processor /proc/cpuinfo) test_flags: --nojournal --storageEngine=mmapv1 --excludeWithAnyTags=requires_wiredtiger,requires_journaling,requires_document_locking,requires_majority_read_concern use_scons_cache: true tooltags: "" build_mongoreplay: false tasks: - name: compile distros: - rhel62-large - name: compile_all distros: - rhel62-large - name: aggregation - name: auth - name: bulk_gle_passthrough - name: concurrency - name: concurrency_replication - name: concurrency_sharded - name: concurrency_simultaneous - name: dbtest - name: disk - name: durability - name: dur_jscore_passthrough - name: failpoints - name: failpoints_auth - name: jsCore - name: jsCore_auth - name: jsCore_compatibility - name: jstestfuzz - name: jstestfuzz_concurrent - name: jstestfuzz_concurrent_replication - name: jstestfuzz_concurrent_sharded - name: jstestfuzz_replication - name: jstestfuzz_sharded - name: logical_session_cache_replication_1sec_refresh_jscore_passthrough - name: logical_session_cache_sharding_1sec_refresh_jscore_passthrough - name: logical_session_cache_standalone_1sec_refresh_jscore_passthrough - name: master_slave - name: master_slave_jscore_passthrough - name: mmap - name: noPassthrough - name: noPassthroughWithMongod - name: parallel - name: parallel_compatibility - name: replica_sets - name: replica_sets_auth - name: replica_sets_jscore_passthrough - name: serial_run - name: sharding - name: sharding_auth - name: sharding_jscore_passthrough - name: slow1 - name: tool