stepback: false command_type: system pre: post: # Do cluster teardown first to ensure runtime is below Evergreen's post timeout. Other post tasks # will not have been run if the timeout is exceeded. - command: shell.exec params: working_dir: work script: | source ./dsienv.sh if [ -e /data/infrastructure_provisioning/terraform/provisioned.${cluster} ]; then mark_idle.sh fi - command: shell.exec params: working_dir: work script: | set -e set -v source ./dsienv.sh ../src/dsi/dsi/run-dsi infrastructure_teardown.py - command: shell.exec params: working_dir: work script: | source ./dsienv.sh make_artifact.sh - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: work/dsi-artifacts.tgz remote_file: ${project_dir}/${build_variant}/${revision}/${task_id}/${version_id}/logs/dsi-artifacts-${task_name}-${build_id}-${execution}.${ext|tgz} bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: Dsi Artifacts - Execution ${execution} - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/workloads/workloads/jsdoc/jsdocs-redirect.html remote_file: ${project_dir}/${build_variant}/${revision}/${task_id}/${version_id}/logs/workloads-${task_name}-${build_id}.html bucket: mciuploads permissions: public-read content_type: text/html display_name: workloads documentation - command: attach.results params: file_location: work/report.json - command: "json.send" params: name: "perf" file: "work/perf.json" - func: "upload pip requirements" - command: shell.exec params: working_dir: src script: | # removes files from the (local) scons cache when it's over a # threshold, to the $prune_ratio percentage. Ideally override # these default values in the distro config in evergreen. if [ -d "${scons_cache_path}" ]; then /opt/mongodbtoolchain/v3/bin/python3 buildscripts/scons_cache_prune.py --cache-dir ${scons_cache_path} --cache-size ${scons_cache_size|200} --prune-ratio ${scons_prune_ratio|0.8} fi functions: "git get project": &git_get_project command: git.get_project params: directory: src revisions: # for each module include revision as : ${_rev} dsi: ${dsi_rev} enterprise: ${enterprise_rev} linkbench: ${linkbench_rev} genny: ${genny_rev} workloads: ${workloads_rev} mongo-tools: ${mongo-tools_rev} "compile mongodb": # We create a virtual environment with the Python dependencies for compiling the server # installed. - command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose /opt/mongodbtoolchain/v3/bin/virtualenv --python /opt/mongodbtoolchain/v3/bin/python3 "${workdir}/compile_venv" /opt/mongodbtoolchain/v3/bin/virtualenv --python /opt/mongodbtoolchain/v3/bin/python2 "${workdir}/venv" source "${workdir}/compile_venv/bin/activate" python -m pip install -r etc/pip/compile-requirements.txt - command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose mkdir -p mongodb/bin # We get the raw version string (r1.2.3-45-gabcdef) from git MONGO_VERSION=$(git describe --abbrev=7) # If this is a patch build, we add the patch version id to the version string so we know # this build was a patch, and which evergreen task it came from if [ "${is_patch|false}" = "true" ]; then MONGO_VERSION="$MONGO_VERSION-patch-${version_id}" fi # This script converts the generated version string into a sanitized version string for # use by scons and uploading artifacts as well as information about for the scons cache. source "${workdir}/compile_venv/bin/activate" MONGO_VERSION=$MONGO_VERSION USE_SCONS_CACHE=${use_scons_cache|false} python buildscripts/generate_compile_expansions.py --out compile_expansions.yml # Then we load the generated version data into the agent so we can use it in task definitions - command: expansions.update params: file: src/compile_expansions.yml - command: shell.exec params: working_dir: src/mongo-tools/src/github.com/mongodb/mongo-tools script: | set -o verbose set -o errexit # make sure newlines in the scripts are handled correctly by windows if [ "Windows_NT" = "$OS" ]; then set -o igncr fi; # set_goenv provides set_goenv(), print_ldflags() and print_tags() used below . ./set_goenv.sh GOROOT="" set_goenv || exit go version build_tools="bsondump mongostat mongofiles mongoexport mongoimport mongorestore mongodump mongotop" if [ "${build_mongoreplay}" = "true" ]; then build_tools="$build_tools mongoreplay" fi for i in $build_tools; do go build -ldflags "$(print_ldflags)" ${args} -tags "$(print_tags ${tooltags})" -o "../../../../../mongodb/bin/$i${exe|}" $i/main/$i.go "../../../../../mongodb/bin/$i${exe|}" --version done - command: shell.exec params: working_dir: src script: | set -o errexit set -o verbose source "${workdir}/compile_venv/bin/activate" python ./buildscripts/scons.py ${compile_flags|} ${scons_cache_args|} --install-mode=hygienic install-core MONGO_VERSION=${version} DESTDIR=$(pwd)/mongodb mkdir -p mongodb/jstests/hooks if [ -d jstests/hooks ] then echo "Fetching JS test DB correctness checks from directory jstests" cp -a jstests/* mongodb/jstests echo "Now adding our own special run_validate_collections.js wrapper" mv mongodb/jstests/hooks/run_validate_collections.js mongodb/jstests/hooks/run_validate_collections.actual.js cat << EOF > mongodb/jstests/hooks/run_validate_collections.js print("NOTE: run_validate_collections.js will skip the oplog!"); TestData = { skipValidationNamespaces: ['local.oplog.rs'] }; load('jstests/hooks/run_validate_collections.actual.js'); EOF fi tar czf mongodb${compile-variant|}.tar.gz mongodb - command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: src/mongodb${compile-variant|}.tar.gz remote_file: ${project_dir}/${version_id}/${revision}/${platform}/mongodb${compile-variant|}-${version_id}.tar.gz bucket: mciuploads permissions: public-read content_type: ${content_type|application/x-gzip} display_name: mongodb${compile-variant|}.tar.gz "write yml config": &write_yml_config command: shell.exec params: working_dir: work script: | cat > bootstrap.yml < runtime.yml < aws_ssh_key.pem chmod 400 aws_ssh_key.pem cat > runtime_secret.yml < post_run_check.status - command: shell.exec params: working_dir: work silent: true script: | set -o errexit is_patch=${is_patch} task_id=${task_id} perf_jira_user=${perf_jira_user} perf_jira_pw=${perf_jira_pw} analysis_user=${dsi_analysis_atlas_user} analysis_password=${dsi_analysis_atlas_pw} evergreen_api_key=${evergreen_api_key} evergreen_api_user=${evergreen_api_user} source ../src/buildscripts/signal_processing_setup.sh - command: shell.exec params: working_dir: work script: | set -o verbose source ./signal_processing_venv/bin/activate detect-changes --config .signal-processing.yml - command: shell.exec params: working_dir: work script: | set -o verbose source ./signal_processing_venv/bin/activate detect-outliers --config .signal-processing.yml - command: shell.exec type: setup params: working_dir: work script: | set -o verbose filename=rejects.json if [ -s "$filename" ]; then echo "Rejecting task due to the following outliers:" cat "$filename" exit ${detected_outlier_exit_code|0} fi - command: shell.exec type: test params: working_dir: work script: | set -o verbose exit $(cat post_run_check.status) "upload pip requirements": command: s3.put params: aws_key: ${aws_key} aws_secret: ${aws_secret} local_file: pip-requirements.txt remote_file: ${project}/${build_variant}/${revision}/pip-requirements-${task_id}-${execution}.txt bucket: mciuploads permissions: public-read content_type: atext-plain display_name: Pip Requirements ####################################### # Tasks # ####################################### tasks: - name: compile commands: - command: manifest.load - func: "git get project" - func: "compile mongodb" - name: linkbench priority: 5 commands: - func: "prepare environment" vars: test: "linkbench" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: linkbench2 priority: 5 exec_timeout_secs: 43200 # 12 hours commands: - func: "prepare environment" vars: test: "linkbench2" additional_tfvars: "tags: {expire-on-delta: 12}" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: tpcc priority: 5 commands: - func: "prepare environment" vars: test: "tpcc" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: industry_benchmarks priority: 5 commands: - func: "prepare environment" vars: test: "ycsb" - func: "deploy cluster" - func: "run test" - func: "analyze" vars: script_flags: --ycsb-throughput-analysis reports - name: ycsb_60GB priority: 5 commands: - func: "prepare environment" vars: test: "ycsb-60GB" - func: "deploy cluster" - func: "run test" - func: "analyze" vars: script_flags: --ycsb-throughput-analysis reports - name: industry_benchmarks_secondary_reads priority: 5 commands: - func: "prepare environment" vars: test: "ycsb-secondary-reads" - func: "deploy cluster" - func: "run test" - func: "analyze" vars: script_flags: --ycsb-throughput-analysis reports - name: industry_benchmarks_wmajority priority: 5 commands: - func: "prepare environment" vars: test: "ycsb-wmajority" - func: "deploy cluster" - func: "run test" - func: "analyze" vars: script_flags: --ycsb-throughput-analysis reports - name: crud_workloads priority: 5 commands: - func: "prepare environment" vars: test: "crud_workloads" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: crud_workloads_majority priority: 5 commands: - func: "prepare environment" vars: test: "crud_workloads_majority" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: cursor_manager priority: 5 commands: - func: "prepare environment" vars: test: "cursor_manager" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: mixed_workloads priority: 5 commands: - func: "prepare environment" vars: test: "mixed_workloads" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: misc_workloads priority: 5 commands: - func: "prepare environment" vars: test: "misc_workloads" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: map_reduce_workloads priority: 5 commands: - func: "prepare environment" vars: test: "map_reduce_workloads" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: smoke_test priority: 5 commands: - func: "prepare environment" vars: test: "short" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: genny_canaries priority: 5 commands: - func: "prepare environment" vars: test: "genny_canaries" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: retryable_writes_workloads priority: 5 commands: - func: "prepare environment" vars: test: "retryable_writes" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: snapshot_reads priority: 5 commands: - func: "prepare environment" vars: test: "snapshot_reads" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: secondary_reads priority: 5 commands: - func: "prepare environment" vars: test: "secondary_reads" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: bestbuy_agg priority: 5 commands: - func: "prepare environment" vars: test: "bestbuy_agg" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: bestbuy_agg_merge_same_db priority: 5 commands: - func: "prepare environment" vars: test: "bestbuy_agg_merge_same_db" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: bestbuy_agg_merge_different_db priority: 5 commands: - func: "prepare environment" vars: test: "bestbuy_agg_merge_different_db" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: bestbuy_agg_merge_target_hashed priority: 5 commands: - func: "prepare environment" vars: test: "bestbuy_agg_merge_target_hashed" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: bestbuy_agg_merge_wordcount priority: 5 commands: - func: "prepare environment" vars: test: "bestbuy_agg_merge_wordcount" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: bestbuy_query priority: 5 commands: - func: "prepare environment" vars: test: "bestbuy_query" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: non_sharded_workloads priority: 5 commands: - func: "prepare environment" vars: test: "non_sharded" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: mongos_workloads priority: 5 commands: - func: "prepare environment" vars: test: "mongos" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: mongos_large_catalog_workloads priority: 5 commands: - func: "prepare environment" vars: test: "mongos_large_catalog" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: move_chunk_workloads priority: 5 commands: - func: "prepare environment" vars: test: "move_chunk" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: move_chunk_waiting_workloads priority: 5 commands: - func: "prepare environment" vars: test: "move_chunk_waiting" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: refine_shard_key_transaction_stress priority: 5 commands: - func: "prepare environment" vars: test: "refine_shard_key_transaction_stress" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: secondary_performance priority: 5 commands: - func: "prepare environment" vars: # Unfortunately the dash/underscore style is different for mongodb_setup and test_control test: "secondary_performance" setup: "secondary-performance" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: initialsync priority: 5 commands: - func: "prepare environment" vars: test: "initialsync" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: initialsync-logkeeper-short priority: 5 commands: - func: "prepare environment" vars: test: "initialsync-logkeeper" setup: "initialsync-logkeeper-short" # Logkeeper dataset with FCV set to 4.4 dataset: "https://s3-us-west-2.amazonaws.com/dsi-donot-remove/InitialSyncLogKeeper/logkeeper-slice-data-mongodb-4.4.tgz" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: initialsync-logkeeper priority: 5 exec_timeout_secs: 216000 # 2.5 days commands: - func: "prepare environment" vars: test: "initialsync-logkeeper" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: initialsync-large priority: 5 commands: - func: "prepare environment" vars: test: "initialsync-large" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: change_streams_throughput priority: 5 commands: - func: "prepare environment" vars: test: "change_streams_throughput" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: change_streams_latency priority: 5 commands: - func: "prepare environment" vars: test: "change_streams_latency" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: change_streams_multi_mongos priority: 5 commands: - func: "prepare environment" vars: test: "change_streams_multi_mongos" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: genny_generate_all_tasks priority: 5 commands: - func: "git get project" - command: shell.exec params: working_dir: src script: | genny/genny/scripts/genny_auto_tasks.sh --generate-all-tasks --output build/all_tasks.json cat ../src/genny/genny/build/all_tasks.json - command: generate.tasks params: files: - src/genny/genny/build/all_tasks.json - name: genny_auto_tasks priority: 5 commands: - func: "git get project" - command: shell.exec params: script: | mkdir work - *write_yml_config - command: shell.exec params: working_dir: work script: | ../src/genny/genny/scripts/genny_auto_tasks.sh --output build/auto_tasks.json --variants "${build_variant}" --autorun cat ../src/genny/genny/build/auto_tasks.json - command: generate.tasks params: files: - src/genny/genny/build/auto_tasks.json - name: genny_patch_tasks patch_only: true priority: 5 commands: - func: "git get project" - command: shell.exec params: working_dir: src script: | # add --forced-workloads to run specific workloads in addition, i.e. adding the following to the end of the command below: # --forced-workloads scale/BigUpdate.yml execution/CreateIndex.yml # will run all locally added/modified workloads, plus big_update and create_index genny/genny/scripts/genny_auto_tasks.sh --output build/patch_tasks.json --variants "${build_variant}" --modified cat genny/genny/build/patch_tasks.json - command: generate.tasks params: files: - src/genny/genny/build/patch_tasks.json - name: sb_large_scale priority: 5 exec_timeout_secs: 43200 # 12 hours commands: - func: "prepare environment" vars: test: "sb_large_scale" additional_tfvars: "tags: {expire-on-delta: 12}" - func: "deploy cluster" - func: "run test" - func: "analyze" - name: sb_timeseries priority: 5 commands: - func: "prepare environment" vars: test: "sb_timeseries" - func: "deploy cluster" - func: "run test" - func: "analyze" ####################################### # Modules # ####################################### # if a module is added and to be added to the manifest # be sure to add the module to git.get_project revisions parameter modules: - name: dsi repo: git@github.com:10gen/dsi.git prefix: dsi branch: master - name: genny repo: git@github.com:10gen/genny.git prefix: genny branch: master - name: workloads repo: git@github.com:10gen/workloads.git prefix: workloads branch: master - name: linkbench repo: git@github.com:10gen/linkbench.git prefix: linkbench branch: master - name: enterprise repo: git@github.com:10gen/mongo-enterprise-modules.git prefix: src/mongo/db/modules branch: v4.4 - name: mongo-tools repo: git@github.com:mongodb/mongo-tools.git prefix: mongo-tools/src/github.com/mongodb branch: master ####################################### # Buildvariants # ####################################### buildvariants: # We are explicitly tracking the Amazon Linux 2 variant compile options from etc/evergreen.yml. If we can get # proper artifacts directly from that project, we should do that and remove these tasks. - &compile-amazon2 name: compile-amazon2 display_name: Compile modules: - mongo-tools - enterprise - genny batchtime: 1440 # 24 hours expansions: &compile-expansions platform: linux project_dir: &project_dir dsi tooltags: "" use_scons_cache: true compile_flags: >- --ssl MONGO_DISTMOD=amazon2 -j$(grep -c ^processor /proc/cpuinfo) --release --variables-files=etc/scons/mongodbtoolchain_v3_gcc.vars install-mongocryptd run_on: - "amazon2-build" tasks: - name: compile - name: genny_generate_all_tasks - name: compile-rhel70 display_name: Compile for Atlas-like modules: - mongo-tools - enterprise batchtime: 1440 # 24 hours expansions: <<: *compile-expansions compile_flags: >- --ssl MONGO_DISTMOD=rhel70 -j$(grep -c ^processor /proc/cpuinfo) --release --variables-files=etc/scons/mongodbtoolchain_v3_gcc.vars compile-variant: -rhel70 run_on: - rhel70-small tasks: - name: compile ####################################### # Linux Buildvariants # ####################################### - name: linux-1-node-replSet display_name: Linux 1-Node ReplSet batchtime: 10080 # 7 days modules: &modules - dsi - genny - workloads - linkbench expansions: setup: single-replica cluster: single platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: &1nodetasks - name: industry_benchmarks - name: ycsb_60GB - name: crud_workloads - name: mixed_workloads - name: misc_workloads - name: map_reduce_workloads - name: smoke_test - name: retryable_writes_workloads - name: non_sharded_workloads - name: bestbuy_agg - name: bestbuy_agg_merge_different_db - name: bestbuy_agg_merge_same_db - name: bestbuy_agg_merge_wordcount - name: bestbuy_query - name: change_streams_throughput - name: change_streams_latency - name: snapshot_reads - name: linkbench - name: linkbench2 - name: tpcc - name: genny_patch_tasks - name: genny_auto_tasks - name: industry_benchmarks_wmajority - name: sb_large_scale - name: sb_timeseries - name: linux-standalone display_name: Linux Standalone batchtime: 1440 # 24 hours modules: *modules expansions: setup: standalone cluster: single platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: &standalonetasks - name: industry_benchmarks - name: ycsb_60GB - name: crud_workloads - name: genny_canaries - name: cursor_manager - name: mixed_workloads - name: misc_workloads - name: map_reduce_workloads - name: smoke_test - name: non_sharded_workloads - name: bestbuy_agg - name: bestbuy_agg_merge_different_db - name: bestbuy_agg_merge_same_db - name: bestbuy_agg_merge_wordcount - name: bestbuy_query - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-standalone-audit display_name: Linux Standalone Audit batchtime: 10080 # 7 days modules: *modules expansions: setup: standalone-audit cluster: single platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: crud_workloads - name: smoke_test - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-1-node-replSet-fle display_name: Linux 1-Node ReplSet FLE batchtime: 1440 # 24 hours modules: *modules expansions: setup: single-replica-fle cluster: single platform: linux project_dir: *project_dir authentication: enabled fle: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: linkbench - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-1-node-replSet-cwrwc display_name: Linux 1-Node ReplSet CWRWC batchtime: 2880 # 48 hours modules: *modules expansions: setup: single-replica-cwrwc cluster: single platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: linkbench - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-1-node-replSet-ese-cbc display_name: Linux 1-Node ReplSet ESE CBC batchtime: 1440 # 24 hours modules: *modules expansions: setup: single-replica-ese-cbc cluster: single platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: smoke_test - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-1-node-replSet-ese-gcm display_name: Linux 1-Node ReplSet ESE GCM batchtime: 1440 # 24 hours modules: *modules expansions: setup: single-replica-ese-gcm cluster: single platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: smoke_test - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-1-node-15gbwtcache display_name: Linux 1-Node ReplSet 15 GB WiredTiger Cache batchtime: 1440 # 24 hours modules: *modules expansions: setup: single-replica-15gbwtcache cluster: replica platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-single" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: smoke_test - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-3-shard display_name: Linux 3-Shard Cluster batchtime: 10080 # 7 days modules: *modules expansions: setup: shard cluster: shard platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-shard" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: crud_workloads - name: mixed_workloads - name: misc_workloads - name: map_reduce_workloads - name: smoke_test - name: industry_benchmarks_wmajority - name: mongos_workloads - name: mongos_large_catalog_workloads - name: change_streams_throughput - name: change_streams_latency - name: change_streams_multi_mongos - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-shard-lite display_name: Linux Shard Lite Cluster batchtime: 2880 # 48 hours modules: *modules expansions: setup: shard-lite cluster: shard-lite platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-shard-lite" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: &shardlitetasks - name: bestbuy_agg - name: bestbuy_agg_merge_different_db - name: bestbuy_agg_merge_same_db - name: bestbuy_agg_merge_target_hashed - name: bestbuy_agg_merge_wordcount - name: bestbuy_query - name: change_streams_latency - name: change_streams_throughput - name: industry_benchmarks - name: industry_benchmarks_wmajority - name: linkbench - name: mixed_workloads - name: mongos_workloads - name: mongos_large_catalog_workloads - name: move_chunk_workloads - name: move_chunk_waiting_workloads - name: retryable_writes_workloads - name: smoke_test - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-shard-lite-cwrwc display_name: Linux Shard Lite Cluster CWRWC batchtime: 10080 # 7 days modules: *modules expansions: setup: shard-lite-cwrwc cluster: shard-lite platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-shard-lite" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: linkbench - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-3-node-replSet display_name: Linux 3-Node ReplSet batchtime: 1440 # 24 hours modules: *modules expansions: setup: replica cluster: replica platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-replset" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: &3nodetasks - name: industry_benchmarks - name: ycsb_60GB - name: industry_benchmarks_secondary_reads - name: crud_workloads - name: crud_workloads_majority - name: mixed_workloads - name: misc_workloads - name: map_reduce_workloads - name: refine_shard_key_transaction_stress - name: smoke_test - name: retryable_writes_workloads - name: industry_benchmarks_wmajority - name: secondary_performance # Uses a special 2 node mongodb setup - name: non_sharded_workloads - name: bestbuy_agg - name: bestbuy_agg_merge_different_db - name: bestbuy_agg_merge_same_db - name: bestbuy_agg_merge_wordcount - name: bestbuy_query - name: change_streams_throughput - name: change_streams_latency - name: snapshot_reads - name: secondary_reads - name: tpcc - name: linkbench - name: linkbench2 - name: genny_patch_tasks - name: genny_auto_tasks - name: sb_large_scale - name: sb_timeseries - name: linux-3-node-replSet-noflowcontrol display_name: Linux 3-Node ReplSet (Flow Control off) batchtime: 10080 # 7 days modules: *modules expansions: setup: replica-noflowcontrol cluster: replica platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-replset" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: industry_benchmarks_secondary_reads - name: crud_workloads - name: crud_workloads_majority - name: mixed_workloads - name: smoke_test - name: industry_benchmarks_wmajority - name: change_streams_throughput - name: change_streams_latency - name: tpcc - name: linkbench - name: linkbench2 - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-3-node-replSet-ssl display_name: Linux 3-Node ReplSet (SSL) batchtime: 10080 # 7 days modules: *modules expansions: setup: replica-ssl cluster: replica platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger run_on: - "rhel70-perf-replset" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: industry_benchmarks - name: mixed_workloads - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-3-node-replSet-initialsync display_name: Linux 3-Node ReplSet Initial Sync batchtime: 1440 # 24 hours modules: *modules expansions: setup: replica-2node cluster: replica platform: linux authentication: disabled storageEngine: wiredTiger project_dir: *project_dir depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 run_on: - "rhel70-perf-replset" tasks: - name: initialsync - name: initialsync-logkeeper-short - name: initialsync-large - name: genny_patch_tasks - name: genny_auto_tasks - name: linux-replSet-initialsync-logkeeper display_name: Linux ReplSet Initial Sync LogKeeper batchtime: 10080 # 7 days modules: *modules expansions: setup: initialsync-logkeeper cluster: initialsync-logkeeper # EBS logkeeper snapshot with FCV set to 4.4 snapshotId: snap-0b0b8b2ad16e8f14e platform: linux authentication: disabled storageEngine: wiredTiger project_dir: *project_dir run_on: - "rhel70-perf-initialsync-logkeeper" depends_on: - name: compile variant: compile-amazon2 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: - name: initialsync-logkeeper - name: genny_patch_tasks - name: genny_auto_tasks ####################################### # Atlas Like Buildvariants # ####################################### - name: atlas-like-M60 display_name: M60-Like 3-Node ReplSet batchtime: 2880 # 48 hours modules: *modules expansions: setup: atlas-like-replica cluster: M60-like-replica platform: linux project_dir: *project_dir authentication: enabled storageEngine: wiredTiger compile-variant: -rhel70 run_on: - "rhel70-perf-M60-like" depends_on: - name: compile variant: compile-rhel70 - name: genny_generate_all_tasks variant: compile-amazon2 tasks: # Cannot use *3nodetasks because secondary_performance uses a special mongodb setup. - name: industry_benchmarks - name: ycsb_60GB - name: industry_benchmarks_secondary_reads - name: crud_workloads - name: crud_workloads_majority - name: mixed_workloads - name: misc_workloads - name: map_reduce_workloads - name: smoke_test - name: retryable_writes_workloads - name: industry_benchmarks_wmajority - name: non_sharded_workloads - name: bestbuy_agg - name: bestbuy_agg_merge_different_db - name: bestbuy_agg_merge_same_db - name: bestbuy_agg_merge_wordcount - name: bestbuy_query - name: change_streams_throughput - name: change_streams_latency - name: snapshot_reads - name: secondary_reads # - name: tpcc # TPCC with SSL currently broken https://jira.mongodb.org/browse/TIG-1681 - name: linkbench - name: genny_patch_tasks - name: genny_auto_tasks