summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--jstests/sharding/resharding_large_number_of_initial_chunks.js18
1 files changed, 11 insertions, 7 deletions
diff --git a/jstests/sharding/resharding_large_number_of_initial_chunks.js b/jstests/sharding/resharding_large_number_of_initial_chunks.js
index accb9f2a94a..798131c0fb8 100644
--- a/jstests/sharding/resharding_large_number_of_initial_chunks.js
+++ b/jstests/sharding/resharding_large_number_of_initial_chunks.js
@@ -13,11 +13,17 @@
load("jstests/sharding/libs/find_chunks_util.js");
-// The total size of the zones and chunks documents that the config server writes in a single
-// replica set transaction totals to around 60 MB. WWhen combined with the other operations and
-// transactions occurring in the config server, this large transaction causes WiredTiger to run out
-// of dirty cache space. Hence, we need to increase the wiredTigerCacheSizeGB to 1 GB.
-const st = new ShardingTest({mongos: 1, shards: 2, config: 1});
+const st = new ShardingTest({
+ mongos: 1,
+ shards: 2,
+ config: 1,
+ other: {
+ configOptions: {
+ setParameter:
+ {reshardingCriticalSectionTimeoutMillis: 24 * 60 * 60 * 1000 /* 1 day */}
+ }
+ }
+});
const kDbName = 'db';
const collName = 'foo';
@@ -27,8 +33,6 @@ const mongos = st.s;
assert.commandWorked(mongos.adminCommand({enableSharding: kDbName}));
assert.commandWorked(mongos.adminCommand({shardCollection: ns, key: {oldKey: 1}}));
-// In debug builds, the resharding operation takes significantly longer when there's a large number
-// of chunks and zones. We reduce nZones so the test completes in a reasonable amount of time.
let nZones = 10000;
let zones = [];
let shard0Zones = [];