diff options
-rw-r--r-- | jstests/sharding/resharding_large_number_of_initial_chunks.js | 18 |
1 files changed, 11 insertions, 7 deletions
diff --git a/jstests/sharding/resharding_large_number_of_initial_chunks.js b/jstests/sharding/resharding_large_number_of_initial_chunks.js index accb9f2a94a..798131c0fb8 100644 --- a/jstests/sharding/resharding_large_number_of_initial_chunks.js +++ b/jstests/sharding/resharding_large_number_of_initial_chunks.js @@ -13,11 +13,17 @@ load("jstests/sharding/libs/find_chunks_util.js"); -// The total size of the zones and chunks documents that the config server writes in a single -// replica set transaction totals to around 60 MB. WWhen combined with the other operations and -// transactions occurring in the config server, this large transaction causes WiredTiger to run out -// of dirty cache space. Hence, we need to increase the wiredTigerCacheSizeGB to 1 GB. -const st = new ShardingTest({mongos: 1, shards: 2, config: 1}); +const st = new ShardingTest({ + mongos: 1, + shards: 2, + config: 1, + other: { + configOptions: { + setParameter: + {reshardingCriticalSectionTimeoutMillis: 24 * 60 * 60 * 1000 /* 1 day */} + } + } +}); const kDbName = 'db'; const collName = 'foo'; @@ -27,8 +33,6 @@ const mongos = st.s; assert.commandWorked(mongos.adminCommand({enableSharding: kDbName})); assert.commandWorked(mongos.adminCommand({shardCollection: ns, key: {oldKey: 1}})); -// In debug builds, the resharding operation takes significantly longer when there's a large number -// of chunks and zones. We reduce nZones so the test completes in a reasonable amount of time. let nZones = 10000; let zones = []; let shard0Zones = []; |