summaryrefslogtreecommitdiff
path: root/jstests/sharding/sync_cluster_config/sync6.js
blob: c6b9529b46d3fdf2ea1e630eca2d757c3371e6c4 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
// Test that distributed lock forcing does not result in inconsistencies, using a
// fast timeout. See comments about _testDistLockWithSkew in distlock_test.cpp for
// more details.

// Note that this test will always have random factors, since we can't control the
// thread scheduling.

// NOTE: this test is skipped when running smoke.py with --auth or --keyFile to force authentication
// in all tests.
var test = new SyncCCTest("sync6");

// Startup another process to handle our commands to the cluster, mostly so it's
// easier to read.
var commandConn = MongoRunner.runMongod({});

// Up the log level for this test
commandConn.getDB("admin").runCommand({setParameter: 1, logLevel: 1});

// Have lots of threads, so use larger i
// Can't test too many, we get socket exceptions... possibly due to the
// javascript console.
for (var i = 8; i < 9; i++) {
    // Our force time is 8 seconds
    // Slower machines can't keep up the LockPinger rate, which can lead to lock failures
    // since our locks are only valid if the LockPinger pings faster than the force time.
    // Actual lock timeout is 15 minutes, so a few seconds is extremely aggressive
    var takeoverMS = 8000;

    // Generate valid sleep and skew for this timeout
    var threadSleepWithLock = takeoverMS / 2;
    var configServerTimeSkew = [0, 0, 0];
    for (var h = 0; h < 3; h++) {
        // Skew by 1/30th the takeover time either way, at max
        configServerTimeSkew[h] = (i + h) % Math.floor(takeoverMS / 60);
        // Make skew pos or neg
        configServerTimeSkew[h] *= ((i + h) % 2) ? -1 : 1;
    }

    // Build command
    var command = {
        _testDistLockWithSkew: 1
    };

    // Basic test parameters
    command["lockName"] = "TimeSkewFailNewTest_lock_" + i;
    command["host"] = test.url;
    command["seed"] = i;
    command["numThreads"] = (i % 50) + 1;

    // Critical values so we're sure of correct operation
    command["takeoverMS"] = takeoverMS;
    command["wait"] = 4 * takeoverMS;  // so we must force the lock
    command["skewHosts"] = configServerTimeSkew;
    command["threadWait"] = threadSleepWithLock;

    // Less critical test params

    // 1/3 of threads will not release the lock
    command["hangThreads"] = 3;
    // Amount of time to wait before trying lock again
    command["threadSleep"] = 1;  // ( ( i + 1 ) * 100 ) % (takeoverMS / 4)
    // Amount of total clock skew possible between locking threads (processes)
    // This can be large now.
    command["skewRange"] = (command["takeoverMS"] * 3) * 60 * 1000;

    // Double-check our sleep, host skew, and takeoverMS values again

    // At maximum, our threads must sleep only half the lock timeout time.
    assert(command["threadWait"] <= command["takeoverMS"] / 2);
    for (var h = 0; h < command["skewHosts"].length; h++) {
        // At maximum, our config server time skew needs to be less than 1/30th
        // the total time skew (1/60th either way).
        assert(Math.abs(command["skewHosts"][h]) <= (command["takeoverMS"] / 60));
    }

    var result = commandConn.getDB("admin").runCommand(command);
    printjson(result);
    printjson(command);
    assert(result.ok, "Skewed threads did not increment correctly.");
}

MongoRunner.stopMongod(commandConn);
test.stop();