summaryrefslogtreecommitdiff
path: root/jstests/replsets/initial_sync_cloner_dups.js
blob: 23b1b989400bb5638925770a79488c97d796f5a8 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
/**
 * Test for SERVER-17487
 * 3 node replset
 * insert docs with numeric _ids
 * start deleting/re-inserting docs from collection in a loop
 * add new secondary to force initialSync
 * verify collection and both indexes on the secondary have the right number of docs
 */
(function(doNotRun) {
    "use strict";

    if (doNotRun) {
        return;
    }

    load('jstests/libs/parallelTester.js');

    Random.setRandomSeed();

    // used to parse RAM log file
    var contains = function(logLines, func) {
        var i = logLines.length;
        while (i--) {
            printjson(logLines[i]);
            if (func(logLines[i])) {
                return true;
            }
        }
        return false;
    };

    var replTest = new ReplSetTest({name: 'cloner', nodes: 3, oplogSize: 150 /*~1.5x data size*/});
    replTest.startSet();
    var conf = replTest.getReplSetConfig();
    conf.settings = {};
    conf.settings.chainingAllowed = false;
    replTest.initiate(conf);
    replTest.awaitSecondaryNodes();
    var primary = replTest.getPrimary();
    var coll = primary.getDB('test').cloner;
    coll.drop();
    coll.createIndex({k: 1});

    // These need to be big enough to force initial-sync to use many batches
    var numDocs = 100 * 1000;
    var bigStr = Array(1001).toString();
    var batch = coll.initializeUnorderedBulkOp();
    for (var i = 0; i < numDocs; i++) {
        batch.insert({_id: i, bigStr: bigStr});
    }
    batch.execute();

    replTest.awaitReplication();

    jsTestLog("Start remove/insert on primary");
    var insertAndRemove = function(host) {
        jsTestLog("starting bg writes on " + host);
        var m = new Mongo(host);
        var db = m.getDB('test');
        var coll = db.cloner;
        var numDocs = coll.count();
        for (var i = 0; !db.stop.findOne(); i++) {
            var id = Random.randInt(numDocs);
            coll.remove({_id: id});
            coll.insert({_id: id});

            var id = i % numDocs;
            // print(id);
            coll.remove({_id: id});
            coll.insert({_id: id});

            // Try to throttle this thread to prevent overloading slow machines.
            sleep(1);
        }

        jsTestLog("finished bg writes on " + host);
    };
    var worker = new ScopedThread(insertAndRemove, primary.host);
    worker.start();

    jsTestLog("add a new secondary");
    var secondary = replTest.add({});
    replTest.reInitiate();
    secondary.setSlaveOk();
    // Wait for the secondary to get ReplSetInitiate command.
    replTest.waitForState(
        secondary,
        [ReplSetTest.State.STARTUP_2, ReplSetTest.State.RECOVERING, ReplSetTest.State.SECONDARY]);

    // This fail point will cause the first intial sync to fail, and leave an op in the buffer to
    // verify the fix from SERVER-17807
    print("=================== failpoint enabled ==============");
    printjson(assert.commandWorked(secondary.getDB("admin").adminCommand(
        {configureFailPoint: 'failInitSyncWithBufferedEntriesLeft', mode: {times: 1}})));
    printjson(assert.commandWorked(secondary.getDB("admin").adminCommand({resync: true})));

    // NOTE: This is here to prevent false negatives, but it is racy and dependent on magic numbers.
    // Removed the assertion because it was too flaky.  Printing a warning instead (dan)
    jsTestLog("making sure we dropped some dups");
    var res = secondary.adminCommand({getLog: "global"});
    var droppedDups = (contains(res.log, function(v) {
        return v.indexOf("index build dropped" /* NNN dups*/) != -1;
    }));
    if (!droppedDups) {
        jsTestLog(
            "Warning: Test did not trigger duplicate documents, this run will be a false negative");
    }

    jsTestLog("stopping writes and waiting for replica set to coalesce");
    primary.getDB('test').stop.insert({});
    worker.join();
    // make sure all secondaries are caught up, after init sync
    reconnect(secondary.getDB("test"));
    replTest.awaitSecondaryNodes();
    replTest.awaitReplication();

    jsTestLog("check that secondary has correct counts");
    var secondaryColl = secondary.getDB('test').getCollection('cloner');
    var index = secondaryColl.find({}, {_id: 1}).hint({_id: 1}).itcount();
    var secondary_index = secondaryColl.find({}, {_id: 1}).hint({k: 1}).itcount();
    var table = secondaryColl.find({}, {_id: 1}).hint({$natural: 1}).itcount();
    if (index != table || index != secondary_index) {
        printjson({
            name: coll,
            _id_index_count: index,
            secondary_index_count: secondary_index,
            table_count: table
        });
    }
    assert.eq(index, table);
    assert.eq(table, secondary_index);
})(true /* Disabled until SERVER-23476 re-enabled rsync command */);