summaryrefslogtreecommitdiff
path: root/jstests/replsets/replset7.js
blob: 94dff59b33fbf3ad10a431cd554065efdff0c906 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54

// test for SERVER-5040 - if documents move forward during an initial sync.

var rt = new ReplSetTest({name: "replset7tests", nodes: 1});

var nodes = rt.startSet();
rt.initiate();
var master = rt.getPrimary();

var md = master.getDB('d');
var mdc = md['c'];

// prep the data
var doccount = 5000;
var bulk = mdc.initializeUnorderedBulkOp();
for (i = 0; i < doccount; ++i) {
    bulk.insert({_id: i, x: i});
}
assert.writeOK(bulk.execute());

assert.commandWorked(mdc.ensureIndex({x: 1}, {unique: true}));

// add a secondary
var slave = rt.add();
rt.reInitiate();
print("initiation complete!");
var sc = slave.getDB('d')['c'];
slave.setSlaveOk();

// Wait for slave to start cloning.
// assert.soon( function() { c = sc.find( { _id:1, x:1 } ); print( c ); return c > 0; } );

// Move all documents to the end by growing it
bulk = mdc.initializeUnorderedBulkOp();
var bigStr = "ayayayayayayayayayayayayayayayayayayayayayayayayayayayayayayayayay" +
    "ayayayayayayayayayayayay";
for (i = 0; i < doccount; ++i) {
    bulk.find({_id: i, x: i}).remove();
    bulk.insert({_id: doccount + i, x: i, bigstring: bigStr});
}
assert.writeOK(bulk.execute());

// Wait for replication to catch up.
rt.awaitSecondaryNodes();

// Do we have an index?
assert.eq(1,
          slave.getDB('d')['c']
              .getIndexes()
              .filter(function(doc) {
                  return (doc.v === 1 && JSON.stringify(doc.key) === JSON.stringify({x: 1}) &&
                          doc.ns === 'd.c' && doc.name === 'x_1');
              })
              .length);