1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
|
/**
* Initial sync runs in several phases - the first 3 are as follows:
* 1) fetches the last oplog entry (op_start1) on the source;
* 2) copies all non-local databases from the source; and
* 3) fetches and applies operations from the source after op_start1.
*
* This test updates and deletes a document on the source between phases 1 and 2 in a prepared
* transaction. The secondary will initially fail to apply the update operation in phase 3 and
* subsequently have to attempt to check the source for a new copy of the document. Before the
* secondary checks the source, we insert a new copy of the document on the source so that the
* secondary can fetch it.
*
* @tags: [uses_transactions, uses_prepare_transaction]
*/
(function() {
load("jstests/core/txns/libs/prepare_helpers.js");
load("jstests/replsets/libs/initial_sync_update_missing_doc.js");
load("jstests/libs/check_log.js");
function doTest(doTransactionWork, numDocuments) {
const name = 'initial_sync_update_missing_doc_with_prepare';
const replSet = new ReplSetTest({
name: name,
nodes: 1,
});
replSet.startSet();
replSet.initiate();
const primary = replSet.getPrimary();
const dbName = 'test';
const coll = primary.getDB(dbName).getCollection(name);
assert.commandWorked(coll.insert({_id: 0, x: 1}));
// Add a secondary node with priority: 0 and votes: 0 so that we prevent elections while
// it is syncing from the primary.
const secondaryConfig = {rsConfig: {votes: 0, priority: 0}};
const secondary = reInitiateSetWithSecondary(replSet, secondaryConfig);
const session = primary.startSession();
const sessionDB = session.getDatabase(dbName);
const sessionColl = sessionDB.getCollection(name);
session.startTransaction();
doTransactionWork(sessionColl, {_id: 0});
const prepareTimestamp = PrepareHelpers.prepareTransaction(session);
assert.commandWorked(PrepareHelpers.commitTransaction(session, prepareTimestamp));
turnOffHangBeforeCopyingDatabasesFailPoint(secondary);
// Re-insert deleted document on the sync source. The secondary should be able to fetch and
// insert this document after failing to apply the udpate.
assert.commandWorked(coll.insert({_id: 0, x: 3}));
const res = assert.commandWorked(secondary.adminCommand({replSetGetStatus: 1}));
assert.eq(res.initialSyncStatus.fetchedMissingDocs, 0);
const firstOplogEnd = res.initialSyncStatus.initialSyncOplogEnd;
// Temporarily increase log levels so that we can see the 'Inserted missing document' log
// line.
secondary.getDB('test').setLogLevel(1, 'replication');
turnOffHangBeforeGettingMissingDocFailPoint(primary, secondary, name, 1 /* numInserted */);
secondary.getDB('test').setLogLevel(0, 'replication');
finishAndValidate(replSet, name, firstOplogEnd, 1 /* numInserted */, numDocuments);
assert.docEq(
{_id: 0, x: 3}, coll.findOne({_id: 0}), 'document on secondary matches primary');
replSet.stopSet();
}
jsTestLog("Testing with prepared transaction");
// Passing in a function to update and remove document on primary in a prepared transaction
// between phrase 1 and 2. Once the secondary receives the commit for the transaction, the
// secondary should apply each operation separately (one update, and one delete) during initial
// sync.
doTest(updateRemove, 1 /* numDocuments after initial sync */);
jsTestLog("Testing with large prepared transaction");
// Passing in a function to insert, update and remove large documents on primary in a large
// prepared transaction. Once the secondary receives the commit for the transaction, the
// secondary should apply each operation separately (one insert, one update, and one delete)
// during initial sync.
doTest(insertUpdateRemoveLarge, 2 /* numDocuments after initial sync */);
})();
|