summaryrefslogtreecommitdiff
path: root/src/mongo/dbtests/validate_tests.cpp
diff options
context:
space:
mode:
authorMaria van Keulen <maria@mongodb.com>2017-03-07 12:00:08 -0500
committerMaria van Keulen <maria@mongodb.com>2017-03-07 12:00:08 -0500
commit589a5c169ced8f6e9ddcd3d182ae1b75db6b7d79 (patch)
treec7a090ffdd56a91ae677e2492c61b820af44f964 /src/mongo/dbtests/validate_tests.cpp
parent3cba97198638df3750e3b455e2ad57af7ee536ae (diff)
downloadmongo-589a5c169ced8f6e9ddcd3d182ae1b75db6b7d79.tar.gz
SERVER-27938 Rename all OperationContext variables to opCtx
This commit is an automated rename of all whole word instances of txn, _txn, and txnPtr to opCtx, _opCtx, and opCtxPtr, respectively in all .cpp and .h files in src/mongo.
Diffstat (limited to 'src/mongo/dbtests/validate_tests.cpp')
-rw-r--r--src/mongo/dbtests/validate_tests.cpp272
1 files changed, 146 insertions, 126 deletions
diff --git a/src/mongo/dbtests/validate_tests.cpp b/src/mongo/dbtests/validate_tests.cpp
index 3eb1706fbce..9b949ec3b93 100644
--- a/src/mongo/dbtests/validate_tests.cpp
+++ b/src/mongo/dbtests/validate_tests.cpp
@@ -59,7 +59,7 @@ static const char* const _ns = "unittests.validate_tests";
*/
class ValidateBase {
public:
- explicit ValidateBase(bool full) : _ctx(&_txn, _ns), _client(&_txn), _full(full) {
+ explicit ValidateBase(bool full) : _ctx(&_opCtx, _ns), _client(&_opCtx), _full(full) {
_client.createCollection(_ns);
}
~ValidateBase() {
@@ -75,7 +75,7 @@ protected:
ValidateResults results;
BSONObjBuilder output;
ASSERT_OK(collection()->validate(
- &_txn, _full ? kValidateFull : kValidateIndex, &results, &output));
+ &_opCtx, _full ? kValidateFull : kValidateIndex, &results, &output));
// Check if errors are reported if and only if valid is set to false.
ASSERT_EQ(results.valid, results.errors.empty());
@@ -94,7 +94,7 @@ protected:
}
const ServiceContext::UniqueOperationContext _txnPtr = cc().makeOperationContext();
- OperationContext& _txn = *_txnPtr;
+ OperationContext& _opCtx = *_txnPtr;
OldClientWriteContext _ctx;
DBDirectClient _client;
bool _full;
@@ -112,13 +112,13 @@ public:
RecordId id1;
{
OpDebug* const nullOpDebug = nullptr;
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2), nullOpDebug, true));
+ ASSERT_OK(coll->insertDocument(&_opCtx, BSON("_id" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(coll->insertDocument(&_opCtx, BSON("_id" << 2), nullOpDebug, true));
wunit.commit();
}
@@ -128,8 +128,8 @@ public:
// Remove {_id: 1} from the record store, so we get more _id entries than records.
{
- WriteUnitOfWork wunit(&_txn);
- rs->deleteRecord(&_txn, id1);
+ WriteUnitOfWork wunit(&_opCtx);
+ rs->deleteRecord(&_opCtx, id1);
wunit.commit();
}
@@ -138,11 +138,11 @@ public:
// Insert records {_id: 0} and {_id: 1} , so we get too few _id entries, and verify
// validate fails.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
for (int j = 0; j < 2; j++) {
auto doc = BSON("_id" << j);
- ASSERT_OK(
- rs->insertRecord(&_txn, doc.objdata(), doc.objsize(), /*enforceQuota*/ false));
+ ASSERT_OK(rs->insertRecord(
+ &_opCtx, doc.objdata(), doc.objsize(), /*enforceQuota*/ false));
}
wunit.commit();
}
@@ -162,16 +162,18 @@ public:
RecordId id1;
{
OpDebug* const nullOpDebug = nullptr;
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
wunit.commit();
}
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "a"
@@ -191,8 +193,8 @@ public:
// Remove a record, so we get more _id entries than records, and verify validate fails.
{
- WriteUnitOfWork wunit(&_txn);
- rs->deleteRecord(&_txn, id1);
+ WriteUnitOfWork wunit(&_opCtx);
+ rs->deleteRecord(&_opCtx, id1);
wunit.commit();
}
@@ -201,11 +203,11 @@ public:
// Insert two more records, so we get too few entries for a non-sparse index, and
// verify validate fails.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
for (int j = 0; j < 2; j++) {
auto doc = BSON("_id" << j);
- ASSERT_OK(
- rs->insertRecord(&_txn, doc.objdata(), doc.objsize(), /*enforceQuota*/ false));
+ ASSERT_OK(rs->insertRecord(
+ &_opCtx, doc.objdata(), doc.objsize(), /*enforceQuota*/ false));
}
wunit.commit();
}
@@ -224,17 +226,20 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 3 << "b" << 3), nullOpDebug, true));
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 3 << "b" << 3), nullOpDebug, true));
wunit.commit();
}
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "a"
@@ -255,10 +260,10 @@ public:
// Update {a: 1} to {a: 9} without updating the index, so we get inconsistent values
// between the index and the document. Verify validate fails.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1 << "a" << 9);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
@@ -279,13 +284,13 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2), nullOpDebug, true));
+ ASSERT_OK(coll->insertDocument(&_opCtx, BSON("_id" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(coll->insertDocument(&_opCtx, BSON("_id" << 2), nullOpDebug, true));
wunit.commit();
}
@@ -296,10 +301,10 @@ public:
// Update {_id: 1} to {_id: 9} without updating the index, so we get inconsistent values
// between the index and the document. Verify validate fails.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 9);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -308,10 +313,10 @@ public:
// Revert {_id: 9} to {_id: 1} and verify that validate succeeds.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -322,11 +327,12 @@ public:
// will still be the same number of index entries and documents, but one document will not
// have an index entry.
{
- WriteUnitOfWork wunit(&_txn);
- rs->deleteRecord(&_txn, id1);
+ WriteUnitOfWork wunit(&_opCtx);
+ rs->deleteRecord(&_opCtx, id1);
auto doc = BSON("_id" << 3);
- ASSERT_OK(rs->insertRecord(&_txn, doc.objdata(), doc.objsize(), /*enforceQuota*/ false)
- .getStatus());
+ ASSERT_OK(
+ rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), /*enforceQuota*/ false)
+ .getStatus());
wunit.commit();
}
@@ -354,22 +360,22 @@ public:
// {a: [c: 1]}
auto doc3 = BSON("_id" << 3 << "a" << BSON_ARRAY(BSON("c" << 1)));
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
- ASSERT_OK(coll->insertDocument(&_txn, doc1, nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, doc2, nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, doc3, nullOpDebug, true));
+ ASSERT_OK(coll->insertDocument(&_opCtx, doc1, nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(coll->insertDocument(&_opCtx, doc2, nullOpDebug, true));
+ ASSERT_OK(coll->insertDocument(&_opCtx, doc3, nullOpDebug, true));
wunit.commit();
}
ASSERT_TRUE(checkValid());
// Create multi-key index.
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "multikey_index"
@@ -389,9 +395,9 @@ public:
// Update a document's indexed field without updating the index.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc1_b.objdata(), doc1_b.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc1_b.objdata(), doc1_b.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -401,9 +407,9 @@ public:
// Update a document's non-indexed field without updating the index.
// Index validation should still be valid.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc1_c.objdata(), doc1_c.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc1_c.objdata(), doc1_c.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -423,19 +429,22 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
-
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 3 << "b" << 1), nullOpDebug, true));
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
+
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 3 << "b" << 1), nullOpDebug, true));
wunit.commit();
}
// Create a sparse index.
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "sparse_index"
@@ -457,10 +466,10 @@ public:
// Update a document's indexed field without updating the index.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 2 << "a" << 3);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -480,22 +489,24 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
-
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
+
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
// Explicitly test that multi-key partial indexes containing documents that
// don't match the filter expression are handled correctly.
ASSERT_OK(coll->insertDocument(
- &_txn, BSON("_id" << 3 << "a" << BSON_ARRAY(-1 << -2 << -3)), nullOpDebug, true));
+ &_opCtx, BSON("_id" << 3 << "a" << BSON_ARRAY(-1 << -2 << -3)), nullOpDebug, true));
wunit.commit();
}
// Create a partial index.
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "partial_index"
@@ -517,10 +528,10 @@ public:
// Update an unindexed document without updating the index.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -541,16 +552,16 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
ASSERT_OK(coll->insertDocument(
- &_txn, BSON("_id" << 1 << "x" << 1 << "a" << 2), nullOpDebug, true));
+ &_opCtx, BSON("_id" << 1 << "x" << 1 << "a" << 2), nullOpDebug, true));
wunit.commit();
}
// Create a partial geo index that indexes the document. This should throw an error.
- ASSERT_THROWS(dbtests::createIndexFromSpec(&_txn,
+ ASSERT_THROWS(dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "partial_index"
@@ -568,7 +579,7 @@ public:
UserException);
// Create a partial geo index that does not index the document.
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "partial_index"
@@ -599,24 +610,27 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
ASSERT_OK(coll->insertDocument(
- &_txn, BSON("_id" << 1 << "a" << 1 << "b" << 4), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
+ &_opCtx, BSON("_id" << 1 << "a" << 1 << "b" << 4), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
ASSERT_OK(coll->insertDocument(
- &_txn, BSON("_id" << 2 << "a" << 2 << "b" << 5), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 3 << "a" << 3), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 4 << "b" << 6), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 5 << "c" << 7), nullOpDebug, true));
+ &_opCtx, BSON("_id" << 2 << "a" << 2 << "b" << 5), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 3 << "a" << 3), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 4 << "b" << 6), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 5 << "c" << 7), nullOpDebug, true));
wunit.commit();
}
// Create two compound indexes, one forward and one reverse, to test
// validate()'s index direction parsing.
- auto status = dbtests::createIndexFromSpec(&_txn,
+ auto status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "compound_index_1"
@@ -630,7 +644,7 @@ public:
<< false));
ASSERT_OK(status);
- status = dbtests::createIndexFromSpec(&_txn,
+ status = dbtests::createIndexFromSpec(&_opCtx,
coll->ns().ns(),
BSON("name"
<< "compound_index_2"
@@ -650,10 +664,10 @@ public:
// Update a document's indexed field without updating the index.
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1 << "a" << 1 << "b" << 3);
auto updateStatus = rs->updateRecord(
- &_txn, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -673,20 +687,23 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
-
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 3 << "b" << 1), nullOpDebug, true));
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
+
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 3 << "b" << 1), nullOpDebug, true));
wunit.commit();
}
const std::string indexName = "bad_index";
auto status = dbtests::createIndexFromSpec(
- &_txn,
+ &_opCtx,
coll->ns().ns(),
BSON("name" << indexName << "ns" << coll->ns().ns() << "key" << BSON("a" << 1) << "v"
<< static_cast<int>(kIndexVersion)
@@ -698,11 +715,11 @@ public:
// Replace a correct index entry with a bad one and check it's invalid.
IndexCatalog* indexCatalog = coll->getIndexCatalog();
- IndexDescriptor* descriptor = indexCatalog->findIndexByName(&_txn, indexName);
+ IndexDescriptor* descriptor = indexCatalog->findIndexByName(&_opCtx, indexName);
IndexAccessMethod* iam = indexCatalog->getIndex(descriptor);
{
- WriteUnitOfWork wunit(&_txn);
+ WriteUnitOfWork wunit(&_opCtx);
int64_t numDeleted;
int64_t numInserted;
const BSONObj actualKey = BSON("a" << 1);
@@ -710,8 +727,8 @@ public:
InsertDeleteOptions options;
options.dupsAllowed = true;
options.logIfError = true;
- auto removeStatus = iam->remove(&_txn, actualKey, id1, options, &numDeleted);
- auto insertStatus = iam->insert(&_txn, badKey, id1, options, &numInserted);
+ auto removeStatus = iam->remove(&_opCtx, actualKey, id1, options, &numDeleted);
+ auto insertStatus = iam->insert(&_opCtx, badKey, id1, options, &numInserted);
ASSERT_EQUALS(numDeleted, 1);
ASSERT_EQUALS(numInserted, 1);
@@ -735,20 +752,23 @@ public:
Collection* coll;
RecordId id1;
{
- WriteUnitOfWork wunit(&_txn);
- ASSERT_OK(db->dropCollection(&_txn, _ns));
- coll = db->createCollection(&_txn, _ns);
-
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
- id1 = coll->getCursor(&_txn)->next()->id;
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
- ASSERT_OK(coll->insertDocument(&_txn, BSON("_id" << 3 << "b" << 1), nullOpDebug, true));
+ WriteUnitOfWork wunit(&_opCtx);
+ ASSERT_OK(db->dropCollection(&_opCtx, _ns));
+ coll = db->createCollection(&_opCtx, _ns);
+
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 1 << "a" << 1), nullOpDebug, true));
+ id1 = coll->getCursor(&_opCtx)->next()->id;
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 2 << "a" << 2), nullOpDebug, true));
+ ASSERT_OK(
+ coll->insertDocument(&_opCtx, BSON("_id" << 3 << "b" << 1), nullOpDebug, true));
wunit.commit();
}
const std::string indexName = "bad_index";
auto status = dbtests::createIndexFromSpec(
- &_txn,
+ &_opCtx,
coll->ns().ns(),
BSON("name" << indexName << "ns" << coll->ns().ns() << "key" << BSON("a" << 1) << "v"
<< static_cast<int>(kIndexVersion)
@@ -761,7 +781,7 @@ public:
// Change the IndexDescriptor's keyPattern to descending so the index ordering
// appears wrong.
IndexCatalog* indexCatalog = coll->getIndexCatalog();
- IndexDescriptor* descriptor = indexCatalog->findIndexByName(&_txn, indexName);
+ IndexDescriptor* descriptor = indexCatalog->findIndexByName(&_opCtx, indexName);
descriptor->setKeyPatternForTest(BSON("a" << -1));
ASSERT_FALSE(checkValid());