summaryrefslogtreecommitdiff
path: root/src/mongo
diff options
context:
space:
mode:
authorDianna Hohensee <dianna.hohensee@10gen.com>2018-07-10 16:01:21 -0400
committerDianna Hohensee <dianna.hohensee@10gen.com>2018-07-13 19:01:45 -0400
commit93accf1620db95d4aa7edbd5bbc017264b28e0f2 (patch)
treefab09fb1f5618338bc0c6a9232d11936e972ed56 /src/mongo
parent77ab40c05f28c6c0afbc84bb72375a9867f79ecb (diff)
downloadmongo-93accf1620db95d4aa7edbd5bbc017264b28e0f2.tar.gz
SERVER-35626 remove old MMAP 'enforceQuota' code
Diffstat (limited to 'src/mongo')
-rw-r--r--src/mongo/db/catalog/collection.h21
-rw-r--r--src/mongo/db/catalog/collection_impl.cpp35
-rw-r--r--src/mongo/db/catalog/collection_impl.h13
-rw-r--r--src/mongo/db/catalog/collection_mock.h6
-rw-r--r--src/mongo/db/catalog/rename_collection_test.cpp3
-rw-r--r--src/mongo/db/commands/mr.cpp2
-rw-r--r--src/mongo/db/exec/update.cpp3
-rw-r--r--src/mongo/db/ops/write_ops_exec.cpp3
-rw-r--r--src/mongo/db/repl/collection_bulk_loader_impl.cpp4
-rw-r--r--src/mongo/db/repl/mock_repl_coord_server_fixture.cpp1
-rw-r--r--src/mongo/db/s/collection_sharding_state_test.cpp2
-rw-r--r--src/mongo/db/session.cpp3
-rw-r--r--src/mongo/db/storage/devnull/devnull_kv_engine.cpp7
-rw-r--r--src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp11
-rw-r--r--src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h9
-rw-r--r--src/mongo/db/storage/kv/kv_catalog.cpp17
-rw-r--r--src/mongo/db/storage/kv/kv_engine_test_harness.cpp4
-rw-r--r--src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp5
-rw-r--r--src/mongo/db/storage/mobile/mobile_record_store.cpp9
-rw-r--r--src/mongo/db/storage/mobile/mobile_record_store.h4
-rw-r--r--src/mongo/db/storage/record_store.h14
-rw-r--r--src/mongo/db/storage/record_store_test_capped_visibility.cpp2
-rw-r--r--src/mongo/db/storage/record_store_test_datafor.cpp4
-rw-r--r--src/mongo/db/storage/record_store_test_datasize.cpp2
-rw-r--r--src/mongo/db/storage/record_store_test_deleterecord.cpp4
-rw-r--r--src/mongo/db/storage/record_store_test_harness.cpp21
-rw-r--r--src/mongo/db/storage/record_store_test_insertrecord.cpp4
-rw-r--r--src/mongo/db/storage/record_store_test_manyiter.cpp2
-rw-r--r--src/mongo/db/storage/record_store_test_randomiter.cpp5
-rw-r--r--src/mongo/db/storage/record_store_test_recorditer.cpp16
-rw-r--r--src/mongo/db/storage/record_store_test_repairiter.cpp5
-rw-r--r--src/mongo/db/storage/record_store_test_storagesize.cpp2
-rw-r--r--src/mongo/db/storage/record_store_test_touch.cpp4
-rw-r--r--src/mongo/db/storage/record_store_test_truncate.cpp2
-rw-r--r--src/mongo/db/storage/record_store_test_updaterecord.cpp17
-rw-r--r--src/mongo/db/storage/record_store_test_updatewithdamages.cpp8
-rw-r--r--src/mongo/db/storage/record_store_test_validate.h4
-rw-r--r--src/mongo/db/storage/recovery_unit_test_harness.cpp4
-rw-r--r--src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp24
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp4
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp10
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h10
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp43
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp5
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp5
-rw-r--r--src/mongo/db/views/durable_view_catalog.cpp15
-rw-r--r--src/mongo/dbtests/indexupdatetests.cpp4
-rw-r--r--src/mongo/dbtests/multikey_paths_test.cpp36
-rw-r--r--src/mongo/dbtests/query_stage_cached_plan.cpp4
-rw-r--r--src/mongo/dbtests/query_stage_count.cpp4
-rw-r--r--src/mongo/dbtests/query_stage_sort.cpp5
-rw-r--r--src/mongo/dbtests/storage_timestamp_tests.cpp3
-rw-r--r--src/mongo/dbtests/validate_tests.cpp36
53 files changed, 189 insertions, 301 deletions
diff --git a/src/mongo/db/catalog/collection.h b/src/mongo/db/catalog/collection.h
index 018cfb3f51a..2eec1337022 100644
--- a/src/mongo/db/catalog/collection.h
+++ b/src/mongo/db/catalog/collection.h
@@ -228,13 +228,11 @@ public:
std::vector<InsertStatement>::const_iterator begin,
std::vector<InsertStatement>::const_iterator end,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate) = 0;
virtual Status insertDocument(OperationContext* opCtx,
const InsertStatement& doc,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate) = 0;
virtual Status insertDocumentsForOplog(OperationContext* opCtx,
@@ -244,14 +242,12 @@ public:
virtual Status insertDocument(OperationContext* opCtx,
const BSONObj& doc,
- const std::vector<MultiIndexBlock*>& indexBlocks,
- bool enforceQuota) = 0;
+ const std::vector<MultiIndexBlock*>& indexBlocks) = 0;
virtual RecordId updateDocument(OperationContext* opCtx,
const RecordId& oldLocation,
const Snapshotted<BSONObj>& oldDoc,
const BSONObj& newDoc,
- bool enforceQuota,
bool indexesAffected,
OpDebug* opDebug,
OplogUpdateEntryArgs* args) = 0;
@@ -468,9 +464,8 @@ public:
const std::vector<InsertStatement>::const_iterator begin,
const std::vector<InsertStatement>::const_iterator end,
OpDebug* const opDebug,
- const bool enforceQuota,
const bool fromMigrate = false) {
- return this->_impl().insertDocuments(opCtx, begin, end, opDebug, enforceQuota, fromMigrate);
+ return this->_impl().insertDocuments(opCtx, begin, end, opDebug, fromMigrate);
}
/**
@@ -478,14 +473,12 @@ public:
* i.e. will not add an _id field for documents that are missing it
*
* 'opDebug' Optional argument. When not null, will be used to record operation statistics.
- * 'enforceQuota' If false, quotas will be ignored.
*/
inline Status insertDocument(OperationContext* const opCtx,
const InsertStatement& doc,
OpDebug* const opDebug,
- const bool enforceQuota,
const bool fromMigrate = false) {
- return this->_impl().insertDocument(opCtx, doc, opDebug, enforceQuota, fromMigrate);
+ return this->_impl().insertDocument(opCtx, doc, opDebug, fromMigrate);
}
/**
@@ -506,9 +499,8 @@ public:
*/
inline Status insertDocument(OperationContext* const opCtx,
const BSONObj& doc,
- const std::vector<MultiIndexBlock*>& indexBlocks,
- const bool enforceQuota) {
- return this->_impl().insertDocument(opCtx, doc, indexBlocks, enforceQuota);
+ const std::vector<MultiIndexBlock*>& indexBlocks) {
+ return this->_impl().insertDocument(opCtx, doc, indexBlocks);
}
/**
@@ -524,12 +516,11 @@ public:
const RecordId& oldLocation,
const Snapshotted<BSONObj>& oldDoc,
const BSONObj& newDoc,
- const bool enforceQuota,
const bool indexesAffected,
OpDebug* const opDebug,
OplogUpdateEntryArgs* const args) {
return this->_impl().updateDocument(
- opCtx, oldLocation, oldDoc, newDoc, enforceQuota, indexesAffected, opDebug, args);
+ opCtx, oldLocation, oldDoc, newDoc, indexesAffected, opDebug, args);
}
inline bool updateWithDamagesSupported() const {
diff --git a/src/mongo/db/catalog/collection_impl.cpp b/src/mongo/db/catalog/collection_impl.cpp
index 67e689bf569..1b5b0a5ad33 100644
--- a/src/mongo/db/catalog/collection_impl.cpp
+++ b/src/mongo/db/catalog/collection_impl.cpp
@@ -329,7 +329,6 @@ Status CollectionImpl::insertDocuments(OperationContext* opCtx,
const vector<InsertStatement>::const_iterator begin,
const vector<InsertStatement>::const_iterator end,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate) {
MONGO_FAIL_POINT_BLOCK(failCollectionInserts, extraData) {
@@ -363,7 +362,7 @@ Status CollectionImpl::insertDocuments(OperationContext* opCtx,
const SnapshotId sid = opCtx->recoveryUnit()->getSnapshotId();
- Status status = _insertDocuments(opCtx, begin, end, enforceQuota, opDebug);
+ Status status = _insertDocuments(opCtx, begin, end, opDebug);
if (!status.isOK())
return status;
invariant(sid == opCtx->recoveryUnit()->getSnapshotId());
@@ -380,17 +379,15 @@ Status CollectionImpl::insertDocuments(OperationContext* opCtx,
Status CollectionImpl::insertDocument(OperationContext* opCtx,
const InsertStatement& docToInsert,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate) {
vector<InsertStatement> docs;
docs.push_back(docToInsert);
- return insertDocuments(opCtx, docs.begin(), docs.end(), opDebug, enforceQuota, fromMigrate);
+ return insertDocuments(opCtx, docs.begin(), docs.end(), opDebug, fromMigrate);
}
Status CollectionImpl::insertDocument(OperationContext* opCtx,
const BSONObj& doc,
- const std::vector<MultiIndexBlock*>& indexBlocks,
- bool enforceQuota) {
+ const std::vector<MultiIndexBlock*>& indexBlocks) {
MONGO_FAIL_POINT_BLOCK(failCollectionInserts, extraData) {
const BSONObj& data = extraData.getData();
@@ -415,8 +412,8 @@ Status CollectionImpl::insertDocument(OperationContext* opCtx,
// TODO SERVER-30638: using timestamp 0 for these inserts, which are non-oplog so we don't yet
// care about their correct timestamps.
- StatusWith<RecordId> loc = _recordStore->insertRecord(
- opCtx, doc.objdata(), doc.objsize(), Timestamp(), _enforceQuota(enforceQuota));
+ StatusWith<RecordId> loc =
+ _recordStore->insertRecord(opCtx, doc.objdata(), doc.objsize(), Timestamp());
if (!loc.isOK())
return loc.getStatus();
@@ -450,7 +447,6 @@ Status CollectionImpl::insertDocument(OperationContext* opCtx,
Status CollectionImpl::_insertDocuments(OperationContext* opCtx,
const vector<InsertStatement>::const_iterator begin,
const vector<InsertStatement>::const_iterator end,
- bool enforceQuota,
OpDebug* opDebug) {
dassert(opCtx->lockState()->isCollectionLockedForMode(ns().toString(), MODE_IX));
@@ -483,8 +479,7 @@ Status CollectionImpl::_insertDocuments(OperationContext* opCtx,
Timestamp timestamp = Timestamp(it->oplogSlot.opTime.getTimestamp());
timestamps.push_back(timestamp);
}
- Status status =
- _recordStore->insertRecords(opCtx, &records, &timestamps, _enforceQuota(enforceQuota));
+ Status status = _recordStore->insertRecords(opCtx, &records, &timestamps);
if (!status.isOK())
return status;
@@ -584,7 +579,6 @@ RecordId CollectionImpl::updateDocument(OperationContext* opCtx,
const RecordId& oldLocation,
const Snapshotted<BSONObj>& oldDoc,
const BSONObj& newDoc,
- bool enforceQuota,
bool indexesAffected,
OpDebug* opDebug,
OplogUpdateEntryArgs* args) {
@@ -664,8 +658,8 @@ RecordId CollectionImpl::updateDocument(OperationContext* opCtx,
args->preImageDoc = oldDoc.value().getOwned();
- Status updateStatus = _recordStore->updateRecord(
- opCtx, oldLocation, newDoc.objdata(), newDoc.objsize(), _enforceQuota(enforceQuota), this);
+ Status updateStatus =
+ _recordStore->updateRecord(opCtx, oldLocation, newDoc.objdata(), newDoc.objsize(), this);
// Update each index with each respective UpdateTicket.
if (indexesAffected) {
@@ -733,19 +727,6 @@ StatusWith<RecordData> CollectionImpl::updateDocumentWithDamages(
return newRecStatus;
}
-bool CollectionImpl::_enforceQuota(bool userEnforeQuota) const {
- if (!userEnforeQuota)
- return false;
-
- if (_ns.db() == "local")
- return false;
-
- if (_ns.isSpecial())
- return false;
-
- return true;
-}
-
bool CollectionImpl::isCapped() const {
return _cappedNotifier.get();
}
diff --git a/src/mongo/db/catalog/collection_impl.h b/src/mongo/db/catalog/collection_impl.h
index 1530760be64..5944e19ba78 100644
--- a/src/mongo/db/catalog/collection_impl.h
+++ b/src/mongo/db/catalog/collection_impl.h
@@ -163,7 +163,6 @@ public:
std::vector<InsertStatement>::const_iterator begin,
std::vector<InsertStatement>::const_iterator end,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate = false) final;
/**
@@ -171,12 +170,10 @@ public:
* i.e. will not add an _id field for documents that are missing it
*
* 'opDebug' Optional argument. When not null, will be used to record operation statistics.
- * 'enforceQuota' If false, quotas will be ignored.
*/
Status insertDocument(OperationContext* opCtx,
const InsertStatement& doc,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate = false) final;
/**
@@ -195,8 +192,7 @@ public:
*/
Status insertDocument(OperationContext* opCtx,
const BSONObj& doc,
- const std::vector<MultiIndexBlock*>& indexBlocks,
- bool enforceQuota) final;
+ const std::vector<MultiIndexBlock*>& indexBlocks) final;
/**
* Updates the document @ oldLocation with newDoc.
@@ -211,7 +207,6 @@ public:
const RecordId& oldLocation,
const Snapshotted<BSONObj>& oldDoc,
const BSONObj& newDoc,
- bool enforceQuota,
bool indexesAffected,
OpDebug* opDebug,
OplogUpdateEntryArgs* args) final;
@@ -392,17 +387,13 @@ private:
* - some user error checks
* - adjust padding
*/
- Status _insertDocument(OperationContext* opCtx, const BSONObj& doc, bool enforceQuota);
+ Status _insertDocument(OperationContext* opCtx, const BSONObj& doc);
Status _insertDocuments(OperationContext* opCtx,
std::vector<InsertStatement>::const_iterator begin,
std::vector<InsertStatement>::const_iterator end,
- bool enforceQuota,
OpDebug* opDebug);
-
- bool _enforceQuota(bool userEnforeQuota) const;
-
int _magic;
const NamespaceString _ns;
diff --git a/src/mongo/db/catalog/collection_mock.h b/src/mongo/db/catalog/collection_mock.h
index ae4a816f63f..ba0034df7ff 100644
--- a/src/mongo/db/catalog/collection_mock.h
+++ b/src/mongo/db/catalog/collection_mock.h
@@ -137,7 +137,6 @@ public:
std::vector<InsertStatement>::const_iterator begin,
std::vector<InsertStatement>::const_iterator end,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate) {
std::abort();
}
@@ -145,7 +144,6 @@ public:
Status insertDocument(OperationContext* opCtx,
const InsertStatement& doc,
OpDebug* opDebug,
- bool enforceQuota,
bool fromMigrate) {
std::abort();
}
@@ -159,8 +157,7 @@ public:
Status insertDocument(OperationContext* opCtx,
const BSONObj& doc,
- const std::vector<MultiIndexBlock*>& indexBlocks,
- bool enforceQuota) {
+ const std::vector<MultiIndexBlock*>& indexBlocks) {
std::abort();
}
@@ -168,7 +165,6 @@ public:
const RecordId& oldLocation,
const Snapshotted<BSONObj>& oldDoc,
const BSONObj& newDoc,
- bool enforceQuota,
bool indexesAffected,
OpDebug* opDebug,
OplogUpdateEntryArgs* args) {
diff --git a/src/mongo/db/catalog/rename_collection_test.cpp b/src/mongo/db/catalog/rename_collection_test.cpp
index 092b1328406..943fd3506ca 100644
--- a/src/mongo/db/catalog/rename_collection_test.cpp
+++ b/src/mongo/db/catalog/rename_collection_test.cpp
@@ -426,8 +426,7 @@ void _insertDocument(OperationContext* opCtx, const NamespaceString& nss, const
WriteUnitOfWork wuow(opCtx);
OpDebug* const opDebug = nullptr;
- bool enforceQuota = true;
- ASSERT_OK(collection->insertDocument(opCtx, InsertStatement(doc), opDebug, enforceQuota));
+ ASSERT_OK(collection->insertDocument(opCtx, InsertStatement(doc), opDebug));
wuow.commit();
});
}
diff --git a/src/mongo/db/commands/mr.cpp b/src/mongo/db/commands/mr.cpp
index dee774f7af2..29efeb7f3a9 100644
--- a/src/mongo/db/commands/mr.cpp
+++ b/src/mongo/db/commands/mr.cpp
@@ -843,7 +843,7 @@ void State::_insertToInc(BSONObj& o) {
// TODO: Consider whether to pass OpDebug for stats tracking under SERVER-23261.
OpDebug* const nullOpDebug = nullptr;
uassertStatusOK(autoColl.getCollection()->insertDocument(
- _opCtx, InsertStatement(o), nullOpDebug, true, false));
+ _opCtx, InsertStatement(o), nullOpDebug, false));
wuow.commit();
});
}
diff --git a/src/mongo/db/exec/update.cpp b/src/mongo/db/exec/update.cpp
index fabf092fa08..964b70866a7 100644
--- a/src/mongo/db/exec/update.cpp
+++ b/src/mongo/db/exec/update.cpp
@@ -334,7 +334,6 @@ BSONObj UpdateStage::transformAndUpdate(const Snapshotted<BSONObj>& oldObj, Reco
recordId,
oldObj,
newObj,
- true,
driver->modsAffectIndices(),
_params.opDebug,
&args);
@@ -469,11 +468,9 @@ void UpdateStage::doInsert() {
writeConflictRetry(getOpCtx(), "upsert", _collection->ns().ns(), [&] {
WriteUnitOfWork wunit(getOpCtx());
invariant(_collection);
- const bool enforceQuota = !request->isGod();
uassertStatusOK(_collection->insertDocument(getOpCtx(),
InsertStatement(request->getStmtId(), newObj),
_params.opDebug,
- enforceQuota,
request->isFromMigration()));
// Technically, we should save/restore state here, but since we are going to return
diff --git a/src/mongo/db/ops/write_ops_exec.cpp b/src/mongo/db/ops/write_ops_exec.cpp
index 9718457f0be..1c6f944d75b 100644
--- a/src/mongo/db/ops/write_ops_exec.cpp
+++ b/src/mongo/db/ops/write_ops_exec.cpp
@@ -354,8 +354,7 @@ void insertDocuments(OperationContext* opCtx,
}
}
- uassertStatusOK(collection->insertDocuments(
- opCtx, begin, end, &CurOp::get(opCtx)->debug(), /*enforceQuota*/ true));
+ uassertStatusOK(collection->insertDocuments(opCtx, begin, end, &CurOp::get(opCtx)->debug()));
wuow.commit();
}
diff --git a/src/mongo/db/repl/collection_bulk_loader_impl.cpp b/src/mongo/db/repl/collection_bulk_loader_impl.cpp
index b0ccc6dfcc9..276aaba8c11 100644
--- a/src/mongo/db/repl/collection_bulk_loader_impl.cpp
+++ b/src/mongo/db/repl/collection_bulk_loader_impl.cpp
@@ -126,7 +126,7 @@ Status CollectionBulkLoaderImpl::insertDocuments(const std::vector<BSONObj>::con
// This flavor of insertDocument will not update any pre-existing indexes,
// only the indexers passed in.
const auto status = _autoColl->getCollection()->insertDocument(
- _opCtx.get(), *iter, indexers, false);
+ _opCtx.get(), *iter, indexers);
if (!status.isOK()) {
return status;
}
@@ -134,7 +134,7 @@ Status CollectionBulkLoaderImpl::insertDocuments(const std::vector<BSONObj>::con
// For capped collections, we use regular insertDocument, which will update
// pre-existing indexes.
const auto status = _autoColl->getCollection()->insertDocument(
- _opCtx.get(), InsertStatement(*iter), nullptr, false);
+ _opCtx.get(), InsertStatement(*iter), nullptr);
if (!status.isOK()) {
return status;
}
diff --git a/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp b/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp
index e173b709602..d03b88b1253 100644
--- a/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp
+++ b/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp
@@ -102,7 +102,6 @@ void MockReplCoordServerFixture::insertOplogEntry(const repl::OplogEntry& entry)
auto status = coll->insertDocument(opCtx(),
InsertStatement(entry.toBSON()),
&CurOp::get(opCtx())->debug(),
- /* enforceQuota */ false,
/* fromMigrate */ false);
ASSERT_OK(status);
}
diff --git a/src/mongo/db/s/collection_sharding_state_test.cpp b/src/mongo/db/s/collection_sharding_state_test.cpp
index 6934bddb5e3..3fda81a7d21 100644
--- a/src/mongo/db/s/collection_sharding_state_test.cpp
+++ b/src/mongo/db/s/collection_sharding_state_test.cpp
@@ -95,7 +95,7 @@ TEST_F(CollectionShardingStateTest, GlobalInitDoesntGetCalledIfWriteAborts) {
WriteUnitOfWork wuow(operationContext());
ASSERT_OK(autoColl.getCollection()->insertDocument(
- operationContext(), shardIdentity.toShardIdentityDocument(), {}, false));
+ operationContext(), shardIdentity.toShardIdentityDocument(), {}));
ASSERT_EQ(0, getInitCallCount());
}
diff --git a/src/mongo/db/session.cpp b/src/mongo/db/session.cpp
index df347194ea7..26f621facbb 100644
--- a/src/mongo/db/session.cpp
+++ b/src/mongo/db/session.cpp
@@ -231,7 +231,7 @@ void updateSessionEntry(OperationContext* opCtx, const UpdateRequest& updateRequ
if (recordId.isNull()) {
// Upsert case.
auto status = collection->insertDocument(
- opCtx, InsertStatement(updateRequest.getUpdates()), nullptr, true, false);
+ opCtx, InsertStatement(updateRequest.getUpdates()), nullptr, false);
if (status == ErrorCodes::DuplicateKey) {
throw WriteConflictException();
@@ -266,7 +266,6 @@ void updateSessionEntry(OperationContext* opCtx, const UpdateRequest& updateRequ
recordId,
Snapshotted<BSONObj>(startingSnapshotId, originalDoc),
updateRequest.getUpdates(),
- true, // enforceQuota
false, // indexesAffected = false because _id is the only index
nullptr,
&args);
diff --git a/src/mongo/db/storage/devnull/devnull_kv_engine.cpp b/src/mongo/db/storage/devnull/devnull_kv_engine.cpp
index 7f0b16fa26b..bd48a8cfddd 100644
--- a/src/mongo/db/storage/devnull/devnull_kv_engine.cpp
+++ b/src/mongo/db/storage/devnull/devnull_kv_engine.cpp
@@ -101,8 +101,10 @@ public:
virtual void deleteRecord(OperationContext* opCtx, const RecordId& dl) {}
- virtual StatusWith<RecordId> insertRecord(
- OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota) {
+ virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx,
+ const char* data,
+ int len,
+ Timestamp) {
_numInserts++;
return StatusWith<RecordId>(RecordId(6, 4));
}
@@ -125,7 +127,6 @@ public:
const RecordId& oldLocation,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier) {
return Status::OK();
}
diff --git a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp
index f2ed2f99de2..aa69447703a 100644
--- a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp
+++ b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp
@@ -397,8 +397,10 @@ StatusWith<RecordId> EphemeralForTestRecordStore::extractAndCheckLocForOplog(con
return status;
}
-StatusWith<RecordId> EphemeralForTestRecordStore::insertRecord(
- OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota) {
+StatusWith<RecordId> EphemeralForTestRecordStore::insertRecord(OperationContext* opCtx,
+ const char* data,
+ int len,
+ Timestamp) {
if (_isCapped && len > _cappedMaxSize) {
// We use dataSize for capped rollover and we don't want to delete everything if we know
// this won't fit.
@@ -473,7 +475,6 @@ Status EphemeralForTestRecordStore::updateRecord(OperationContext* opCtx,
const RecordId& loc,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier) {
stdx::lock_guard<stdx::recursive_mutex> lock(_data->recordsMutex);
EphemeralForTestRecord* oldRecord = recordFor(loc);
@@ -620,9 +621,7 @@ Status EphemeralForTestRecordStore::touch(OperationContext* opCtx, BSONObjBuilde
return Status::OK();
}
-void EphemeralForTestRecordStore::increaseStorageSize(OperationContext* opCtx,
- int size,
- bool enforceQuota) {
+void EphemeralForTestRecordStore::increaseStorageSize(OperationContext* opCtx, int size) {
// unclear what this would mean for this class. For now, just error if called.
invariant(!"increaseStorageSize not yet implemented");
}
diff --git a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h
index 5e0d7ba1e5a..1b39400cc51 100644
--- a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h
+++ b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h
@@ -66,8 +66,10 @@ public:
virtual void deleteRecord(OperationContext* opCtx, const RecordId& dl);
- virtual StatusWith<RecordId> insertRecord(
- OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota);
+ virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx,
+ const char* data,
+ int len,
+ Timestamp);
virtual Status insertRecordsWithDocWriter(OperationContext* opCtx,
const DocWriter* const* docs,
@@ -79,7 +81,6 @@ public:
const RecordId& oldLocation,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier);
virtual bool updateWithDamagesSupported() const;
@@ -109,7 +110,7 @@ public:
virtual Status touch(OperationContext* opCtx, BSONObjBuilder* output) const;
- virtual void increaseStorageSize(OperationContext* opCtx, int size, bool enforceQuota);
+ virtual void increaseStorageSize(OperationContext* opCtx, int size);
virtual int64_t storageSize(OperationContext* opCtx,
BSONObjBuilder* extraInfo = NULL,
diff --git a/src/mongo/db/storage/kv/kv_catalog.cpp b/src/mongo/db/storage/kv/kv_catalog.cpp
index 283b0090a7e..89ad94b6544 100644
--- a/src/mongo/db/storage/kv/kv_catalog.cpp
+++ b/src/mongo/db/storage/kv/kv_catalog.cpp
@@ -306,17 +306,14 @@ void KVCatalog::FeatureTracker::putInfo(OperationContext* opCtx, const FeatureBi
if (_rid.isNull()) {
// This is the first time a feature is being marked as in-use or not in-use, so we must
// insert the feature document rather than update it.
- const bool enforceQuota = false;
// TODO SERVER-30638: using timestamp 0 for these inserts
- auto rid = _catalog->_rs->insertRecord(
- opCtx, obj.objdata(), obj.objsize(), Timestamp(), enforceQuota);
+ auto rid = _catalog->_rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), Timestamp());
fassert(40113, rid.getStatus());
_rid = rid.getValue();
} else {
- const bool enforceQuota = false;
UpdateNotifier* notifier = nullptr;
- auto status = _catalog->_rs->updateRecord(
- opCtx, _rid, obj.objdata(), obj.objsize(), enforceQuota, notifier);
+ auto status =
+ _catalog->_rs->updateRecord(opCtx, _rid, obj.objdata(), obj.objsize(), notifier);
fassert(40114, status);
}
}
@@ -425,10 +422,8 @@ Status KVCatalog::newCollection(OperationContext* opCtx,
b.append("md", md.toBSON());
obj = b.obj();
}
- const bool enforceQuota = false;
// TODO SERVER-30638: using timestamp 0 for these inserts.
- StatusWith<RecordId> res =
- _rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), Timestamp(), enforceQuota);
+ StatusWith<RecordId> res = _rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), Timestamp());
if (!res.isOK())
return res.getStatus();
@@ -524,7 +519,7 @@ void KVCatalog::putMetaData(OperationContext* opCtx,
}
LOG(3) << "recording new metadata: " << obj;
- Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), false, NULL);
+ Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), NULL);
fassert(28521, status.isOK());
}
@@ -549,7 +544,7 @@ Status KVCatalog::renameCollection(OperationContext* opCtx,
b.appendElementsUnique(old);
BSONObj obj = b.obj();
- Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), false, NULL);
+ Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), NULL);
fassert(28522, status.isOK());
}
diff --git a/src/mongo/db/storage/kv/kv_engine_test_harness.cpp b/src/mongo/db/storage/kv/kv_engine_test_harness.cpp
index e0eebfaa73d..6b9b671fd37 100644
--- a/src/mongo/db/storage/kv/kv_engine_test_harness.cpp
+++ b/src/mongo/db/storage/kv/kv_engine_test_harness.cpp
@@ -75,7 +75,7 @@ TEST(KVEngineTestHarness, SimpleRS1) {
{
MyOperationContext opCtx(engine);
WriteUnitOfWork uow(&opCtx);
- StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -115,7 +115,7 @@ TEST(KVEngineTestHarness, Restart1) {
{
MyOperationContext opCtx(engine);
WriteUnitOfWork uow(&opCtx);
- StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp b/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp
index d2c24f3b44e..90c51d942c8 100644
--- a/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp
+++ b/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp
@@ -101,7 +101,7 @@ public:
}
RecordId insertRecord(OperationContext* opCtx, std::string contents = "abcd") {
- auto id = rs->insertRecord(opCtx, contents.c_str(), contents.length() + 1, _counter, false);
+ auto id = rs->insertRecord(opCtx, contents.c_str(), contents.length() + 1, _counter);
ASSERT_OK(id);
return id.getValue();
}
@@ -118,8 +118,7 @@ public:
auto op = makeOperation();
WriteUnitOfWork wuow(op);
ASSERT_OK(op->recoveryUnit()->setTimestamp(_counter));
- ASSERT_OK(
- rs->updateRecord(op, id, contents.c_str(), contents.length() + 1, false, nullptr));
+ ASSERT_OK(rs->updateRecord(op, id, contents.c_str(), contents.length() + 1, nullptr));
wuow.commit();
}
diff --git a/src/mongo/db/storage/mobile/mobile_record_store.cpp b/src/mongo/db/storage/mobile/mobile_record_store.cpp
index 4e62e87b29b..2e1aed4c5e2 100644
--- a/src/mongo/db/storage/mobile/mobile_record_store.cpp
+++ b/src/mongo/db/storage/mobile/mobile_record_store.cpp
@@ -291,8 +291,10 @@ void MobileRecordStore::deleteRecord(OperationContext* opCtx, const RecordId& re
deleteStmt.step(SQLITE_DONE);
}
-StatusWith<RecordId> MobileRecordStore::insertRecord(
- OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota) {
+StatusWith<RecordId> MobileRecordStore::insertRecord(OperationContext* opCtx,
+ const char* data,
+ int len,
+ Timestamp) {
// Inserts record into SQLite table (or replaces if duplicate record id).
MobileSession* session = MobileRecoveryUnit::get(opCtx)->getSession(opCtx, false);
@@ -326,7 +328,7 @@ Status MobileRecordStore::insertRecordsWithDocWriter(OperationContext* opCtx,
for (size_t i = 0; i < nDocs; i++) {
docs[i]->writeDocument(pos);
size_t docLen = docs[i]->documentSize();
- StatusWith<RecordId> res = insertRecord(opCtx, pos, docLen, timestamps[i], true);
+ StatusWith<RecordId> res = insertRecord(opCtx, pos, docLen, timestamps[i]);
idsOut[i] = res.getValue();
pos += docLen;
}
@@ -338,7 +340,6 @@ Status MobileRecordStore::updateRecord(OperationContext* opCtx,
const RecordId& recId,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier) {
MobileSession* session = MobileRecoveryUnit::get(opCtx)->getSession(opCtx, false);
std::string dataSizeQuery =
diff --git a/src/mongo/db/storage/mobile/mobile_record_store.h b/src/mongo/db/storage/mobile/mobile_record_store.h
index d0771f0317a..ec8ab13aa2e 100644
--- a/src/mongo/db/storage/mobile/mobile_record_store.h
+++ b/src/mongo/db/storage/mobile/mobile_record_store.h
@@ -63,8 +63,7 @@ public:
StatusWith<RecordId> insertRecord(OperationContext* opCtx,
const char* data,
int len,
- Timestamp timestamp,
- bool enforceQuota) override;
+ Timestamp timestamp) override;
Status insertRecordsWithDocWriter(OperationContext* opCtx,
const DocWriter* const* docs,
@@ -76,7 +75,6 @@ public:
const RecordId& oldLocation,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier) override;
bool updateWithDamagesSupported() const override;
diff --git a/src/mongo/db/storage/record_store.h b/src/mongo/db/storage/record_store.h
index e09ef89980b..32f7b41cf7b 100644
--- a/src/mongo/db/storage/record_store.h
+++ b/src/mongo/db/storage/record_store.h
@@ -372,20 +372,15 @@ public:
virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx,
const char* data,
int len,
- Timestamp timestamp,
- bool enforceQuota) = 0;
+ Timestamp timestamp) = 0;
virtual Status insertRecords(OperationContext* opCtx,
std::vector<Record>* records,
- std::vector<Timestamp>* timestamps,
- bool enforceQuota) {
+ std::vector<Timestamp>* timestamps) {
int index = 0;
for (auto& record : *records) {
- StatusWith<RecordId> res = insertRecord(opCtx,
- record.data.data(),
- record.data.size(),
- (*timestamps)[index++],
- enforceQuota);
+ StatusWith<RecordId> res =
+ insertRecord(opCtx, record.data.data(), record.data.size(), (*timestamps)[index++]);
if (!res.isOK())
return res.getStatus();
@@ -431,7 +426,6 @@ public:
const RecordId& oldLocation,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier) = 0;
/**
diff --git a/src/mongo/db/storage/record_store_test_capped_visibility.cpp b/src/mongo/db/storage/record_store_test_capped_visibility.cpp
index 6b778dd2dce..e66544ce0c4 100644
--- a/src/mongo/db/storage/record_store_test_capped_visibility.cpp
+++ b/src/mongo/db/storage/record_store_test_capped_visibility.cpp
@@ -42,7 +42,7 @@ namespace {
RecordId doInsert(unowned_ptr<OperationContext> opCtx, unowned_ptr<RecordStore> rs) {
static char zeros[16];
- return uassertStatusOK(rs->insertRecord(opCtx, zeros, sizeof(zeros), Timestamp(), false));
+ return uassertStatusOK(rs->insertRecord(opCtx, zeros, sizeof(zeros), Timestamp()));
}
// macro to keep assert line numbers correct.
diff --git a/src/mongo/db/storage/record_store_test_datafor.cpp b/src/mongo/db/storage/record_store_test_datafor.cpp
index 672fa03e017..07f94f9ed9f 100644
--- a/src/mongo/db/storage/record_store_test_datafor.cpp
+++ b/src/mongo/db/storage/record_store_test_datafor.cpp
@@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, DataFor) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -107,7 +107,7 @@ TEST(RecordStoreTestHarness, DataForMultiple) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_datasize.cpp b/src/mongo/db/storage/record_store_test_datasize.cpp
index 3dad519623b..b57cebf06bd 100644
--- a/src/mongo/db/storage/record_store_test_datasize.cpp
+++ b/src/mongo/db/storage/record_store_test_datasize.cpp
@@ -80,7 +80,7 @@ TEST(RecordStoreTestHarness, DataSizeNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
diff --git a/src/mongo/db/storage/record_store_test_deleterecord.cpp b/src/mongo/db/storage/record_store_test_deleterecord.cpp
index 0d630c0d0be..90d8f274c86 100644
--- a/src/mongo/db/storage/record_store_test_deleterecord.cpp
+++ b/src/mongo/db/storage/record_store_test_deleterecord.cpp
@@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, DeleteRecord) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -111,7 +111,7 @@ TEST(RecordStoreTestHarness, DeleteMultipleRecords) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_harness.cpp b/src/mongo/db/storage/record_store_test_harness.cpp
index 2287e03052c..4ee95575612 100644
--- a/src/mongo/db/storage/record_store_test_harness.cpp
+++ b/src/mongo/db/storage/record_store_test_harness.cpp
@@ -60,7 +60,7 @@ TEST(RecordStoreTestHarness, Simple1) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc1 = res.getValue();
uow.commit();
@@ -87,7 +87,7 @@ TEST(RecordStoreTestHarness, Simple1) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
@@ -160,7 +160,7 @@ TEST(RecordStoreTestHarness, Delete1) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -205,9 +205,9 @@ TEST(RecordStoreTestHarness, Delete2) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
- res = rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false);
+ res = rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -248,7 +248,7 @@ TEST(RecordStoreTestHarness, Update1) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s1.c_str(), s1.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s1.c_str(), s1.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -264,8 +264,7 @@ TEST(RecordStoreTestHarness, Update1) {
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
{
WriteUnitOfWork uow(opCtx.get());
- Status status =
- rs->updateRecord(opCtx.get(), loc, s2.c_str(), s2.size() + 1, false, NULL);
+ Status status = rs->updateRecord(opCtx.get(), loc, s2.c_str(), s2.size() + 1, NULL);
ASSERT_OK(status);
uow.commit();
@@ -296,7 +295,7 @@ TEST(RecordStoreTestHarness, UpdateInPlace1) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s1Rec.data(), s1Rec.size(), Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s1Rec.data(), s1Rec.size(), Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -350,7 +349,7 @@ TEST(RecordStoreTestHarness, Truncate1) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -400,7 +399,7 @@ TEST(RecordStoreTestHarness, Cursor1) {
WriteUnitOfWork uow(opCtx.get());
for (int i = 0; i < N; i++) {
string s = str::stream() << "eliot" << i;
- ASSERT_OK(rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false)
+ ASSERT_OK(rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp())
.getStatus());
}
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_insertrecord.cpp b/src/mongo/db/storage/record_store_test_insertrecord.cpp
index adcf2fc2300..0cef303a730 100644
--- a/src/mongo/db/storage/record_store_test_insertrecord.cpp
+++ b/src/mongo/db/storage/record_store_test_insertrecord.cpp
@@ -61,7 +61,7 @@ TEST(RecordStoreTestHarness, InsertRecord) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -96,7 +96,7 @@ TEST(RecordStoreTestHarness, InsertMultipleRecords) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_manyiter.cpp b/src/mongo/db/storage/record_store_test_manyiter.cpp
index 5b477f0d27a..2080f1ca073 100644
--- a/src/mongo/db/storage/record_store_test_manyiter.cpp
+++ b/src/mongo/db/storage/record_store_test_manyiter.cpp
@@ -87,7 +87,7 @@ TEST(RecordStoreTestHarness, GetManyIteratorsNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_randomiter.cpp b/src/mongo/db/storage/record_store_test_randomiter.cpp
index 0644b07a83b..3cdeec17f2d 100644
--- a/src/mongo/db/storage/record_store_test_randomiter.cpp
+++ b/src/mongo/db/storage/record_store_test_randomiter.cpp
@@ -87,7 +87,7 @@ TEST(RecordStoreTestHarness, GetRandomIteratorNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
@@ -141,8 +141,7 @@ TEST(RecordStoreTestHarness, GetRandomIteratorSingleton) {
{
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), "some data", 10, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "some data", 10, Timestamp());
ASSERT_OK(res.getStatus());
idToRetrieve = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_recorditer.cpp b/src/mongo/db/storage/record_store_test_recorditer.cpp
index 37f6d284258..968dfbcb760 100644
--- a/src/mongo/db/storage/record_store_test_recorditer.cpp
+++ b/src/mongo/db/storage/record_store_test_recorditer.cpp
@@ -71,7 +71,7 @@ TEST(RecordStoreTestHarness, IterateOverMultipleRecords) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
@@ -122,7 +122,7 @@ TEST(RecordStoreTestHarness, IterateOverMultipleRecordsReversed) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
@@ -173,7 +173,7 @@ TEST(RecordStoreTestHarness, IterateStartFromMiddle) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
@@ -225,7 +225,7 @@ TEST(RecordStoreTestHarness, IterateStartFromMiddleReversed) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
@@ -278,7 +278,7 @@ TEST(RecordStoreTestHarness, RecordIteratorEOF) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
@@ -315,7 +315,7 @@ TEST(RecordStoreTestHarness, RecordIteratorEOF) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
@@ -349,7 +349,7 @@ TEST(RecordStoreTestHarness, RecordIteratorSaveRestore) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
@@ -406,7 +406,7 @@ TEST(RecordStoreTestHarness, SeekAfterEofAndContinue) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
datas[i] = data;
diff --git a/src/mongo/db/storage/record_store_test_repairiter.cpp b/src/mongo/db/storage/record_store_test_repairiter.cpp
index f5397012cf8..a84e7dcab4d 100644
--- a/src/mongo/db/storage/record_store_test_repairiter.cpp
+++ b/src/mongo/db/storage/record_store_test_repairiter.cpp
@@ -89,7 +89,7 @@ TEST(RecordStoreTestHarness, GetIteratorForRepairNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
@@ -136,8 +136,7 @@ TEST(RecordStoreTestHarness, GetIteratorForRepairInvalidateSingleton) {
{
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), "some data", 10, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "some data", 10, Timestamp());
ASSERT_OK(res.getStatus());
idToInvalidate = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_storagesize.cpp b/src/mongo/db/storage/record_store_test_storagesize.cpp
index 521d879cfac..c5bd0bcf7cc 100644
--- a/src/mongo/db/storage/record_store_test_storagesize.cpp
+++ b/src/mongo/db/storage/record_store_test_storagesize.cpp
@@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, StorageSizeNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
diff --git a/src/mongo/db/storage/record_store_test_touch.cpp b/src/mongo/db/storage/record_store_test_touch.cpp
index eae3e0046b0..dfd541250bb 100644
--- a/src/mongo/db/storage/record_store_test_touch.cpp
+++ b/src/mongo/db/storage/record_store_test_touch.cpp
@@ -85,7 +85,7 @@ TEST(RecordStoreTestHarness, TouchNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
@@ -149,7 +149,7 @@ TEST(RecordStoreTestHarness, TouchNonEmptyWithNullStats) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
diff --git a/src/mongo/db/storage/record_store_test_truncate.cpp b/src/mongo/db/storage/record_store_test_truncate.cpp
index 7fda629b4f9..aac72c5caf8 100644
--- a/src/mongo/db/storage/record_store_test_truncate.cpp
+++ b/src/mongo/db/storage/record_store_test_truncate.cpp
@@ -89,7 +89,7 @@ TEST(RecordStoreTestHarness, TruncateNonEmpty) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
diff --git a/src/mongo/db/storage/record_store_test_updaterecord.cpp b/src/mongo/db/storage/record_store_test_updaterecord.cpp
index f9c9e33c375..c6b96def16d 100644
--- a/src/mongo/db/storage/record_store_test_updaterecord.cpp
+++ b/src/mongo/db/storage/record_store_test_updaterecord.cpp
@@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, UpdateRecord) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -80,8 +80,7 @@ TEST(RecordStoreTestHarness, UpdateRecord) {
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
{
WriteUnitOfWork uow(opCtx.get());
- Status res =
- rs->updateRecord(opCtx.get(), loc, data.c_str(), data.size() + 1, false, NULL);
+ Status res = rs->updateRecord(opCtx.get(), loc, data.c_str(), data.size() + 1, NULL);
ASSERT_OK(res);
uow.commit();
@@ -119,7 +118,7 @@ TEST(RecordStoreTestHarness, UpdateMultipleRecords) {
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
locs[i] = res.getValue();
uow.commit();
@@ -140,7 +139,7 @@ TEST(RecordStoreTestHarness, UpdateMultipleRecords) {
WriteUnitOfWork uow(opCtx.get());
Status res =
- rs->updateRecord(opCtx.get(), locs[i], data.c_str(), data.size() + 1, false, NULL);
+ rs->updateRecord(opCtx.get(), locs[i], data.c_str(), data.size() + 1, NULL);
ASSERT_OK(res);
uow.commit();
@@ -177,8 +176,8 @@ TEST(RecordStoreTestHarness, UpdateRecordWithMoveNotifier) {
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
{
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(
- opCtx.get(), oldData.c_str(), oldData.size() + 1, Timestamp(), false);
+ StatusWith<RecordId> res =
+ rs->insertRecord(opCtx.get(), oldData.c_str(), oldData.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -197,8 +196,8 @@ TEST(RecordStoreTestHarness, UpdateRecordWithMoveNotifier) {
UpdateNotifierSpy umn(opCtx.get(), loc, oldData.c_str(), oldData.size());
WriteUnitOfWork uow(opCtx.get());
- Status res = rs->updateRecord(
- opCtx.get(), loc, newData.c_str(), newData.size() + 1, false, &umn);
+ Status res =
+ rs->updateRecord(opCtx.get(), loc, newData.c_str(), newData.size() + 1, &umn);
ASSERT_OK(res);
ASSERT_GTE(1, umn.numInPlaceCallbacks());
diff --git a/src/mongo/db/storage/record_store_test_updatewithdamages.cpp b/src/mongo/db/storage/record_store_test_updatewithdamages.cpp
index 2e839f28421..106d8f6e17c 100644
--- a/src/mongo/db/storage/record_store_test_updatewithdamages.cpp
+++ b/src/mongo/db/storage/record_store_test_updatewithdamages.cpp
@@ -65,7 +65,7 @@ TEST(RecordStoreTestHarness, UpdateWithDamages) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false);
+ rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -131,7 +131,7 @@ TEST(RecordStoreTestHarness, UpdateWithOverlappingDamageEvents) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false);
+ rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -195,7 +195,7 @@ TEST(RecordStoreTestHarness, UpdateWithOverlappingDamageEventsReversed) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false);
+ rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
@@ -257,7 +257,7 @@ TEST(RecordStoreTestHarness, UpdateWithNoDamages) {
{
WriteUnitOfWork uow(opCtx.get());
StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false);
+ rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp());
ASSERT_OK(res.getStatus());
loc = res.getValue();
uow.commit();
diff --git a/src/mongo/db/storage/record_store_test_validate.h b/src/mongo/db/storage/record_store_test_validate.h
index 7edc8e152ec..4cecc65bf56 100644
--- a/src/mongo/db/storage/record_store_test_validate.h
+++ b/src/mongo/db/storage/record_store_test_validate.h
@@ -96,8 +96,8 @@ public:
ASSERT(_remain.insert(data).second);
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = _rs->insertRecord(
- opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false);
+ StatusWith<RecordId> res =
+ _rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
diff --git a/src/mongo/db/storage/recovery_unit_test_harness.cpp b/src/mongo/db/storage/recovery_unit_test_harness.cpp
index 1ec45a1f929..a5e4f210c75 100644
--- a/src/mongo/db/storage/recovery_unit_test_harness.cpp
+++ b/src/mongo/db/storage/recovery_unit_test_harness.cpp
@@ -53,7 +53,7 @@ public:
TEST_F(RecoveryUnitTestHarness, CommitUnitOfWork) {
const auto rs = harnessHelper->createRecordStore(opCtx.get(), "table1");
ru->beginUnitOfWork(opCtx.get());
- StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp(), false);
+ StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp());
ASSERT_TRUE(s.isOK());
ASSERT_EQUALS(1, rs->numRecords(NULL));
ru->commitUnitOfWork();
@@ -64,7 +64,7 @@ TEST_F(RecoveryUnitTestHarness, CommitUnitOfWork) {
TEST_F(RecoveryUnitTestHarness, AbortUnitOfWork) {
const auto rs = harnessHelper->createRecordStore(opCtx.get(), "table1");
ru->beginUnitOfWork(opCtx.get());
- StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp(), false);
+ StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp());
ASSERT_TRUE(s.isOK());
ASSERT_EQUALS(1, rs->numRecords(NULL));
ru->abortUnitOfWork();
diff --git a/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp b/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp
index d85d6e35e16..73e84d06f05 100644
--- a/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp
+++ b/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp
@@ -44,8 +44,7 @@ StatusWith<RecordId> insertBSON(ServiceContext::UniqueOperationContext& opCtx,
Status status = rs->oplogDiskLocRegister(opCtx.get(), opTime, false);
if (!status.isOK())
return StatusWith<RecordId>(status);
- StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime, false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime);
if (res.isOK())
wuow.commit();
return res;
@@ -58,7 +57,7 @@ RecordId _oplogOrderInsertOplog(OperationContext* opCtx,
Status status = rs->oplogDiskLocRegister(opCtx, opTime, false);
ASSERT_OK(status);
BSONObj obj = BSON("ts" << opTime);
- StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime, false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime);
ASSERT_OK(res.getStatus());
return res.getValue();
}
@@ -79,19 +78,17 @@ TEST(RecordStore_Oplog, OplogHack) {
{
WriteUnitOfWork wuow(opCtx.get());
BSONObj obj = BSON("not_ts" << Timestamp(2, 1));
- ASSERT_EQ(
- rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, 1), false)
- .getStatus(),
- ErrorCodes::BadValue);
+ ASSERT_EQ(rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, 1))
+ .getStatus(),
+ ErrorCodes::BadValue);
}
{
WriteUnitOfWork wuow(opCtx.get());
BSONObj obj = BSON("ts"
<< "not a Timestamp");
- ASSERT_EQ(
- rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(), false)
- .getStatus(),
- ErrorCodes::BadValue);
+ ASSERT_EQ(rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp())
+ .getStatus(),
+ ErrorCodes::BadValue);
}
ASSERT_EQ(insertBSON(opCtx, rs, Timestamp(-2, 1)).getStatus(), ErrorCodes::BadValue);
@@ -167,9 +164,8 @@ TEST(RecordStore_Oplog, OplogHackOnNonOplog) {
BSONObj obj = BSON("ts" << Timestamp(2, -1));
{
WriteUnitOfWork wuow(opCtx.get());
- ASSERT_OK(
- rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, -1), false)
- .getStatus());
+ ASSERT_OK(rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, -1))
+ .getStatus());
wuow.commit();
}
ASSERT_EQ(rs->oplogStartHack(opCtx.get(), RecordId(0, 1)), boost::none);
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp
index f9d431af36a..c205e361415 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp
@@ -213,7 +213,7 @@ TEST(WiredTigerRecordStoreTest, PrefixedTableScan) {
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
for (int num = 0; num < numDocs; ++num) {
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
@@ -239,7 +239,7 @@ TEST(WiredTigerRecordStoreTest, PrefixedSeekingCursor) {
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
for (int num = 0; num < numDocs; ++num) {
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
if (startRecordId.isNull()) {
startRecordId = res.getValue();
}
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp
index 19acc2dbc6d..ab92b1935bd 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp
@@ -1239,8 +1239,7 @@ void WiredTigerRecordStore::reclaimOplog(OperationContext* opCtx, Timestamp pers
Status WiredTigerRecordStore::insertRecords(OperationContext* opCtx,
std::vector<Record>* records,
- std::vector<Timestamp>* timestamps,
- bool enforceQuota) {
+ std::vector<Timestamp>* timestamps) {
return _insertRecords(opCtx, records->data(), timestamps->data(), records->size());
}
@@ -1324,8 +1323,10 @@ Status WiredTigerRecordStore::_insertRecords(OperationContext* opCtx,
return Status::OK();
}
-StatusWith<RecordId> WiredTigerRecordStore::insertRecord(
- OperationContext* opCtx, const char* data, int len, Timestamp timestamp, bool enforceQuota) {
+StatusWith<RecordId> WiredTigerRecordStore::insertRecord(OperationContext* opCtx,
+ const char* data,
+ int len,
+ Timestamp timestamp) {
Record record = {RecordId(), RecordData(data, len)};
Status status = _insertRecords(opCtx, &record, &timestamp, 1);
if (!status.isOK())
@@ -1399,7 +1400,6 @@ Status WiredTigerRecordStore::updateRecord(OperationContext* opCtx,
const RecordId& id,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier) {
dassert(opCtx->lockState()->isWriteLocked());
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h
index 5ef7017faca..9aa98659c70 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h
@@ -143,11 +143,12 @@ public:
virtual Status insertRecords(OperationContext* opCtx,
std::vector<Record>* records,
- std::vector<Timestamp>* timestamps,
- bool enforceQuota);
+ std::vector<Timestamp>* timestamps);
- virtual StatusWith<RecordId> insertRecord(
- OperationContext* opCtx, const char* data, int len, Timestamp timestamp, bool enforceQuota);
+ virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx,
+ const char* data,
+ int len,
+ Timestamp timestamp);
virtual Status insertRecordsWithDocWriter(OperationContext* opCtx,
const DocWriter* const* docs,
@@ -159,7 +160,6 @@ public:
const RecordId& oldLocation,
const char* data,
int len,
- bool enforceQuota,
UpdateNotifier* notifier);
virtual bool updateWithDamagesSupported() const;
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp
index 73841f20e29..28b6efcebaf 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp
@@ -113,11 +113,11 @@ TEST(WiredTigerRecordStoreTest, Isolation1) {
{
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
id1 = res.getValue();
- res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
id2 = res.getValue();
@@ -136,12 +136,12 @@ TEST(WiredTigerRecordStoreTest, Isolation1) {
rs->dataFor(t1.get(), id1);
rs->dataFor(t2.get(), id1);
- ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, false, NULL));
- ASSERT_OK(rs->updateRecord(t1.get(), id2, "B", 2, false, NULL));
+ ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, NULL));
+ ASSERT_OK(rs->updateRecord(t1.get(), id2, "B", 2, NULL));
try {
// this should fail
- rs->updateRecord(t2.get(), id1, "c", 2, false, NULL).transitional_ignore();
+ rs->updateRecord(t2.get(), id1, "c", 2, NULL).transitional_ignore();
ASSERT(0);
} catch (WriteConflictException&) {
w2.reset(NULL);
@@ -164,11 +164,11 @@ TEST(WiredTigerRecordStoreTest, Isolation2) {
{
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
id1 = res.getValue();
- res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
id2 = res.getValue();
@@ -187,7 +187,7 @@ TEST(WiredTigerRecordStoreTest, Isolation2) {
{
WriteUnitOfWork w(t1.get());
- ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, false, NULL));
+ ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, NULL));
w.commit();
}
@@ -196,7 +196,7 @@ TEST(WiredTigerRecordStoreTest, Isolation2) {
ASSERT_EQUALS(string("a"), rs->dataFor(t2.get(), id1).data());
try {
// this should fail as our version of id1 is too old
- rs->updateRecord(t2.get(), id1, "c", 2, false, NULL).transitional_ignore();
+ rs->updateRecord(t2.get(), id1, "c", 2, NULL).transitional_ignore();
ASSERT(0);
} catch (WriteConflictException&) {
}
@@ -215,8 +215,7 @@ StatusWith<RecordId> insertBSON(ServiceContext::UniqueOperationContext& opCtx,
Status status = wrs->oplogDiskLocRegister(opCtx.get(), opTime, false);
if (!status.isOK())
return StatusWith<RecordId>(status);
- StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime, false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime);
if (res.isOK())
wuow.commit();
return res;
@@ -230,7 +229,7 @@ TEST(WiredTigerRecordStoreTest, CappedCursorRollover) {
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
for (int i = 0; i < 3; ++i) {
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
@@ -250,7 +249,7 @@ TEST(WiredTigerRecordStoreTest, CappedCursorRollover) {
auto opCtx = harnessHelper->newOperationContext(client3.get());
for (int i = 0; i < 100; i++) {
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
uow.commit();
}
@@ -268,7 +267,7 @@ RecordId _oplogOrderInsertOplog(OperationContext* opCtx,
Status status = rs->oplogDiskLocRegister(opCtx, opTime, false);
ASSERT_OK(status);
BSONObj obj = BSON("ts" << opTime);
- StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime, false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime);
ASSERT_OK(res.getStatus());
return res.getValue();
}
@@ -384,7 +383,7 @@ TEST(WiredTigerRecordStoreTest, CappedCursorYieldFirst) {
{ // first insert a document
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
WriteUnitOfWork uow(opCtx.get());
- StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
id1 = res.getValue();
uow.commit();
@@ -428,7 +427,7 @@ StatusWith<RecordId> insertBSONWithSize(OperationContext* opCtx,
if (!status.isOK()) {
return StatusWith<RecordId>(status);
}
- StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime, false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime);
if (res.isOK()) {
wuow.commit();
}
@@ -526,9 +525,9 @@ TEST(WiredTigerRecordStoreTest, OplogStones_UpdateRecord) {
WriteUnitOfWork wuow(opCtx.get());
ASSERT_NOT_OK(rs->updateRecord(
- opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), false, nullptr));
+ opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), nullptr));
ASSERT_NOT_OK(rs->updateRecord(
- opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), false, nullptr));
+ opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), nullptr));
}
// Attempts to shrink the records should also fail.
@@ -540,9 +539,9 @@ TEST(WiredTigerRecordStoreTest, OplogStones_UpdateRecord) {
WriteUnitOfWork wuow(opCtx.get());
ASSERT_NOT_OK(rs->updateRecord(
- opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), false, nullptr));
+ opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), nullptr));
ASSERT_NOT_OK(rs->updateRecord(
- opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), false, nullptr));
+ opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), nullptr));
}
// Changing the contents of the records without changing their size should succeed.
@@ -554,9 +553,9 @@ TEST(WiredTigerRecordStoreTest, OplogStones_UpdateRecord) {
WriteUnitOfWork wuow(opCtx.get());
ASSERT_OK(rs->updateRecord(
- opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), false, nullptr));
+ opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), nullptr));
ASSERT_OK(rs->updateRecord(
- opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), false, nullptr));
+ opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), nullptr));
wuow.commit();
ASSERT_EQ(1U, oplogStones->numStones());
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp
index 75dc1b4d029..0c3f74516f9 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp
@@ -184,8 +184,7 @@ TEST_F(WiredTigerRecoveryUnitTestFixture, CreateAndCheckForCachePressure) {
WriteUnitOfWork wu(opCtx);
ASSERT_OK(ru1->setTimestamp(Timestamp(time++)));
std::string str = str::stream() << "foobarbaz";
- StatusWith<RecordId> ress =
- rs->insertRecord(opCtx, str.c_str(), str.size() + 1, Timestamp(), false);
+ StatusWith<RecordId> ress = rs->insertRecord(opCtx, str.c_str(), str.size() + 1, Timestamp());
ASSERT_OK(ress.getStatus());
auto recordId = ress.getValue();
wu.commit();
@@ -203,7 +202,7 @@ TEST_F(WiredTigerRecoveryUnitTestFixture, CreateAndCheckForCachePressure) {
ASSERT_OK(ru1->setTimestamp(Timestamp(time++)));
std::string s = str::stream()
<< "abcbcdcdedefefgfghghihijijkjklklmlmnmnomopopqpqrqrsrststutuv" << j;
- ASSERT_OK(rs->updateRecord(opCtx, recordId, s.c_str(), s.size() + 1, false, nullptr));
+ ASSERT_OK(rs->updateRecord(opCtx, recordId, s.c_str(), s.size() + 1, nullptr));
wuow.commit();
} catch (const DBException& ex) {
invariant(ex.toStatus().code() == ErrorCodes::WriteConflict);
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp
index 479bd470bf5..b914b4d6d57 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp
@@ -227,8 +227,7 @@ TEST(WiredTigerRecordStoreTest, SizeStorer1) {
{
WriteUnitOfWork uow(opCtx.get());
for (int i = 0; i < N; i++) {
- StatusWith<RecordId> res =
- rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false);
+ StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp());
ASSERT_OK(res.getStatus());
}
uow.commit();
@@ -329,7 +328,7 @@ private:
ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext());
WriteUnitOfWork uow(opCtx.get());
for (int i = 0; i < expectedNumRecords; i++) {
- ASSERT_OK(rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false).getStatus());
+ ASSERT_OK(rs->insertRecord(opCtx.get(), "a", 2, Timestamp()).getStatus());
}
uow.commit();
}
diff --git a/src/mongo/db/views/durable_view_catalog.cpp b/src/mongo/db/views/durable_view_catalog.cpp
index c9a3718fd39..b8691d5f15c 100644
--- a/src/mongo/db/views/durable_view_catalog.cpp
+++ b/src/mongo/db/views/durable_view_catalog.cpp
@@ -143,12 +143,11 @@ void DurableViewCatalogImpl::upsert(OperationContext* opCtx,
const bool requireIndex = false;
RecordId id = Helpers::findOne(opCtx, systemViews, BSON("_id" << name.ns()), requireIndex);
- const bool enforceQuota = true;
Snapshotted<BSONObj> oldView;
if (!id.isNormal() || !systemViews->findDoc(opCtx, id, &oldView)) {
LOG(2) << "insert view " << view << " into " << _db->getSystemViewsName();
- uassertStatusOK(systemViews->insertDocument(
- opCtx, InsertStatement(view), &CurOp::get(opCtx)->debug(), enforceQuota));
+ uassertStatusOK(
+ systemViews->insertDocument(opCtx, InsertStatement(view), &CurOp::get(opCtx)->debug()));
} else {
OplogUpdateEntryArgs args;
args.nss = systemViewsNs;
@@ -157,14 +156,8 @@ void DurableViewCatalogImpl::upsert(OperationContext* opCtx,
args.fromMigrate = false;
const bool assumeIndexesAreAffected = true;
- systemViews->updateDocument(opCtx,
- id,
- oldView,
- view,
- enforceQuota,
- assumeIndexesAreAffected,
- &CurOp::get(opCtx)->debug(),
- &args);
+ systemViews->updateDocument(
+ opCtx, id, oldView, view, assumeIndexesAreAffected, &CurOp::get(opCtx)->debug(), &args);
}
}
diff --git a/src/mongo/dbtests/indexupdatetests.cpp b/src/mongo/dbtests/indexupdatetests.cpp
index 697a8ce1abd..6856646b3c3 100644
--- a/src/mongo/dbtests/indexupdatetests.cpp
+++ b/src/mongo/dbtests/indexupdatetests.cpp
@@ -283,11 +283,11 @@ public:
coll = db->createCollection(&_opCtx, _ns);
// Drop all indexes including id index.
coll->getIndexCatalog()->dropAllIndexes(&_opCtx, true);
- // Insert some documents with enforceQuota=true.
+ // Insert some documents.
int32_t nDocs = 1000;
OpDebug* const nullOpDebug = nullptr;
for (int32_t i = 0; i < nDocs; ++i) {
- coll->insertDocument(&_opCtx, InsertStatement(BSON("a" << i)), nullOpDebug, true)
+ coll->insertDocument(&_opCtx, InsertStatement(BSON("a" << i)), nullOpDebug)
.transitional_ignore();
}
wunit.commit();
diff --git a/src/mongo/dbtests/multikey_paths_test.cpp b/src/mongo/dbtests/multikey_paths_test.cpp
index d3ee08fcc8a..e875ed9da1b 100644
--- a/src/mongo/dbtests/multikey_paths_test.cpp
+++ b/src/mongo/dbtests/multikey_paths_test.cpp
@@ -132,12 +132,10 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnIndexCreation) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
wuow.commit();
}
@@ -164,17 +162,14 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnIndexCreationWithMultipleDocuments) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(BSON("_id" << 1 << "a" << BSON_ARRAY(1 << 2 << 3) << "b" << 5)),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
wuow.commit();
}
@@ -213,12 +208,10 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentInsert) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
wuow.commit();
}
@@ -227,12 +220,10 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentInsert) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(BSON("_id" << 1 << "a" << BSON_ARRAY(1 << 2 << 3) << "b" << 5)),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
wuow.commit();
}
@@ -259,11 +250,8 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentUpdate) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
- ASSERT_OK(collection->insertDocument(_opCtx.get(),
- InsertStatement(BSON("_id" << 0 << "a" << 5)),
- nullOpDebug,
- enforceQuota));
+ ASSERT_OK(collection->insertDocument(
+ _opCtx.get(), InsertStatement(BSON("_id" << 0 << "a" << 5)), nullOpDebug));
wuow.commit();
}
@@ -277,7 +265,6 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentUpdate) {
auto oldDoc = collection->docFor(_opCtx.get(), record->id);
{
WriteUnitOfWork wuow(_opCtx.get());
- const bool enforceQuota = true;
const bool indexesAffected = true;
OpDebug* opDebug = nullptr;
OplogUpdateEntryArgs args;
@@ -286,7 +273,6 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentUpdate) {
record->id,
oldDoc,
BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3)),
- enforceQuota,
indexesAffected,
opDebug,
&args);
@@ -317,12 +303,10 @@ TEST_F(MultikeyPathsTest, PathsNotUpdatedOnDocumentDelete) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
wuow.commit();
}
@@ -375,13 +359,11 @@ TEST_F(MultikeyPathsTest, PathsUpdatedForMultipleIndexesOnDocumentInsert) {
{
WriteUnitOfWork wuow(_opCtx.get());
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = true;
ASSERT_OK(collection->insertDocument(
_opCtx.get(),
InsertStatement(
BSON("_id" << 0 << "a" << BSON_ARRAY(1 << 2 << 3) << "b" << 5 << "c" << 8)),
- nullOpDebug,
- enforceQuota));
+ nullOpDebug));
wuow.commit();
}
diff --git a/src/mongo/dbtests/query_stage_cached_plan.cpp b/src/mongo/dbtests/query_stage_cached_plan.cpp
index f5afe597563..51503ac4056 100644
--- a/src/mongo/dbtests/query_stage_cached_plan.cpp
+++ b/src/mongo/dbtests/query_stage_cached_plan.cpp
@@ -103,10 +103,8 @@ public:
void insertDocument(Collection* collection, BSONObj obj) {
WriteUnitOfWork wuow(&_opCtx);
- const bool enforceQuota = false;
OpDebug* const nullOpDebug = nullptr;
- ASSERT_OK(
- collection->insertDocument(&_opCtx, InsertStatement(obj), nullOpDebug, enforceQuota));
+ ASSERT_OK(collection->insertDocument(&_opCtx, InsertStatement(obj), nullOpDebug));
wuow.commit();
}
diff --git a/src/mongo/dbtests/query_stage_count.cpp b/src/mongo/dbtests/query_stage_count.cpp
index e28d388d0aa..a23c692ae1c 100644
--- a/src/mongo/dbtests/query_stage_count.cpp
+++ b/src/mongo/dbtests/query_stage_count.cpp
@@ -108,8 +108,7 @@ public:
void insert(const BSONObj& doc) {
WriteUnitOfWork wunit(&_opCtx);
OpDebug* const nullOpDebug = nullptr;
- _coll->insertDocument(&_opCtx, InsertStatement(doc), nullOpDebug, false)
- .transitional_ignore();
+ _coll->insertDocument(&_opCtx, InsertStatement(doc), nullOpDebug).transitional_ignore();
wunit.commit();
}
@@ -129,7 +128,6 @@ public:
oldrecordId,
Snapshotted<BSONObj>(_opCtx.recoveryUnit()->getSnapshotId(), oldDoc),
newDoc,
- false,
true,
NULL,
&args);
diff --git a/src/mongo/dbtests/query_stage_sort.cpp b/src/mongo/dbtests/query_stage_sort.cpp
index 9ff58f996cd..859bfaf8bce 100644
--- a/src/mongo/dbtests/query_stage_sort.cpp
+++ b/src/mongo/dbtests/query_stage_sort.cpp
@@ -362,7 +362,7 @@ public:
args.nss = coll->ns();
{
WriteUnitOfWork wuow(&_opCtx);
- coll->updateDocument(&_opCtx, *it, oldDoc, newDoc(oldDoc), false, false, NULL, &args);
+ coll->updateDocument(&_opCtx, *it, oldDoc, newDoc(oldDoc), false, NULL, &args);
wuow.commit();
}
ASSERT_OK(exec->restoreState());
@@ -380,8 +380,7 @@ public:
oldDoc = coll->docFor(&_opCtx, *it);
{
WriteUnitOfWork wuow(&_opCtx);
- coll->updateDocument(
- &_opCtx, *it++, oldDoc, newDoc(oldDoc), false, false, NULL, &args);
+ coll->updateDocument(&_opCtx, *it++, oldDoc, newDoc(oldDoc), false, NULL, &args);
wuow.commit();
}
}
diff --git a/src/mongo/dbtests/storage_timestamp_tests.cpp b/src/mongo/dbtests/storage_timestamp_tests.cpp
index 0a26e2d3002..7501d8ebdb5 100644
--- a/src/mongo/dbtests/storage_timestamp_tests.cpp
+++ b/src/mongo/dbtests/storage_timestamp_tests.cpp
@@ -216,9 +216,8 @@ public:
void insertDocument(Collection* coll, const InsertStatement& stmt) {
// Insert some documents.
OpDebug* const nullOpDebug = nullptr;
- const bool enforceQuota = false;
const bool fromMigrate = false;
- ASSERT_OK(coll->insertDocument(_opCtx, stmt, nullOpDebug, enforceQuota, fromMigrate));
+ ASSERT_OK(coll->insertDocument(_opCtx, stmt, nullOpDebug, fromMigrate));
}
void createIndex(Collection* coll, std::string indexName, const BSONObj& indexKey) {
diff --git a/src/mongo/dbtests/validate_tests.cpp b/src/mongo/dbtests/validate_tests.cpp
index 28d4f5e8498..87c4aaad88a 100644
--- a/src/mongo/dbtests/validate_tests.cpp
+++ b/src/mongo/dbtests/validate_tests.cpp
@@ -190,8 +190,7 @@ public:
WriteUnitOfWork wunit(&_opCtx);
for (int j = 0; j < 2; j++) {
auto doc = BSON("_id" << j);
- ASSERT_OK(rs->insertRecord(
- &_opCtx, doc.objdata(), doc.objsize(), Timestamp(), /*enforceQuota*/ false));
+ ASSERT_OK(rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), Timestamp()));
}
wunit.commit();
}
@@ -265,8 +264,7 @@ public:
WriteUnitOfWork wunit(&_opCtx);
for (int j = 0; j < 2; j++) {
auto doc = BSON("_id" << j);
- ASSERT_OK(rs->insertRecord(
- &_opCtx, doc.objdata(), doc.objsize(), Timestamp(), /*enforceQuota*/ false));
+ ASSERT_OK(rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), Timestamp()));
}
wunit.commit();
}
@@ -330,8 +328,7 @@ public:
{
WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1 << "a" << 9);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
@@ -382,8 +379,7 @@ public:
{
WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 9);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -396,8 +392,7 @@ public:
{
WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -414,9 +409,7 @@ public:
rs->deleteRecord(&_opCtx, id1);
auto doc = BSON("_id" << 3);
ASSERT_OK(
- rs->insertRecord(
- &_opCtx, doc.objdata(), doc.objsize(), Timestamp(), /*enforceQuota*/ false)
- .getStatus());
+ rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), Timestamp()).getStatus());
wunit.commit();
}
@@ -491,8 +484,8 @@ public:
// Update a document's indexed field without updating the index.
{
WriteUnitOfWork wunit(&_opCtx);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc1_b.objdata(), doc1_b.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus =
+ rs->updateRecord(&_opCtx, id1, doc1_b.objdata(), doc1_b.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -505,8 +498,8 @@ public:
// Index validation should still be valid.
{
WriteUnitOfWork wunit(&_opCtx);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc1_c.objdata(), doc1_c.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus =
+ rs->updateRecord(&_opCtx, id1, doc1_c.objdata(), doc1_c.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -574,8 +567,7 @@ public:
{
WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 2 << "a" << 3);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -648,8 +640,7 @@ public:
{
WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}
@@ -811,8 +802,7 @@ public:
{
WriteUnitOfWork wunit(&_opCtx);
auto doc = BSON("_id" << 1 << "a" << 1 << "b" << 3);
- auto updateStatus = rs->updateRecord(
- &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL);
+ auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL);
ASSERT_OK(updateStatus);
wunit.commit();
}