diff options
Diffstat (limited to 'src/mongo')
53 files changed, 189 insertions, 301 deletions
diff --git a/src/mongo/db/catalog/collection.h b/src/mongo/db/catalog/collection.h index 018cfb3f51a..2eec1337022 100644 --- a/src/mongo/db/catalog/collection.h +++ b/src/mongo/db/catalog/collection.h @@ -228,13 +228,11 @@ public: std::vector<InsertStatement>::const_iterator begin, std::vector<InsertStatement>::const_iterator end, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate) = 0; virtual Status insertDocument(OperationContext* opCtx, const InsertStatement& doc, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate) = 0; virtual Status insertDocumentsForOplog(OperationContext* opCtx, @@ -244,14 +242,12 @@ public: virtual Status insertDocument(OperationContext* opCtx, const BSONObj& doc, - const std::vector<MultiIndexBlock*>& indexBlocks, - bool enforceQuota) = 0; + const std::vector<MultiIndexBlock*>& indexBlocks) = 0; virtual RecordId updateDocument(OperationContext* opCtx, const RecordId& oldLocation, const Snapshotted<BSONObj>& oldDoc, const BSONObj& newDoc, - bool enforceQuota, bool indexesAffected, OpDebug* opDebug, OplogUpdateEntryArgs* args) = 0; @@ -468,9 +464,8 @@ public: const std::vector<InsertStatement>::const_iterator begin, const std::vector<InsertStatement>::const_iterator end, OpDebug* const opDebug, - const bool enforceQuota, const bool fromMigrate = false) { - return this->_impl().insertDocuments(opCtx, begin, end, opDebug, enforceQuota, fromMigrate); + return this->_impl().insertDocuments(opCtx, begin, end, opDebug, fromMigrate); } /** @@ -478,14 +473,12 @@ public: * i.e. will not add an _id field for documents that are missing it * * 'opDebug' Optional argument. When not null, will be used to record operation statistics. - * 'enforceQuota' If false, quotas will be ignored. */ inline Status insertDocument(OperationContext* const opCtx, const InsertStatement& doc, OpDebug* const opDebug, - const bool enforceQuota, const bool fromMigrate = false) { - return this->_impl().insertDocument(opCtx, doc, opDebug, enforceQuota, fromMigrate); + return this->_impl().insertDocument(opCtx, doc, opDebug, fromMigrate); } /** @@ -506,9 +499,8 @@ public: */ inline Status insertDocument(OperationContext* const opCtx, const BSONObj& doc, - const std::vector<MultiIndexBlock*>& indexBlocks, - const bool enforceQuota) { - return this->_impl().insertDocument(opCtx, doc, indexBlocks, enforceQuota); + const std::vector<MultiIndexBlock*>& indexBlocks) { + return this->_impl().insertDocument(opCtx, doc, indexBlocks); } /** @@ -524,12 +516,11 @@ public: const RecordId& oldLocation, const Snapshotted<BSONObj>& oldDoc, const BSONObj& newDoc, - const bool enforceQuota, const bool indexesAffected, OpDebug* const opDebug, OplogUpdateEntryArgs* const args) { return this->_impl().updateDocument( - opCtx, oldLocation, oldDoc, newDoc, enforceQuota, indexesAffected, opDebug, args); + opCtx, oldLocation, oldDoc, newDoc, indexesAffected, opDebug, args); } inline bool updateWithDamagesSupported() const { diff --git a/src/mongo/db/catalog/collection_impl.cpp b/src/mongo/db/catalog/collection_impl.cpp index 67e689bf569..1b5b0a5ad33 100644 --- a/src/mongo/db/catalog/collection_impl.cpp +++ b/src/mongo/db/catalog/collection_impl.cpp @@ -329,7 +329,6 @@ Status CollectionImpl::insertDocuments(OperationContext* opCtx, const vector<InsertStatement>::const_iterator begin, const vector<InsertStatement>::const_iterator end, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate) { MONGO_FAIL_POINT_BLOCK(failCollectionInserts, extraData) { @@ -363,7 +362,7 @@ Status CollectionImpl::insertDocuments(OperationContext* opCtx, const SnapshotId sid = opCtx->recoveryUnit()->getSnapshotId(); - Status status = _insertDocuments(opCtx, begin, end, enforceQuota, opDebug); + Status status = _insertDocuments(opCtx, begin, end, opDebug); if (!status.isOK()) return status; invariant(sid == opCtx->recoveryUnit()->getSnapshotId()); @@ -380,17 +379,15 @@ Status CollectionImpl::insertDocuments(OperationContext* opCtx, Status CollectionImpl::insertDocument(OperationContext* opCtx, const InsertStatement& docToInsert, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate) { vector<InsertStatement> docs; docs.push_back(docToInsert); - return insertDocuments(opCtx, docs.begin(), docs.end(), opDebug, enforceQuota, fromMigrate); + return insertDocuments(opCtx, docs.begin(), docs.end(), opDebug, fromMigrate); } Status CollectionImpl::insertDocument(OperationContext* opCtx, const BSONObj& doc, - const std::vector<MultiIndexBlock*>& indexBlocks, - bool enforceQuota) { + const std::vector<MultiIndexBlock*>& indexBlocks) { MONGO_FAIL_POINT_BLOCK(failCollectionInserts, extraData) { const BSONObj& data = extraData.getData(); @@ -415,8 +412,8 @@ Status CollectionImpl::insertDocument(OperationContext* opCtx, // TODO SERVER-30638: using timestamp 0 for these inserts, which are non-oplog so we don't yet // care about their correct timestamps. - StatusWith<RecordId> loc = _recordStore->insertRecord( - opCtx, doc.objdata(), doc.objsize(), Timestamp(), _enforceQuota(enforceQuota)); + StatusWith<RecordId> loc = + _recordStore->insertRecord(opCtx, doc.objdata(), doc.objsize(), Timestamp()); if (!loc.isOK()) return loc.getStatus(); @@ -450,7 +447,6 @@ Status CollectionImpl::insertDocument(OperationContext* opCtx, Status CollectionImpl::_insertDocuments(OperationContext* opCtx, const vector<InsertStatement>::const_iterator begin, const vector<InsertStatement>::const_iterator end, - bool enforceQuota, OpDebug* opDebug) { dassert(opCtx->lockState()->isCollectionLockedForMode(ns().toString(), MODE_IX)); @@ -483,8 +479,7 @@ Status CollectionImpl::_insertDocuments(OperationContext* opCtx, Timestamp timestamp = Timestamp(it->oplogSlot.opTime.getTimestamp()); timestamps.push_back(timestamp); } - Status status = - _recordStore->insertRecords(opCtx, &records, ×tamps, _enforceQuota(enforceQuota)); + Status status = _recordStore->insertRecords(opCtx, &records, ×tamps); if (!status.isOK()) return status; @@ -584,7 +579,6 @@ RecordId CollectionImpl::updateDocument(OperationContext* opCtx, const RecordId& oldLocation, const Snapshotted<BSONObj>& oldDoc, const BSONObj& newDoc, - bool enforceQuota, bool indexesAffected, OpDebug* opDebug, OplogUpdateEntryArgs* args) { @@ -664,8 +658,8 @@ RecordId CollectionImpl::updateDocument(OperationContext* opCtx, args->preImageDoc = oldDoc.value().getOwned(); - Status updateStatus = _recordStore->updateRecord( - opCtx, oldLocation, newDoc.objdata(), newDoc.objsize(), _enforceQuota(enforceQuota), this); + Status updateStatus = + _recordStore->updateRecord(opCtx, oldLocation, newDoc.objdata(), newDoc.objsize(), this); // Update each index with each respective UpdateTicket. if (indexesAffected) { @@ -733,19 +727,6 @@ StatusWith<RecordData> CollectionImpl::updateDocumentWithDamages( return newRecStatus; } -bool CollectionImpl::_enforceQuota(bool userEnforeQuota) const { - if (!userEnforeQuota) - return false; - - if (_ns.db() == "local") - return false; - - if (_ns.isSpecial()) - return false; - - return true; -} - bool CollectionImpl::isCapped() const { return _cappedNotifier.get(); } diff --git a/src/mongo/db/catalog/collection_impl.h b/src/mongo/db/catalog/collection_impl.h index 1530760be64..5944e19ba78 100644 --- a/src/mongo/db/catalog/collection_impl.h +++ b/src/mongo/db/catalog/collection_impl.h @@ -163,7 +163,6 @@ public: std::vector<InsertStatement>::const_iterator begin, std::vector<InsertStatement>::const_iterator end, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate = false) final; /** @@ -171,12 +170,10 @@ public: * i.e. will not add an _id field for documents that are missing it * * 'opDebug' Optional argument. When not null, will be used to record operation statistics. - * 'enforceQuota' If false, quotas will be ignored. */ Status insertDocument(OperationContext* opCtx, const InsertStatement& doc, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate = false) final; /** @@ -195,8 +192,7 @@ public: */ Status insertDocument(OperationContext* opCtx, const BSONObj& doc, - const std::vector<MultiIndexBlock*>& indexBlocks, - bool enforceQuota) final; + const std::vector<MultiIndexBlock*>& indexBlocks) final; /** * Updates the document @ oldLocation with newDoc. @@ -211,7 +207,6 @@ public: const RecordId& oldLocation, const Snapshotted<BSONObj>& oldDoc, const BSONObj& newDoc, - bool enforceQuota, bool indexesAffected, OpDebug* opDebug, OplogUpdateEntryArgs* args) final; @@ -392,17 +387,13 @@ private: * - some user error checks * - adjust padding */ - Status _insertDocument(OperationContext* opCtx, const BSONObj& doc, bool enforceQuota); + Status _insertDocument(OperationContext* opCtx, const BSONObj& doc); Status _insertDocuments(OperationContext* opCtx, std::vector<InsertStatement>::const_iterator begin, std::vector<InsertStatement>::const_iterator end, - bool enforceQuota, OpDebug* opDebug); - - bool _enforceQuota(bool userEnforeQuota) const; - int _magic; const NamespaceString _ns; diff --git a/src/mongo/db/catalog/collection_mock.h b/src/mongo/db/catalog/collection_mock.h index ae4a816f63f..ba0034df7ff 100644 --- a/src/mongo/db/catalog/collection_mock.h +++ b/src/mongo/db/catalog/collection_mock.h @@ -137,7 +137,6 @@ public: std::vector<InsertStatement>::const_iterator begin, std::vector<InsertStatement>::const_iterator end, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate) { std::abort(); } @@ -145,7 +144,6 @@ public: Status insertDocument(OperationContext* opCtx, const InsertStatement& doc, OpDebug* opDebug, - bool enforceQuota, bool fromMigrate) { std::abort(); } @@ -159,8 +157,7 @@ public: Status insertDocument(OperationContext* opCtx, const BSONObj& doc, - const std::vector<MultiIndexBlock*>& indexBlocks, - bool enforceQuota) { + const std::vector<MultiIndexBlock*>& indexBlocks) { std::abort(); } @@ -168,7 +165,6 @@ public: const RecordId& oldLocation, const Snapshotted<BSONObj>& oldDoc, const BSONObj& newDoc, - bool enforceQuota, bool indexesAffected, OpDebug* opDebug, OplogUpdateEntryArgs* args) { diff --git a/src/mongo/db/catalog/rename_collection_test.cpp b/src/mongo/db/catalog/rename_collection_test.cpp index 092b1328406..943fd3506ca 100644 --- a/src/mongo/db/catalog/rename_collection_test.cpp +++ b/src/mongo/db/catalog/rename_collection_test.cpp @@ -426,8 +426,7 @@ void _insertDocument(OperationContext* opCtx, const NamespaceString& nss, const WriteUnitOfWork wuow(opCtx); OpDebug* const opDebug = nullptr; - bool enforceQuota = true; - ASSERT_OK(collection->insertDocument(opCtx, InsertStatement(doc), opDebug, enforceQuota)); + ASSERT_OK(collection->insertDocument(opCtx, InsertStatement(doc), opDebug)); wuow.commit(); }); } diff --git a/src/mongo/db/commands/mr.cpp b/src/mongo/db/commands/mr.cpp index dee774f7af2..29efeb7f3a9 100644 --- a/src/mongo/db/commands/mr.cpp +++ b/src/mongo/db/commands/mr.cpp @@ -843,7 +843,7 @@ void State::_insertToInc(BSONObj& o) { // TODO: Consider whether to pass OpDebug for stats tracking under SERVER-23261. OpDebug* const nullOpDebug = nullptr; uassertStatusOK(autoColl.getCollection()->insertDocument( - _opCtx, InsertStatement(o), nullOpDebug, true, false)); + _opCtx, InsertStatement(o), nullOpDebug, false)); wuow.commit(); }); } diff --git a/src/mongo/db/exec/update.cpp b/src/mongo/db/exec/update.cpp index fabf092fa08..964b70866a7 100644 --- a/src/mongo/db/exec/update.cpp +++ b/src/mongo/db/exec/update.cpp @@ -334,7 +334,6 @@ BSONObj UpdateStage::transformAndUpdate(const Snapshotted<BSONObj>& oldObj, Reco recordId, oldObj, newObj, - true, driver->modsAffectIndices(), _params.opDebug, &args); @@ -469,11 +468,9 @@ void UpdateStage::doInsert() { writeConflictRetry(getOpCtx(), "upsert", _collection->ns().ns(), [&] { WriteUnitOfWork wunit(getOpCtx()); invariant(_collection); - const bool enforceQuota = !request->isGod(); uassertStatusOK(_collection->insertDocument(getOpCtx(), InsertStatement(request->getStmtId(), newObj), _params.opDebug, - enforceQuota, request->isFromMigration())); // Technically, we should save/restore state here, but since we are going to return diff --git a/src/mongo/db/ops/write_ops_exec.cpp b/src/mongo/db/ops/write_ops_exec.cpp index 9718457f0be..1c6f944d75b 100644 --- a/src/mongo/db/ops/write_ops_exec.cpp +++ b/src/mongo/db/ops/write_ops_exec.cpp @@ -354,8 +354,7 @@ void insertDocuments(OperationContext* opCtx, } } - uassertStatusOK(collection->insertDocuments( - opCtx, begin, end, &CurOp::get(opCtx)->debug(), /*enforceQuota*/ true)); + uassertStatusOK(collection->insertDocuments(opCtx, begin, end, &CurOp::get(opCtx)->debug())); wuow.commit(); } diff --git a/src/mongo/db/repl/collection_bulk_loader_impl.cpp b/src/mongo/db/repl/collection_bulk_loader_impl.cpp index b0ccc6dfcc9..276aaba8c11 100644 --- a/src/mongo/db/repl/collection_bulk_loader_impl.cpp +++ b/src/mongo/db/repl/collection_bulk_loader_impl.cpp @@ -126,7 +126,7 @@ Status CollectionBulkLoaderImpl::insertDocuments(const std::vector<BSONObj>::con // This flavor of insertDocument will not update any pre-existing indexes, // only the indexers passed in. const auto status = _autoColl->getCollection()->insertDocument( - _opCtx.get(), *iter, indexers, false); + _opCtx.get(), *iter, indexers); if (!status.isOK()) { return status; } @@ -134,7 +134,7 @@ Status CollectionBulkLoaderImpl::insertDocuments(const std::vector<BSONObj>::con // For capped collections, we use regular insertDocument, which will update // pre-existing indexes. const auto status = _autoColl->getCollection()->insertDocument( - _opCtx.get(), InsertStatement(*iter), nullptr, false); + _opCtx.get(), InsertStatement(*iter), nullptr); if (!status.isOK()) { return status; } diff --git a/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp b/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp index e173b709602..d03b88b1253 100644 --- a/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp +++ b/src/mongo/db/repl/mock_repl_coord_server_fixture.cpp @@ -102,7 +102,6 @@ void MockReplCoordServerFixture::insertOplogEntry(const repl::OplogEntry& entry) auto status = coll->insertDocument(opCtx(), InsertStatement(entry.toBSON()), &CurOp::get(opCtx())->debug(), - /* enforceQuota */ false, /* fromMigrate */ false); ASSERT_OK(status); } diff --git a/src/mongo/db/s/collection_sharding_state_test.cpp b/src/mongo/db/s/collection_sharding_state_test.cpp index 6934bddb5e3..3fda81a7d21 100644 --- a/src/mongo/db/s/collection_sharding_state_test.cpp +++ b/src/mongo/db/s/collection_sharding_state_test.cpp @@ -95,7 +95,7 @@ TEST_F(CollectionShardingStateTest, GlobalInitDoesntGetCalledIfWriteAborts) { WriteUnitOfWork wuow(operationContext()); ASSERT_OK(autoColl.getCollection()->insertDocument( - operationContext(), shardIdentity.toShardIdentityDocument(), {}, false)); + operationContext(), shardIdentity.toShardIdentityDocument(), {})); ASSERT_EQ(0, getInitCallCount()); } diff --git a/src/mongo/db/session.cpp b/src/mongo/db/session.cpp index df347194ea7..26f621facbb 100644 --- a/src/mongo/db/session.cpp +++ b/src/mongo/db/session.cpp @@ -231,7 +231,7 @@ void updateSessionEntry(OperationContext* opCtx, const UpdateRequest& updateRequ if (recordId.isNull()) { // Upsert case. auto status = collection->insertDocument( - opCtx, InsertStatement(updateRequest.getUpdates()), nullptr, true, false); + opCtx, InsertStatement(updateRequest.getUpdates()), nullptr, false); if (status == ErrorCodes::DuplicateKey) { throw WriteConflictException(); @@ -266,7 +266,6 @@ void updateSessionEntry(OperationContext* opCtx, const UpdateRequest& updateRequ recordId, Snapshotted<BSONObj>(startingSnapshotId, originalDoc), updateRequest.getUpdates(), - true, // enforceQuota false, // indexesAffected = false because _id is the only index nullptr, &args); diff --git a/src/mongo/db/storage/devnull/devnull_kv_engine.cpp b/src/mongo/db/storage/devnull/devnull_kv_engine.cpp index 7f0b16fa26b..bd48a8cfddd 100644 --- a/src/mongo/db/storage/devnull/devnull_kv_engine.cpp +++ b/src/mongo/db/storage/devnull/devnull_kv_engine.cpp @@ -101,8 +101,10 @@ public: virtual void deleteRecord(OperationContext* opCtx, const RecordId& dl) {} - virtual StatusWith<RecordId> insertRecord( - OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota) { + virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx, + const char* data, + int len, + Timestamp) { _numInserts++; return StatusWith<RecordId>(RecordId(6, 4)); } @@ -125,7 +127,6 @@ public: const RecordId& oldLocation, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier) { return Status::OK(); } diff --git a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp index f2ed2f99de2..aa69447703a 100644 --- a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp +++ b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.cpp @@ -397,8 +397,10 @@ StatusWith<RecordId> EphemeralForTestRecordStore::extractAndCheckLocForOplog(con return status; } -StatusWith<RecordId> EphemeralForTestRecordStore::insertRecord( - OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota) { +StatusWith<RecordId> EphemeralForTestRecordStore::insertRecord(OperationContext* opCtx, + const char* data, + int len, + Timestamp) { if (_isCapped && len > _cappedMaxSize) { // We use dataSize for capped rollover and we don't want to delete everything if we know // this won't fit. @@ -473,7 +475,6 @@ Status EphemeralForTestRecordStore::updateRecord(OperationContext* opCtx, const RecordId& loc, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier) { stdx::lock_guard<stdx::recursive_mutex> lock(_data->recordsMutex); EphemeralForTestRecord* oldRecord = recordFor(loc); @@ -620,9 +621,7 @@ Status EphemeralForTestRecordStore::touch(OperationContext* opCtx, BSONObjBuilde return Status::OK(); } -void EphemeralForTestRecordStore::increaseStorageSize(OperationContext* opCtx, - int size, - bool enforceQuota) { +void EphemeralForTestRecordStore::increaseStorageSize(OperationContext* opCtx, int size) { // unclear what this would mean for this class. For now, just error if called. invariant(!"increaseStorageSize not yet implemented"); } diff --git a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h index 5e0d7ba1e5a..1b39400cc51 100644 --- a/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h +++ b/src/mongo/db/storage/ephemeral_for_test/ephemeral_for_test_record_store.h @@ -66,8 +66,10 @@ public: virtual void deleteRecord(OperationContext* opCtx, const RecordId& dl); - virtual StatusWith<RecordId> insertRecord( - OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota); + virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx, + const char* data, + int len, + Timestamp); virtual Status insertRecordsWithDocWriter(OperationContext* opCtx, const DocWriter* const* docs, @@ -79,7 +81,6 @@ public: const RecordId& oldLocation, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier); virtual bool updateWithDamagesSupported() const; @@ -109,7 +110,7 @@ public: virtual Status touch(OperationContext* opCtx, BSONObjBuilder* output) const; - virtual void increaseStorageSize(OperationContext* opCtx, int size, bool enforceQuota); + virtual void increaseStorageSize(OperationContext* opCtx, int size); virtual int64_t storageSize(OperationContext* opCtx, BSONObjBuilder* extraInfo = NULL, diff --git a/src/mongo/db/storage/kv/kv_catalog.cpp b/src/mongo/db/storage/kv/kv_catalog.cpp index 283b0090a7e..89ad94b6544 100644 --- a/src/mongo/db/storage/kv/kv_catalog.cpp +++ b/src/mongo/db/storage/kv/kv_catalog.cpp @@ -306,17 +306,14 @@ void KVCatalog::FeatureTracker::putInfo(OperationContext* opCtx, const FeatureBi if (_rid.isNull()) { // This is the first time a feature is being marked as in-use or not in-use, so we must // insert the feature document rather than update it. - const bool enforceQuota = false; // TODO SERVER-30638: using timestamp 0 for these inserts - auto rid = _catalog->_rs->insertRecord( - opCtx, obj.objdata(), obj.objsize(), Timestamp(), enforceQuota); + auto rid = _catalog->_rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), Timestamp()); fassert(40113, rid.getStatus()); _rid = rid.getValue(); } else { - const bool enforceQuota = false; UpdateNotifier* notifier = nullptr; - auto status = _catalog->_rs->updateRecord( - opCtx, _rid, obj.objdata(), obj.objsize(), enforceQuota, notifier); + auto status = + _catalog->_rs->updateRecord(opCtx, _rid, obj.objdata(), obj.objsize(), notifier); fassert(40114, status); } } @@ -425,10 +422,8 @@ Status KVCatalog::newCollection(OperationContext* opCtx, b.append("md", md.toBSON()); obj = b.obj(); } - const bool enforceQuota = false; // TODO SERVER-30638: using timestamp 0 for these inserts. - StatusWith<RecordId> res = - _rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), Timestamp(), enforceQuota); + StatusWith<RecordId> res = _rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), Timestamp()); if (!res.isOK()) return res.getStatus(); @@ -524,7 +519,7 @@ void KVCatalog::putMetaData(OperationContext* opCtx, } LOG(3) << "recording new metadata: " << obj; - Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), false, NULL); + Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), NULL); fassert(28521, status.isOK()); } @@ -549,7 +544,7 @@ Status KVCatalog::renameCollection(OperationContext* opCtx, b.appendElementsUnique(old); BSONObj obj = b.obj(); - Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), false, NULL); + Status status = _rs->updateRecord(opCtx, loc, obj.objdata(), obj.objsize(), NULL); fassert(28522, status.isOK()); } diff --git a/src/mongo/db/storage/kv/kv_engine_test_harness.cpp b/src/mongo/db/storage/kv/kv_engine_test_harness.cpp index e0eebfaa73d..6b9b671fd37 100644 --- a/src/mongo/db/storage/kv/kv_engine_test_harness.cpp +++ b/src/mongo/db/storage/kv/kv_engine_test_harness.cpp @@ -75,7 +75,7 @@ TEST(KVEngineTestHarness, SimpleRS1) { { MyOperationContext opCtx(engine); WriteUnitOfWork uow(&opCtx); - StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -115,7 +115,7 @@ TEST(KVEngineTestHarness, Restart1) { { MyOperationContext opCtx(engine); WriteUnitOfWork uow(&opCtx); - StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(&opCtx, "abc", 4, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp b/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp index d2c24f3b44e..90c51d942c8 100644 --- a/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp +++ b/src/mongo/db/storage/kv/kv_engine_test_timestamps.cpp @@ -101,7 +101,7 @@ public: } RecordId insertRecord(OperationContext* opCtx, std::string contents = "abcd") { - auto id = rs->insertRecord(opCtx, contents.c_str(), contents.length() + 1, _counter, false); + auto id = rs->insertRecord(opCtx, contents.c_str(), contents.length() + 1, _counter); ASSERT_OK(id); return id.getValue(); } @@ -118,8 +118,7 @@ public: auto op = makeOperation(); WriteUnitOfWork wuow(op); ASSERT_OK(op->recoveryUnit()->setTimestamp(_counter)); - ASSERT_OK( - rs->updateRecord(op, id, contents.c_str(), contents.length() + 1, false, nullptr)); + ASSERT_OK(rs->updateRecord(op, id, contents.c_str(), contents.length() + 1, nullptr)); wuow.commit(); } diff --git a/src/mongo/db/storage/mobile/mobile_record_store.cpp b/src/mongo/db/storage/mobile/mobile_record_store.cpp index 4e62e87b29b..2e1aed4c5e2 100644 --- a/src/mongo/db/storage/mobile/mobile_record_store.cpp +++ b/src/mongo/db/storage/mobile/mobile_record_store.cpp @@ -291,8 +291,10 @@ void MobileRecordStore::deleteRecord(OperationContext* opCtx, const RecordId& re deleteStmt.step(SQLITE_DONE); } -StatusWith<RecordId> MobileRecordStore::insertRecord( - OperationContext* opCtx, const char* data, int len, Timestamp, bool enforceQuota) { +StatusWith<RecordId> MobileRecordStore::insertRecord(OperationContext* opCtx, + const char* data, + int len, + Timestamp) { // Inserts record into SQLite table (or replaces if duplicate record id). MobileSession* session = MobileRecoveryUnit::get(opCtx)->getSession(opCtx, false); @@ -326,7 +328,7 @@ Status MobileRecordStore::insertRecordsWithDocWriter(OperationContext* opCtx, for (size_t i = 0; i < nDocs; i++) { docs[i]->writeDocument(pos); size_t docLen = docs[i]->documentSize(); - StatusWith<RecordId> res = insertRecord(opCtx, pos, docLen, timestamps[i], true); + StatusWith<RecordId> res = insertRecord(opCtx, pos, docLen, timestamps[i]); idsOut[i] = res.getValue(); pos += docLen; } @@ -338,7 +340,6 @@ Status MobileRecordStore::updateRecord(OperationContext* opCtx, const RecordId& recId, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier) { MobileSession* session = MobileRecoveryUnit::get(opCtx)->getSession(opCtx, false); std::string dataSizeQuery = diff --git a/src/mongo/db/storage/mobile/mobile_record_store.h b/src/mongo/db/storage/mobile/mobile_record_store.h index d0771f0317a..ec8ab13aa2e 100644 --- a/src/mongo/db/storage/mobile/mobile_record_store.h +++ b/src/mongo/db/storage/mobile/mobile_record_store.h @@ -63,8 +63,7 @@ public: StatusWith<RecordId> insertRecord(OperationContext* opCtx, const char* data, int len, - Timestamp timestamp, - bool enforceQuota) override; + Timestamp timestamp) override; Status insertRecordsWithDocWriter(OperationContext* opCtx, const DocWriter* const* docs, @@ -76,7 +75,6 @@ public: const RecordId& oldLocation, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier) override; bool updateWithDamagesSupported() const override; diff --git a/src/mongo/db/storage/record_store.h b/src/mongo/db/storage/record_store.h index e09ef89980b..32f7b41cf7b 100644 --- a/src/mongo/db/storage/record_store.h +++ b/src/mongo/db/storage/record_store.h @@ -372,20 +372,15 @@ public: virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx, const char* data, int len, - Timestamp timestamp, - bool enforceQuota) = 0; + Timestamp timestamp) = 0; virtual Status insertRecords(OperationContext* opCtx, std::vector<Record>* records, - std::vector<Timestamp>* timestamps, - bool enforceQuota) { + std::vector<Timestamp>* timestamps) { int index = 0; for (auto& record : *records) { - StatusWith<RecordId> res = insertRecord(opCtx, - record.data.data(), - record.data.size(), - (*timestamps)[index++], - enforceQuota); + StatusWith<RecordId> res = + insertRecord(opCtx, record.data.data(), record.data.size(), (*timestamps)[index++]); if (!res.isOK()) return res.getStatus(); @@ -431,7 +426,6 @@ public: const RecordId& oldLocation, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier) = 0; /** diff --git a/src/mongo/db/storage/record_store_test_capped_visibility.cpp b/src/mongo/db/storage/record_store_test_capped_visibility.cpp index 6b778dd2dce..e66544ce0c4 100644 --- a/src/mongo/db/storage/record_store_test_capped_visibility.cpp +++ b/src/mongo/db/storage/record_store_test_capped_visibility.cpp @@ -42,7 +42,7 @@ namespace { RecordId doInsert(unowned_ptr<OperationContext> opCtx, unowned_ptr<RecordStore> rs) { static char zeros[16]; - return uassertStatusOK(rs->insertRecord(opCtx, zeros, sizeof(zeros), Timestamp(), false)); + return uassertStatusOK(rs->insertRecord(opCtx, zeros, sizeof(zeros), Timestamp())); } // macro to keep assert line numbers correct. diff --git a/src/mongo/db/storage/record_store_test_datafor.cpp b/src/mongo/db/storage/record_store_test_datafor.cpp index 672fa03e017..07f94f9ed9f 100644 --- a/src/mongo/db/storage/record_store_test_datafor.cpp +++ b/src/mongo/db/storage/record_store_test_datafor.cpp @@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, DataFor) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -107,7 +107,7 @@ TEST(RecordStoreTestHarness, DataForMultiple) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_datasize.cpp b/src/mongo/db/storage/record_store_test_datasize.cpp index 3dad519623b..b57cebf06bd 100644 --- a/src/mongo/db/storage/record_store_test_datasize.cpp +++ b/src/mongo/db/storage/record_store_test_datasize.cpp @@ -80,7 +80,7 @@ TEST(RecordStoreTestHarness, DataSizeNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } diff --git a/src/mongo/db/storage/record_store_test_deleterecord.cpp b/src/mongo/db/storage/record_store_test_deleterecord.cpp index 0d630c0d0be..90d8f274c86 100644 --- a/src/mongo/db/storage/record_store_test_deleterecord.cpp +++ b/src/mongo/db/storage/record_store_test_deleterecord.cpp @@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, DeleteRecord) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -111,7 +111,7 @@ TEST(RecordStoreTestHarness, DeleteMultipleRecords) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_harness.cpp b/src/mongo/db/storage/record_store_test_harness.cpp index 2287e03052c..4ee95575612 100644 --- a/src/mongo/db/storage/record_store_test_harness.cpp +++ b/src/mongo/db/storage/record_store_test_harness.cpp @@ -60,7 +60,7 @@ TEST(RecordStoreTestHarness, Simple1) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc1 = res.getValue(); uow.commit(); @@ -87,7 +87,7 @@ TEST(RecordStoreTestHarness, Simple1) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } @@ -160,7 +160,7 @@ TEST(RecordStoreTestHarness, Delete1) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -205,9 +205,9 @@ TEST(RecordStoreTestHarness, Delete2) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); - res = rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false); + res = rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -248,7 +248,7 @@ TEST(RecordStoreTestHarness, Update1) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s1.c_str(), s1.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), s1.c_str(), s1.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -264,8 +264,7 @@ TEST(RecordStoreTestHarness, Update1) { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); { WriteUnitOfWork uow(opCtx.get()); - Status status = - rs->updateRecord(opCtx.get(), loc, s2.c_str(), s2.size() + 1, false, NULL); + Status status = rs->updateRecord(opCtx.get(), loc, s2.c_str(), s2.size() + 1, NULL); ASSERT_OK(status); uow.commit(); @@ -296,7 +295,7 @@ TEST(RecordStoreTestHarness, UpdateInPlace1) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s1Rec.data(), s1Rec.size(), Timestamp(), false); + rs->insertRecord(opCtx.get(), s1Rec.data(), s1Rec.size(), Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -350,7 +349,7 @@ TEST(RecordStoreTestHarness, Truncate1) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -400,7 +399,7 @@ TEST(RecordStoreTestHarness, Cursor1) { WriteUnitOfWork uow(opCtx.get()); for (int i = 0; i < N; i++) { string s = str::stream() << "eliot" << i; - ASSERT_OK(rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp(), false) + ASSERT_OK(rs->insertRecord(opCtx.get(), s.c_str(), s.size() + 1, Timestamp()) .getStatus()); } uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_insertrecord.cpp b/src/mongo/db/storage/record_store_test_insertrecord.cpp index adcf2fc2300..0cef303a730 100644 --- a/src/mongo/db/storage/record_store_test_insertrecord.cpp +++ b/src/mongo/db/storage/record_store_test_insertrecord.cpp @@ -61,7 +61,7 @@ TEST(RecordStoreTestHarness, InsertRecord) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -96,7 +96,7 @@ TEST(RecordStoreTestHarness, InsertMultipleRecords) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_manyiter.cpp b/src/mongo/db/storage/record_store_test_manyiter.cpp index 5b477f0d27a..2080f1ca073 100644 --- a/src/mongo/db/storage/record_store_test_manyiter.cpp +++ b/src/mongo/db/storage/record_store_test_manyiter.cpp @@ -87,7 +87,7 @@ TEST(RecordStoreTestHarness, GetManyIteratorsNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_randomiter.cpp b/src/mongo/db/storage/record_store_test_randomiter.cpp index 0644b07a83b..3cdeec17f2d 100644 --- a/src/mongo/db/storage/record_store_test_randomiter.cpp +++ b/src/mongo/db/storage/record_store_test_randomiter.cpp @@ -87,7 +87,7 @@ TEST(RecordStoreTestHarness, GetRandomIteratorNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); @@ -141,8 +141,7 @@ TEST(RecordStoreTestHarness, GetRandomIteratorSingleton) { { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), "some data", 10, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "some data", 10, Timestamp()); ASSERT_OK(res.getStatus()); idToRetrieve = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_recorditer.cpp b/src/mongo/db/storage/record_store_test_recorditer.cpp index 37f6d284258..968dfbcb760 100644 --- a/src/mongo/db/storage/record_store_test_recorditer.cpp +++ b/src/mongo/db/storage/record_store_test_recorditer.cpp @@ -71,7 +71,7 @@ TEST(RecordStoreTestHarness, IterateOverMultipleRecords) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; @@ -122,7 +122,7 @@ TEST(RecordStoreTestHarness, IterateOverMultipleRecordsReversed) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; @@ -173,7 +173,7 @@ TEST(RecordStoreTestHarness, IterateStartFromMiddle) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; @@ -225,7 +225,7 @@ TEST(RecordStoreTestHarness, IterateStartFromMiddleReversed) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; @@ -278,7 +278,7 @@ TEST(RecordStoreTestHarness, RecordIteratorEOF) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; @@ -315,7 +315,7 @@ TEST(RecordStoreTestHarness, RecordIteratorEOF) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); @@ -349,7 +349,7 @@ TEST(RecordStoreTestHarness, RecordIteratorSaveRestore) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; @@ -406,7 +406,7 @@ TEST(RecordStoreTestHarness, SeekAfterEofAndContinue) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); datas[i] = data; diff --git a/src/mongo/db/storage/record_store_test_repairiter.cpp b/src/mongo/db/storage/record_store_test_repairiter.cpp index f5397012cf8..a84e7dcab4d 100644 --- a/src/mongo/db/storage/record_store_test_repairiter.cpp +++ b/src/mongo/db/storage/record_store_test_repairiter.cpp @@ -89,7 +89,7 @@ TEST(RecordStoreTestHarness, GetIteratorForRepairNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); @@ -136,8 +136,7 @@ TEST(RecordStoreTestHarness, GetIteratorForRepairInvalidateSingleton) { { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), "some data", 10, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "some data", 10, Timestamp()); ASSERT_OK(res.getStatus()); idToInvalidate = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_storagesize.cpp b/src/mongo/db/storage/record_store_test_storagesize.cpp index 521d879cfac..c5bd0bcf7cc 100644 --- a/src/mongo/db/storage/record_store_test_storagesize.cpp +++ b/src/mongo/db/storage/record_store_test_storagesize.cpp @@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, StorageSizeNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } diff --git a/src/mongo/db/storage/record_store_test_touch.cpp b/src/mongo/db/storage/record_store_test_touch.cpp index eae3e0046b0..dfd541250bb 100644 --- a/src/mongo/db/storage/record_store_test_touch.cpp +++ b/src/mongo/db/storage/record_store_test_touch.cpp @@ -85,7 +85,7 @@ TEST(RecordStoreTestHarness, TouchNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } @@ -149,7 +149,7 @@ TEST(RecordStoreTestHarness, TouchNonEmptyWithNullStats) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } diff --git a/src/mongo/db/storage/record_store_test_truncate.cpp b/src/mongo/db/storage/record_store_test_truncate.cpp index 7fda629b4f9..aac72c5caf8 100644 --- a/src/mongo/db/storage/record_store_test_truncate.cpp +++ b/src/mongo/db/storage/record_store_test_truncate.cpp @@ -89,7 +89,7 @@ TEST(RecordStoreTestHarness, TruncateNonEmpty) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } diff --git a/src/mongo/db/storage/record_store_test_updaterecord.cpp b/src/mongo/db/storage/record_store_test_updaterecord.cpp index f9c9e33c375..c6b96def16d 100644 --- a/src/mongo/db/storage/record_store_test_updaterecord.cpp +++ b/src/mongo/db/storage/record_store_test_updaterecord.cpp @@ -63,7 +63,7 @@ TEST(RecordStoreTestHarness, UpdateRecord) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -80,8 +80,7 @@ TEST(RecordStoreTestHarness, UpdateRecord) { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); { WriteUnitOfWork uow(opCtx.get()); - Status res = - rs->updateRecord(opCtx.get(), loc, data.c_str(), data.size() + 1, false, NULL); + Status res = rs->updateRecord(opCtx.get(), loc, data.c_str(), data.size() + 1, NULL); ASSERT_OK(res); uow.commit(); @@ -119,7 +118,7 @@ TEST(RecordStoreTestHarness, UpdateMultipleRecords) { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); locs[i] = res.getValue(); uow.commit(); @@ -140,7 +139,7 @@ TEST(RecordStoreTestHarness, UpdateMultipleRecords) { WriteUnitOfWork uow(opCtx.get()); Status res = - rs->updateRecord(opCtx.get(), locs[i], data.c_str(), data.size() + 1, false, NULL); + rs->updateRecord(opCtx.get(), locs[i], data.c_str(), data.size() + 1, NULL); ASSERT_OK(res); uow.commit(); @@ -177,8 +176,8 @@ TEST(RecordStoreTestHarness, UpdateRecordWithMoveNotifier) { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord( - opCtx.get(), oldData.c_str(), oldData.size() + 1, Timestamp(), false); + StatusWith<RecordId> res = + rs->insertRecord(opCtx.get(), oldData.c_str(), oldData.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -197,8 +196,8 @@ TEST(RecordStoreTestHarness, UpdateRecordWithMoveNotifier) { UpdateNotifierSpy umn(opCtx.get(), loc, oldData.c_str(), oldData.size()); WriteUnitOfWork uow(opCtx.get()); - Status res = rs->updateRecord( - opCtx.get(), loc, newData.c_str(), newData.size() + 1, false, &umn); + Status res = + rs->updateRecord(opCtx.get(), loc, newData.c_str(), newData.size() + 1, &umn); ASSERT_OK(res); ASSERT_GTE(1, umn.numInPlaceCallbacks()); diff --git a/src/mongo/db/storage/record_store_test_updatewithdamages.cpp b/src/mongo/db/storage/record_store_test_updatewithdamages.cpp index 2e839f28421..106d8f6e17c 100644 --- a/src/mongo/db/storage/record_store_test_updatewithdamages.cpp +++ b/src/mongo/db/storage/record_store_test_updatewithdamages.cpp @@ -65,7 +65,7 @@ TEST(RecordStoreTestHarness, UpdateWithDamages) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false); + rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -131,7 +131,7 @@ TEST(RecordStoreTestHarness, UpdateWithOverlappingDamageEvents) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false); + rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -195,7 +195,7 @@ TEST(RecordStoreTestHarness, UpdateWithOverlappingDamageEventsReversed) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false); + rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); @@ -257,7 +257,7 @@ TEST(RecordStoreTestHarness, UpdateWithNoDamages) { { WriteUnitOfWork uow(opCtx.get()); StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp(), false); + rs->insertRecord(opCtx.get(), rec.data(), rec.size(), Timestamp()); ASSERT_OK(res.getStatus()); loc = res.getValue(); uow.commit(); diff --git a/src/mongo/db/storage/record_store_test_validate.h b/src/mongo/db/storage/record_store_test_validate.h index 7edc8e152ec..4cecc65bf56 100644 --- a/src/mongo/db/storage/record_store_test_validate.h +++ b/src/mongo/db/storage/record_store_test_validate.h @@ -96,8 +96,8 @@ public: ASSERT(_remain.insert(data).second); WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = _rs->insertRecord( - opCtx.get(), data.c_str(), data.size() + 1, Timestamp(), false); + StatusWith<RecordId> res = + _rs->insertRecord(opCtx.get(), data.c_str(), data.size() + 1, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } diff --git a/src/mongo/db/storage/recovery_unit_test_harness.cpp b/src/mongo/db/storage/recovery_unit_test_harness.cpp index 1ec45a1f929..a5e4f210c75 100644 --- a/src/mongo/db/storage/recovery_unit_test_harness.cpp +++ b/src/mongo/db/storage/recovery_unit_test_harness.cpp @@ -53,7 +53,7 @@ public: TEST_F(RecoveryUnitTestHarness, CommitUnitOfWork) { const auto rs = harnessHelper->createRecordStore(opCtx.get(), "table1"); ru->beginUnitOfWork(opCtx.get()); - StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp(), false); + StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp()); ASSERT_TRUE(s.isOK()); ASSERT_EQUALS(1, rs->numRecords(NULL)); ru->commitUnitOfWork(); @@ -64,7 +64,7 @@ TEST_F(RecoveryUnitTestHarness, CommitUnitOfWork) { TEST_F(RecoveryUnitTestHarness, AbortUnitOfWork) { const auto rs = harnessHelper->createRecordStore(opCtx.get(), "table1"); ru->beginUnitOfWork(opCtx.get()); - StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp(), false); + StatusWith<RecordId> s = rs->insertRecord(opCtx.get(), "data", 4, Timestamp()); ASSERT_TRUE(s.isOK()); ASSERT_EQUALS(1, rs->numRecords(NULL)); ru->abortUnitOfWork(); diff --git a/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp b/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp index d85d6e35e16..73e84d06f05 100644 --- a/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp +++ b/src/mongo/db/storage/wiredtiger/record_store_test_oplog.cpp @@ -44,8 +44,7 @@ StatusWith<RecordId> insertBSON(ServiceContext::UniqueOperationContext& opCtx, Status status = rs->oplogDiskLocRegister(opCtx.get(), opTime, false); if (!status.isOK()) return StatusWith<RecordId>(status); - StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime, false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime); if (res.isOK()) wuow.commit(); return res; @@ -58,7 +57,7 @@ RecordId _oplogOrderInsertOplog(OperationContext* opCtx, Status status = rs->oplogDiskLocRegister(opCtx, opTime, false); ASSERT_OK(status); BSONObj obj = BSON("ts" << opTime); - StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime, false); + StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime); ASSERT_OK(res.getStatus()); return res.getValue(); } @@ -79,19 +78,17 @@ TEST(RecordStore_Oplog, OplogHack) { { WriteUnitOfWork wuow(opCtx.get()); BSONObj obj = BSON("not_ts" << Timestamp(2, 1)); - ASSERT_EQ( - rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, 1), false) - .getStatus(), - ErrorCodes::BadValue); + ASSERT_EQ(rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, 1)) + .getStatus(), + ErrorCodes::BadValue); } { WriteUnitOfWork wuow(opCtx.get()); BSONObj obj = BSON("ts" << "not a Timestamp"); - ASSERT_EQ( - rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(), false) - .getStatus(), - ErrorCodes::BadValue); + ASSERT_EQ(rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp()) + .getStatus(), + ErrorCodes::BadValue); } ASSERT_EQ(insertBSON(opCtx, rs, Timestamp(-2, 1)).getStatus(), ErrorCodes::BadValue); @@ -167,9 +164,8 @@ TEST(RecordStore_Oplog, OplogHackOnNonOplog) { BSONObj obj = BSON("ts" << Timestamp(2, -1)); { WriteUnitOfWork wuow(opCtx.get()); - ASSERT_OK( - rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, -1), false) - .getStatus()); + ASSERT_OK(rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), Timestamp(2, -1)) + .getStatus()); wuow.commit(); } ASSERT_EQ(rs->oplogStartHack(opCtx.get(), RecordId(0, 1)), boost::none); diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp index f9d431af36a..c205e361415 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_prefixed_record_store_test.cpp @@ -213,7 +213,7 @@ TEST(WiredTigerRecordStoreTest, PrefixedTableScan) { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); for (int num = 0; num < numDocs; ++num) { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } @@ -239,7 +239,7 @@ TEST(WiredTigerRecordStoreTest, PrefixedSeekingCursor) { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); for (int num = 0; num < numDocs; ++num) { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); if (startRecordId.isNull()) { startRecordId = res.getValue(); } diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp index 19acc2dbc6d..ab92b1935bd 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.cpp @@ -1239,8 +1239,7 @@ void WiredTigerRecordStore::reclaimOplog(OperationContext* opCtx, Timestamp pers Status WiredTigerRecordStore::insertRecords(OperationContext* opCtx, std::vector<Record>* records, - std::vector<Timestamp>* timestamps, - bool enforceQuota) { + std::vector<Timestamp>* timestamps) { return _insertRecords(opCtx, records->data(), timestamps->data(), records->size()); } @@ -1324,8 +1323,10 @@ Status WiredTigerRecordStore::_insertRecords(OperationContext* opCtx, return Status::OK(); } -StatusWith<RecordId> WiredTigerRecordStore::insertRecord( - OperationContext* opCtx, const char* data, int len, Timestamp timestamp, bool enforceQuota) { +StatusWith<RecordId> WiredTigerRecordStore::insertRecord(OperationContext* opCtx, + const char* data, + int len, + Timestamp timestamp) { Record record = {RecordId(), RecordData(data, len)}; Status status = _insertRecords(opCtx, &record, ×tamp, 1); if (!status.isOK()) @@ -1399,7 +1400,6 @@ Status WiredTigerRecordStore::updateRecord(OperationContext* opCtx, const RecordId& id, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier) { dassert(opCtx->lockState()->isWriteLocked()); diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h index 5ef7017faca..9aa98659c70 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store.h @@ -143,11 +143,12 @@ public: virtual Status insertRecords(OperationContext* opCtx, std::vector<Record>* records, - std::vector<Timestamp>* timestamps, - bool enforceQuota); + std::vector<Timestamp>* timestamps); - virtual StatusWith<RecordId> insertRecord( - OperationContext* opCtx, const char* data, int len, Timestamp timestamp, bool enforceQuota); + virtual StatusWith<RecordId> insertRecord(OperationContext* opCtx, + const char* data, + int len, + Timestamp timestamp); virtual Status insertRecordsWithDocWriter(OperationContext* opCtx, const DocWriter* const* docs, @@ -159,7 +160,6 @@ public: const RecordId& oldLocation, const char* data, int len, - bool enforceQuota, UpdateNotifier* notifier); virtual bool updateWithDamagesSupported() const; diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp index 73841f20e29..28b6efcebaf 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_record_store_test.cpp @@ -113,11 +113,11 @@ TEST(WiredTigerRecordStoreTest, Isolation1) { { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); id1 = res.getValue(); - res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); id2 = res.getValue(); @@ -136,12 +136,12 @@ TEST(WiredTigerRecordStoreTest, Isolation1) { rs->dataFor(t1.get(), id1); rs->dataFor(t2.get(), id1); - ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, false, NULL)); - ASSERT_OK(rs->updateRecord(t1.get(), id2, "B", 2, false, NULL)); + ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, NULL)); + ASSERT_OK(rs->updateRecord(t1.get(), id2, "B", 2, NULL)); try { // this should fail - rs->updateRecord(t2.get(), id1, "c", 2, false, NULL).transitional_ignore(); + rs->updateRecord(t2.get(), id1, "c", 2, NULL).transitional_ignore(); ASSERT(0); } catch (WriteConflictException&) { w2.reset(NULL); @@ -164,11 +164,11 @@ TEST(WiredTigerRecordStoreTest, Isolation2) { { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); id1 = res.getValue(); - res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); id2 = res.getValue(); @@ -187,7 +187,7 @@ TEST(WiredTigerRecordStoreTest, Isolation2) { { WriteUnitOfWork w(t1.get()); - ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, false, NULL)); + ASSERT_OK(rs->updateRecord(t1.get(), id1, "b", 2, NULL)); w.commit(); } @@ -196,7 +196,7 @@ TEST(WiredTigerRecordStoreTest, Isolation2) { ASSERT_EQUALS(string("a"), rs->dataFor(t2.get(), id1).data()); try { // this should fail as our version of id1 is too old - rs->updateRecord(t2.get(), id1, "c", 2, false, NULL).transitional_ignore(); + rs->updateRecord(t2.get(), id1, "c", 2, NULL).transitional_ignore(); ASSERT(0); } catch (WriteConflictException&) { } @@ -215,8 +215,7 @@ StatusWith<RecordId> insertBSON(ServiceContext::UniqueOperationContext& opCtx, Status status = wrs->oplogDiskLocRegister(opCtx.get(), opTime, false); if (!status.isOK()) return StatusWith<RecordId>(status); - StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime, false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), obj.objdata(), obj.objsize(), opTime); if (res.isOK()) wuow.commit(); return res; @@ -230,7 +229,7 @@ TEST(WiredTigerRecordStoreTest, CappedCursorRollover) { ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); for (int i = 0; i < 3; ++i) { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } @@ -250,7 +249,7 @@ TEST(WiredTigerRecordStoreTest, CappedCursorRollover) { auto opCtx = harnessHelper->newOperationContext(client3.get()); for (int i = 0; i < 100; i++) { WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); uow.commit(); } @@ -268,7 +267,7 @@ RecordId _oplogOrderInsertOplog(OperationContext* opCtx, Status status = rs->oplogDiskLocRegister(opCtx, opTime, false); ASSERT_OK(status); BSONObj obj = BSON("ts" << opTime); - StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime, false); + StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime); ASSERT_OK(res.getStatus()); return res.getValue(); } @@ -384,7 +383,7 @@ TEST(WiredTigerRecordStoreTest, CappedCursorYieldFirst) { { // first insert a document ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); WriteUnitOfWork uow(opCtx.get()); - StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); id1 = res.getValue(); uow.commit(); @@ -428,7 +427,7 @@ StatusWith<RecordId> insertBSONWithSize(OperationContext* opCtx, if (!status.isOK()) { return StatusWith<RecordId>(status); } - StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime, false); + StatusWith<RecordId> res = rs->insertRecord(opCtx, obj.objdata(), obj.objsize(), opTime); if (res.isOK()) { wuow.commit(); } @@ -526,9 +525,9 @@ TEST(WiredTigerRecordStoreTest, OplogStones_UpdateRecord) { WriteUnitOfWork wuow(opCtx.get()); ASSERT_NOT_OK(rs->updateRecord( - opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), false, nullptr)); + opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), nullptr)); ASSERT_NOT_OK(rs->updateRecord( - opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), false, nullptr)); + opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), nullptr)); } // Attempts to shrink the records should also fail. @@ -540,9 +539,9 @@ TEST(WiredTigerRecordStoreTest, OplogStones_UpdateRecord) { WriteUnitOfWork wuow(opCtx.get()); ASSERT_NOT_OK(rs->updateRecord( - opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), false, nullptr)); + opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), nullptr)); ASSERT_NOT_OK(rs->updateRecord( - opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), false, nullptr)); + opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), nullptr)); } // Changing the contents of the records without changing their size should succeed. @@ -554,9 +553,9 @@ TEST(WiredTigerRecordStoreTest, OplogStones_UpdateRecord) { WriteUnitOfWork wuow(opCtx.get()); ASSERT_OK(rs->updateRecord( - opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), false, nullptr)); + opCtx.get(), RecordId(1, 1), changed1.objdata(), changed1.objsize(), nullptr)); ASSERT_OK(rs->updateRecord( - opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), false, nullptr)); + opCtx.get(), RecordId(1, 2), changed2.objdata(), changed2.objsize(), nullptr)); wuow.commit(); ASSERT_EQ(1U, oplogStones->numStones()); diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp index 75dc1b4d029..0c3f74516f9 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_recovery_unit_test.cpp @@ -184,8 +184,7 @@ TEST_F(WiredTigerRecoveryUnitTestFixture, CreateAndCheckForCachePressure) { WriteUnitOfWork wu(opCtx); ASSERT_OK(ru1->setTimestamp(Timestamp(time++))); std::string str = str::stream() << "foobarbaz"; - StatusWith<RecordId> ress = - rs->insertRecord(opCtx, str.c_str(), str.size() + 1, Timestamp(), false); + StatusWith<RecordId> ress = rs->insertRecord(opCtx, str.c_str(), str.size() + 1, Timestamp()); ASSERT_OK(ress.getStatus()); auto recordId = ress.getValue(); wu.commit(); @@ -203,7 +202,7 @@ TEST_F(WiredTigerRecoveryUnitTestFixture, CreateAndCheckForCachePressure) { ASSERT_OK(ru1->setTimestamp(Timestamp(time++))); std::string s = str::stream() << "abcbcdcdedefefgfghghihijijkjklklmlmnmnomopopqpqrqrsrststutuv" << j; - ASSERT_OK(rs->updateRecord(opCtx, recordId, s.c_str(), s.size() + 1, false, nullptr)); + ASSERT_OK(rs->updateRecord(opCtx, recordId, s.c_str(), s.size() + 1, nullptr)); wuow.commit(); } catch (const DBException& ex) { invariant(ex.toStatus().code() == ErrorCodes::WriteConflict); diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp index 479bd470bf5..b914b4d6d57 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_standard_record_store_test.cpp @@ -227,8 +227,7 @@ TEST(WiredTigerRecordStoreTest, SizeStorer1) { { WriteUnitOfWork uow(opCtx.get()); for (int i = 0; i < N; i++) { - StatusWith<RecordId> res = - rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false); + StatusWith<RecordId> res = rs->insertRecord(opCtx.get(), "a", 2, Timestamp()); ASSERT_OK(res.getStatus()); } uow.commit(); @@ -329,7 +328,7 @@ private: ServiceContext::UniqueOperationContext opCtx(harnessHelper->newOperationContext()); WriteUnitOfWork uow(opCtx.get()); for (int i = 0; i < expectedNumRecords; i++) { - ASSERT_OK(rs->insertRecord(opCtx.get(), "a", 2, Timestamp(), false).getStatus()); + ASSERT_OK(rs->insertRecord(opCtx.get(), "a", 2, Timestamp()).getStatus()); } uow.commit(); } diff --git a/src/mongo/db/views/durable_view_catalog.cpp b/src/mongo/db/views/durable_view_catalog.cpp index c9a3718fd39..b8691d5f15c 100644 --- a/src/mongo/db/views/durable_view_catalog.cpp +++ b/src/mongo/db/views/durable_view_catalog.cpp @@ -143,12 +143,11 @@ void DurableViewCatalogImpl::upsert(OperationContext* opCtx, const bool requireIndex = false; RecordId id = Helpers::findOne(opCtx, systemViews, BSON("_id" << name.ns()), requireIndex); - const bool enforceQuota = true; Snapshotted<BSONObj> oldView; if (!id.isNormal() || !systemViews->findDoc(opCtx, id, &oldView)) { LOG(2) << "insert view " << view << " into " << _db->getSystemViewsName(); - uassertStatusOK(systemViews->insertDocument( - opCtx, InsertStatement(view), &CurOp::get(opCtx)->debug(), enforceQuota)); + uassertStatusOK( + systemViews->insertDocument(opCtx, InsertStatement(view), &CurOp::get(opCtx)->debug())); } else { OplogUpdateEntryArgs args; args.nss = systemViewsNs; @@ -157,14 +156,8 @@ void DurableViewCatalogImpl::upsert(OperationContext* opCtx, args.fromMigrate = false; const bool assumeIndexesAreAffected = true; - systemViews->updateDocument(opCtx, - id, - oldView, - view, - enforceQuota, - assumeIndexesAreAffected, - &CurOp::get(opCtx)->debug(), - &args); + systemViews->updateDocument( + opCtx, id, oldView, view, assumeIndexesAreAffected, &CurOp::get(opCtx)->debug(), &args); } } diff --git a/src/mongo/dbtests/indexupdatetests.cpp b/src/mongo/dbtests/indexupdatetests.cpp index 697a8ce1abd..6856646b3c3 100644 --- a/src/mongo/dbtests/indexupdatetests.cpp +++ b/src/mongo/dbtests/indexupdatetests.cpp @@ -283,11 +283,11 @@ public: coll = db->createCollection(&_opCtx, _ns); // Drop all indexes including id index. coll->getIndexCatalog()->dropAllIndexes(&_opCtx, true); - // Insert some documents with enforceQuota=true. + // Insert some documents. int32_t nDocs = 1000; OpDebug* const nullOpDebug = nullptr; for (int32_t i = 0; i < nDocs; ++i) { - coll->insertDocument(&_opCtx, InsertStatement(BSON("a" << i)), nullOpDebug, true) + coll->insertDocument(&_opCtx, InsertStatement(BSON("a" << i)), nullOpDebug) .transitional_ignore(); } wunit.commit(); diff --git a/src/mongo/dbtests/multikey_paths_test.cpp b/src/mongo/dbtests/multikey_paths_test.cpp index d3ee08fcc8a..e875ed9da1b 100644 --- a/src/mongo/dbtests/multikey_paths_test.cpp +++ b/src/mongo/dbtests/multikey_paths_test.cpp @@ -132,12 +132,10 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnIndexCreation) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))), - nullOpDebug, - enforceQuota)); + nullOpDebug)); wuow.commit(); } @@ -164,17 +162,14 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnIndexCreationWithMultipleDocuments) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))), - nullOpDebug, - enforceQuota)); + nullOpDebug)); ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement(BSON("_id" << 1 << "a" << BSON_ARRAY(1 << 2 << 3) << "b" << 5)), - nullOpDebug, - enforceQuota)); + nullOpDebug)); wuow.commit(); } @@ -213,12 +208,10 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentInsert) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))), - nullOpDebug, - enforceQuota)); + nullOpDebug)); wuow.commit(); } @@ -227,12 +220,10 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentInsert) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement(BSON("_id" << 1 << "a" << BSON_ARRAY(1 << 2 << 3) << "b" << 5)), - nullOpDebug, - enforceQuota)); + nullOpDebug)); wuow.commit(); } @@ -259,11 +250,8 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentUpdate) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; - ASSERT_OK(collection->insertDocument(_opCtx.get(), - InsertStatement(BSON("_id" << 0 << "a" << 5)), - nullOpDebug, - enforceQuota)); + ASSERT_OK(collection->insertDocument( + _opCtx.get(), InsertStatement(BSON("_id" << 0 << "a" << 5)), nullOpDebug)); wuow.commit(); } @@ -277,7 +265,6 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentUpdate) { auto oldDoc = collection->docFor(_opCtx.get(), record->id); { WriteUnitOfWork wuow(_opCtx.get()); - const bool enforceQuota = true; const bool indexesAffected = true; OpDebug* opDebug = nullptr; OplogUpdateEntryArgs args; @@ -286,7 +273,6 @@ TEST_F(MultikeyPathsTest, PathsUpdatedOnDocumentUpdate) { record->id, oldDoc, BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3)), - enforceQuota, indexesAffected, opDebug, &args); @@ -317,12 +303,10 @@ TEST_F(MultikeyPathsTest, PathsNotUpdatedOnDocumentDelete) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement(BSON("_id" << 0 << "a" << 5 << "b" << BSON_ARRAY(1 << 2 << 3))), - nullOpDebug, - enforceQuota)); + nullOpDebug)); wuow.commit(); } @@ -375,13 +359,11 @@ TEST_F(MultikeyPathsTest, PathsUpdatedForMultipleIndexesOnDocumentInsert) { { WriteUnitOfWork wuow(_opCtx.get()); OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = true; ASSERT_OK(collection->insertDocument( _opCtx.get(), InsertStatement( BSON("_id" << 0 << "a" << BSON_ARRAY(1 << 2 << 3) << "b" << 5 << "c" << 8)), - nullOpDebug, - enforceQuota)); + nullOpDebug)); wuow.commit(); } diff --git a/src/mongo/dbtests/query_stage_cached_plan.cpp b/src/mongo/dbtests/query_stage_cached_plan.cpp index f5afe597563..51503ac4056 100644 --- a/src/mongo/dbtests/query_stage_cached_plan.cpp +++ b/src/mongo/dbtests/query_stage_cached_plan.cpp @@ -103,10 +103,8 @@ public: void insertDocument(Collection* collection, BSONObj obj) { WriteUnitOfWork wuow(&_opCtx); - const bool enforceQuota = false; OpDebug* const nullOpDebug = nullptr; - ASSERT_OK( - collection->insertDocument(&_opCtx, InsertStatement(obj), nullOpDebug, enforceQuota)); + ASSERT_OK(collection->insertDocument(&_opCtx, InsertStatement(obj), nullOpDebug)); wuow.commit(); } diff --git a/src/mongo/dbtests/query_stage_count.cpp b/src/mongo/dbtests/query_stage_count.cpp index e28d388d0aa..a23c692ae1c 100644 --- a/src/mongo/dbtests/query_stage_count.cpp +++ b/src/mongo/dbtests/query_stage_count.cpp @@ -108,8 +108,7 @@ public: void insert(const BSONObj& doc) { WriteUnitOfWork wunit(&_opCtx); OpDebug* const nullOpDebug = nullptr; - _coll->insertDocument(&_opCtx, InsertStatement(doc), nullOpDebug, false) - .transitional_ignore(); + _coll->insertDocument(&_opCtx, InsertStatement(doc), nullOpDebug).transitional_ignore(); wunit.commit(); } @@ -129,7 +128,6 @@ public: oldrecordId, Snapshotted<BSONObj>(_opCtx.recoveryUnit()->getSnapshotId(), oldDoc), newDoc, - false, true, NULL, &args); diff --git a/src/mongo/dbtests/query_stage_sort.cpp b/src/mongo/dbtests/query_stage_sort.cpp index 9ff58f996cd..859bfaf8bce 100644 --- a/src/mongo/dbtests/query_stage_sort.cpp +++ b/src/mongo/dbtests/query_stage_sort.cpp @@ -362,7 +362,7 @@ public: args.nss = coll->ns(); { WriteUnitOfWork wuow(&_opCtx); - coll->updateDocument(&_opCtx, *it, oldDoc, newDoc(oldDoc), false, false, NULL, &args); + coll->updateDocument(&_opCtx, *it, oldDoc, newDoc(oldDoc), false, NULL, &args); wuow.commit(); } ASSERT_OK(exec->restoreState()); @@ -380,8 +380,7 @@ public: oldDoc = coll->docFor(&_opCtx, *it); { WriteUnitOfWork wuow(&_opCtx); - coll->updateDocument( - &_opCtx, *it++, oldDoc, newDoc(oldDoc), false, false, NULL, &args); + coll->updateDocument(&_opCtx, *it++, oldDoc, newDoc(oldDoc), false, NULL, &args); wuow.commit(); } } diff --git a/src/mongo/dbtests/storage_timestamp_tests.cpp b/src/mongo/dbtests/storage_timestamp_tests.cpp index 0a26e2d3002..7501d8ebdb5 100644 --- a/src/mongo/dbtests/storage_timestamp_tests.cpp +++ b/src/mongo/dbtests/storage_timestamp_tests.cpp @@ -216,9 +216,8 @@ public: void insertDocument(Collection* coll, const InsertStatement& stmt) { // Insert some documents. OpDebug* const nullOpDebug = nullptr; - const bool enforceQuota = false; const bool fromMigrate = false; - ASSERT_OK(coll->insertDocument(_opCtx, stmt, nullOpDebug, enforceQuota, fromMigrate)); + ASSERT_OK(coll->insertDocument(_opCtx, stmt, nullOpDebug, fromMigrate)); } void createIndex(Collection* coll, std::string indexName, const BSONObj& indexKey) { diff --git a/src/mongo/dbtests/validate_tests.cpp b/src/mongo/dbtests/validate_tests.cpp index 28d4f5e8498..87c4aaad88a 100644 --- a/src/mongo/dbtests/validate_tests.cpp +++ b/src/mongo/dbtests/validate_tests.cpp @@ -190,8 +190,7 @@ public: WriteUnitOfWork wunit(&_opCtx); for (int j = 0; j < 2; j++) { auto doc = BSON("_id" << j); - ASSERT_OK(rs->insertRecord( - &_opCtx, doc.objdata(), doc.objsize(), Timestamp(), /*enforceQuota*/ false)); + ASSERT_OK(rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), Timestamp())); } wunit.commit(); } @@ -265,8 +264,7 @@ public: WriteUnitOfWork wunit(&_opCtx); for (int j = 0; j < 2; j++) { auto doc = BSON("_id" << j); - ASSERT_OK(rs->insertRecord( - &_opCtx, doc.objdata(), doc.objsize(), Timestamp(), /*enforceQuota*/ false)); + ASSERT_OK(rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), Timestamp())); } wunit.commit(); } @@ -330,8 +328,7 @@ public: { WriteUnitOfWork wunit(&_opCtx); auto doc = BSON("_id" << 1 << "a" << 9); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); @@ -382,8 +379,7 @@ public: { WriteUnitOfWork wunit(&_opCtx); auto doc = BSON("_id" << 9); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } @@ -396,8 +392,7 @@ public: { WriteUnitOfWork wunit(&_opCtx); auto doc = BSON("_id" << 1); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } @@ -414,9 +409,7 @@ public: rs->deleteRecord(&_opCtx, id1); auto doc = BSON("_id" << 3); ASSERT_OK( - rs->insertRecord( - &_opCtx, doc.objdata(), doc.objsize(), Timestamp(), /*enforceQuota*/ false) - .getStatus()); + rs->insertRecord(&_opCtx, doc.objdata(), doc.objsize(), Timestamp()).getStatus()); wunit.commit(); } @@ -491,8 +484,8 @@ public: // Update a document's indexed field without updating the index. { WriteUnitOfWork wunit(&_opCtx); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc1_b.objdata(), doc1_b.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = + rs->updateRecord(&_opCtx, id1, doc1_b.objdata(), doc1_b.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } @@ -505,8 +498,8 @@ public: // Index validation should still be valid. { WriteUnitOfWork wunit(&_opCtx); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc1_c.objdata(), doc1_c.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = + rs->updateRecord(&_opCtx, id1, doc1_c.objdata(), doc1_c.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } @@ -574,8 +567,7 @@ public: { WriteUnitOfWork wunit(&_opCtx); auto doc = BSON("_id" << 2 << "a" << 3); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } @@ -648,8 +640,7 @@ public: { WriteUnitOfWork wunit(&_opCtx); auto doc = BSON("_id" << 1); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } @@ -811,8 +802,7 @@ public: { WriteUnitOfWork wunit(&_opCtx); auto doc = BSON("_id" << 1 << "a" << 1 << "b" << 3); - auto updateStatus = rs->updateRecord( - &_opCtx, id1, doc.objdata(), doc.objsize(), /*enforceQuota*/ false, NULL); + auto updateStatus = rs->updateRecord(&_opCtx, id1, doc.objdata(), doc.objsize(), NULL); ASSERT_OK(updateStatus); wunit.commit(); } |