diff options
author | Eric Milkie <milkie@10gen.com> | 2020-04-22 08:33:42 -0400 |
---|---|---|
committer | Evergreen Agent <no-reply@evergreen.mongodb.com> | 2020-04-22 20:28:50 +0000 |
commit | 0f3ea400ef047d93a46bc353ba7b89b1ed7cfc92 (patch) | |
tree | c1bfe3a8b53432efc1303aea2a9ecca9da8382cc /src/mongo/db | |
parent | 47bdd45ded2b9c16a88877c161023c3364099196 (diff) | |
download | mongo-0f3ea400ef047d93a46bc353ba7b89b1ed7cfc92.tar.gz |
SERVER-47397 enhanced logging for Execution owned files (part 2)
Diffstat (limited to 'src/mongo/db')
-rw-r--r-- | src/mongo/db/catalog/multi_index_block.cpp | 5 | ||||
-rw-r--r-- | src/mongo/db/catalog/rename_collection.cpp | 44 | ||||
-rw-r--r-- | src/mongo/db/catalog/validate_adaptor.cpp | 11 | ||||
-rw-r--r-- | src/mongo/db/concurrency/deferred_writer.cpp | 5 | ||||
-rw-r--r-- | src/mongo/db/read_concern_mongod.cpp | 5 | ||||
-rw-r--r-- | src/mongo/db/repair_database.cpp | 16 | ||||
-rw-r--r-- | src/mongo/db/repair_database_and_check_version.cpp | 11 | ||||
-rw-r--r-- | src/mongo/db/storage/storage_engine_impl.cpp | 53 | ||||
-rw-r--r-- | src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp | 10 | ||||
-rw-r--r-- | src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp | 6 | ||||
-rw-r--r-- | src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp | 20 |
11 files changed, 95 insertions, 91 deletions
diff --git a/src/mongo/db/catalog/multi_index_block.cpp b/src/mongo/db/catalog/multi_index_block.cpp index ce10d03abf4..59ef2232661 100644 --- a/src/mongo/db/catalog/multi_index_block.cpp +++ b/src/mongo/db/catalog/multi_index_block.cpp @@ -475,8 +475,9 @@ Status MultiIndexBlock::insertAllDocumentsInCollection(OperationContext* opCtx, LOGV2(20391, "index build: collection scan done. scanned {n} total records in {t_seconds} seconds", - "n"_attr = n, - "t_seconds"_attr = t.seconds()); + "Index build: collection scan done", + "totalRecords"_attr = n, + "duration"_attr = duration_cast<Milliseconds>(Seconds(t.seconds()))); Status ret = dumpInsertsFromBulk(opCtx); if (!ret.isOK()) diff --git a/src/mongo/db/catalog/rename_collection.cpp b/src/mongo/db/catalog/rename_collection.cpp index 8041004b415..eaecc1bf97b 100644 --- a/src/mongo/db/catalog/rename_collection.cpp +++ b/src/mongo/db/catalog/rename_collection.cpp @@ -173,12 +173,13 @@ Status renameTargetCollectionToTmp(OperationContext* opCtx, LOGV2(20397, "Successfully renamed the target {targetNs} ({targetUUID}) to {tmpName} so that the " "source {sourceNs} ({sourceUUID}) could be renamed to {targetNs2}", - "targetNs"_attr = targetNs, - "targetUUID"_attr = targetUUID, - "tmpName"_attr = tmpName, - "sourceNs"_attr = sourceNs, + "Successfully renamed the target so that the source could be renamed", + "existingTargetNamespace"_attr = targetNs, + "existingTargetUUID"_attr = targetUUID, + "renamedExistingTarget"_attr = tmpName, + "sourceNamespace"_attr = sourceNs, "sourceUUID"_attr = sourceUUID, - "targetNs2"_attr = targetNs); + "newTargetNamespace"_attr = targetNs); return Status::OK(); }); @@ -554,8 +555,9 @@ Status renameBetweenDBs(OperationContext* opCtx, LOGV2(20398, "Attempting to create temporary collection: {tmpName} with the contents of collection: " "{source}", - "tmpName"_attr = tmpName, - "source"_attr = source); + "Attempting to create temporary collection", + "temporaryCollection"_attr = tmpName, + "sourceCollection"_attr = source); Collection* tmpColl = nullptr; { @@ -590,10 +592,11 @@ Status renameBetweenDBs(OperationContext* opCtx, LOGV2(20399, "Unable to drop temporary collection {tmpName} while renaming from {source} to " "{target}: {status}", - "tmpName"_attr = tmpName, + "Unable to drop temporary collection while renaming", + "tempCollection"_attr = tmpName, "source"_attr = source, "target"_attr = target, - "status"_attr = status); + "reason"_attr = status); } }); @@ -822,13 +825,13 @@ Status renameCollection(OperationContext* opCtx, "renaming system.views collection or renaming to system.views is not allowed"); } - const std::string dropTargetMsg = - options.dropTarget ? " and drop " + target.toString() + "." : "."; + StringData dropTargetMsg = options.dropTarget ? "yes"_sd : "no"_sd; LOGV2(20400, "renameCollectionForCommand: rename {source} to {target}{dropTargetMsg}", - "source"_attr = source, - "target"_attr = target, - "dropTargetMsg"_attr = dropTargetMsg); + "renameCollectionForCommand", + "sourceNamespace"_attr = source, + "targetNamespace"_attr = target, + "dropTarget"_attr = dropTargetMsg); if (source.db() == target.db()) return renameCollectionWithinDB(opCtx, source, target, options); @@ -920,16 +923,16 @@ Status renameCollectionForApplyOps(OperationContext* opCtx, << sourceNss.toString()); } - const std::string dropTargetMsg = - uuidToDrop ? " and drop " + uuidToDrop->toString() + "." : "."; + const std::string uuidToDropString = uuidToDrop ? uuidToDrop->toString() : "<none>"; const std::string uuidString = uuidToRename ? uuidToRename->toString() : "UUID unknown"; LOGV2(20401, "renameCollectionForApplyOps: rename {sourceNss} ({uuidString}) to " "{targetNss}{dropTargetMsg}", - "sourceNss"_attr = sourceNss, - "uuidString"_attr = uuidString, - "targetNss"_attr = targetNss, - "dropTargetMsg"_attr = dropTargetMsg); + "renameCollectionForApplyOps", + "sourceNamespace"_attr = sourceNss, + "uuid"_attr = uuidString, + "targetNamespace"_attr = targetNss, + "uuidToDrop"_attr = uuidToDropString); if (sourceNss.db() == targetNss.db()) { return renameCollectionWithinDBForApplyOps( @@ -952,6 +955,7 @@ Status renameCollectionForRollback(OperationContext* opCtx, LOGV2(20402, "renameCollectionForRollback: rename {source} ({uuid}) to {target}.", + "renameCollectionForRollback", "source"_attr = *source, "uuid"_attr = uuid, "target"_attr = target); diff --git a/src/mongo/db/catalog/validate_adaptor.cpp b/src/mongo/db/catalog/validate_adaptor.cpp index 7d1c7c66fc8..4e0219d1fcb 100644 --- a/src/mongo/db/catalog/validate_adaptor.cpp +++ b/src/mongo/db/catalog/validate_adaptor.cpp @@ -71,10 +71,7 @@ Status ValidateAdaptor::validateRecord(OperationContext* opCtx, } if (MONGO_unlikely(_validateState->extraLoggingForTest())) { - LOGV2(46666001, - "[validate](record) {record_id}, Value: {record_data}", - "record_id"_attr = recordId, - "record_data"_attr = recordBson); + LOGV2(46666001, "[validate]", "recordId"_attr = recordId, "recordData"_attr = recordBson); } const Status status = validateBSON( @@ -328,7 +325,6 @@ void ValidateAdaptor::traverseRecordStore(OperationContext* opCtx, // padding, but we still require that they return the unpadded record data. if (!status.isOK() || validatedSize != static_cast<size_t>(dataSize)) { str::stream ss; - ss << "Document with RecordId " << record->id << " is corrupted. "; if (!status.isOK() && validatedSize != static_cast<size_t>(dataSize)) { ss << "Reasons: (1) " << status << "; (2) Validated size of " << validatedSize << " bytes does not equal the record size of " << dataSize << " bytes"; @@ -338,7 +334,10 @@ void ValidateAdaptor::traverseRecordStore(OperationContext* opCtx, ss << "Reason: Validated size of " << validatedSize << " bytes does not equal the record size of " << dataSize << " bytes"; } - LOGV2(20404, "{std_string_ss}", "std_string_ss"_attr = std::string(ss)); + LOGV2(20404, + "Document corruption details", + "recordId"_attr = record->id, + "reasons"_attr = std::string(ss)); // Only log once if (results->valid) { diff --git a/src/mongo/db/concurrency/deferred_writer.cpp b/src/mongo/db/concurrency/deferred_writer.cpp index bcd8bd8814c..3c1ad5639da 100644 --- a/src/mongo/db/concurrency/deferred_writer.cpp +++ b/src/mongo/db/concurrency/deferred_writer.cpp @@ -49,8 +49,9 @@ void DeferredWriter::_logFailure(const Status& status) { if (TimePoint::clock::now() - _lastLogged > kLogInterval) { LOGV2(20516, "Unable to write to collection {nss}: {status}", - "nss"_attr = _nss.toString(), - "status"_attr = status.toString()); + "Unable to write to collection", + "namespace"_attr = _nss.toString(), + "error"_attr = status); _lastLogged = stdx::chrono::system_clock::now(); } } diff --git a/src/mongo/db/read_concern_mongod.cpp b/src/mongo/db/read_concern_mongod.cpp index 205506b9f2f..b5b50da9c48 100644 --- a/src/mongo/db/read_concern_mongod.cpp +++ b/src/mongo/db/read_concern_mongod.cpp @@ -348,8 +348,9 @@ Status waitForReadConcernImpl(OperationContext* opCtx, if (!status.isOK()) { LOGV2(20990, "Failed noop write at clusterTime: {targetClusterTime} due to {status}", - "targetClusterTime"_attr = targetClusterTime->toString(), - "status"_attr = status.toString()); + "Failed noop write", + "targetClusterTime"_attr = targetClusterTime, + "error"_attr = status); } } diff --git a/src/mongo/db/repair_database.cpp b/src/mongo/db/repair_database.cpp index 1ea89de9c86..3badeddc9a6 100644 --- a/src/mongo/db/repair_database.cpp +++ b/src/mongo/db/repair_database.cpp @@ -96,7 +96,8 @@ Status dropUnfinishedIndexes(OperationContext* opCtx, Collection* collection) { LOGV2(3871400, "Dropping unfinished index '{name}' after collection was modified by " "repair", - "name"_attr = indexName); + "Dropping unfinished index after collection was modified by repair", + "index"_attr = indexName); WriteUnitOfWork wuow(opCtx); auto status = durableCatalog->removeIndex(opCtx, collection->getCatalogId(), indexName); if (!status.isOK()) { @@ -120,7 +121,7 @@ Status repairCollections(OperationContext* opCtx, for (const auto& nss : colls) { opCtx->checkForInterrupt(); - LOGV2(21027, "Repairing collection {nss}", "nss"_attr = nss); + LOGV2(21027, "Repairing collection", "namespace"_attr = nss); auto collection = CollectionCatalog::get(opCtx).lookupCollectionByNamespace(opCtx, nss); Status status = engine->repairRecordStore(opCtx, collection->getCatalogId(), nss); @@ -173,9 +174,7 @@ Status repairCollections(OperationContext* opCtx, return status; } - LOGV2(21028, - "Collection validation results: {output_done}", - "output_done"_attr = output.done()); + LOGV2(21028, "Collection validation", "results"_attr = output.done()); if (!validateResults.valid) { status = rebuildIndexesForNamespace(opCtx, nss, engine); @@ -195,7 +194,7 @@ Status repairDatabase(OperationContext* opCtx, StorageEngine* engine, const std: invariant(opCtx->lockState()->isW()); invariant(dbName.find('.') == std::string::npos); - LOGV2(21029, "repairDatabase {dbName}", "dbName"_attr = dbName); + LOGV2(21029, "repairDatabase", "db"_attr = dbName); BackgroundOperation::assertNoBgOpInProgForDb(dbName); @@ -212,8 +211,9 @@ Status repairDatabase(OperationContext* opCtx, StorageEngine* engine, const std: if (!status.isOK()) { LOGV2_FATAL_CONTINUE(21030, "Failed to repair database {dbName}: {status_reason}", - "dbName"_attr = dbName, - "status_reason"_attr = status.reason()); + "Failed to repair database", + "db"_attr = dbName, + "error"_attr = status); } try { diff --git a/src/mongo/db/repair_database_and_check_version.cpp b/src/mongo/db/repair_database_and_check_version.cpp index f87afde7e16..bc0d4e2a4ac 100644 --- a/src/mongo/db/repair_database_and_check_version.cpp +++ b/src/mongo/db/repair_database_and_check_version.cpp @@ -312,8 +312,9 @@ void rebuildIndexes(OperationContext* opCtx, StorageEngine* storageEngine) { for (const auto& indexName : entry.second.first) { LOGV2(21004, "Rebuilding index. Collection: {collNss} Index: {indexName}", - "collNss"_attr = collNss, - "indexName"_attr = indexName); + "Rebuilding index", + "namespace"_attr = collNss, + "index"_attr = indexName); } std::vector<BSONObj> indexSpecs = entry.second.second; @@ -461,12 +462,10 @@ bool repairDatabasesAndCheckVersion(OperationContext* opCtx) { auto repairObserver = StorageRepairObserver::get(opCtx->getServiceContext()); repairObserver->onRepairDone(opCtx); if (repairObserver->getModifications().size() > 0) { - LOGV2_WARNING(21018, "Modifications made by repair:"); const auto& mods = repairObserver->getModifications(); for (const auto& mod : mods) { - LOGV2_WARNING(21019, - " {mod_getDescription}", - "mod_getDescription"_attr = mod.getDescription()); + LOGV2_WARNING( + 21019, "repairModification", "description"_attr = mod.getDescription()); } } if (repairObserver->isDataInvalidated()) { diff --git a/src/mongo/db/storage/storage_engine_impl.cpp b/src/mongo/db/storage/storage_engine_impl.cpp index 06522ecf8ef..99007a754f3 100644 --- a/src/mongo/db/storage/storage_engine_impl.cpp +++ b/src/mongo/db/storage/storage_engine_impl.cpp @@ -95,9 +95,7 @@ void StorageEngineImpl::loadCatalog(OperationContext* opCtx) { Status status = _engine->repairIdent(opCtx, catalogInfo); if (status.code() == ErrorCodes::DataModifiedByRepair) { - LOGV2_WARNING(22264, - "Catalog data modified by repair: {status_reason}", - "status_reason"_attr = status.reason()); + LOGV2_WARNING(22264, "Catalog data modified by repair", "reason"_attr = status); repairObserver->invalidatingModification(str::stream() << "DurableCatalog repaired: " << status.reason()); } else { @@ -163,13 +161,13 @@ void StorageEngineImpl::loadCatalog(OperationContext* opCtx) { wuow.commit(); auto orphanCollNs = statusWithNs.getValue(); LOGV2(22247, - "Successfully created an entry in the catalog for the orphaned " - "collection: {orphanCollNs}", - "orphanCollNs"_attr = orphanCollNs); + "Successfully created an entry in the catalog for orphaned " + "collection", + "namespace"_attr = orphanCollNs); LOGV2_WARNING(22265, - "{orphanCollNs} does not have the _id index. Please manually " - "build the index.", - "orphanCollNs"_attr = orphanCollNs); + "Collection does not have an _id index. Please manually " + "build the index", + "namespace"_attr = orphanCollNs); StorageRepairObserver::get(getGlobalServiceContext()) ->benignModification(str::stream() << "Orphan collection created: " @@ -183,8 +181,7 @@ void StorageEngineImpl::loadCatalog(OperationContext* opCtx) { "Cannot create an entry in the catalog for the orphaned " "collection ident: {ident} due to {statusWithNs_getStatus_reason}", "ident"_attr = ident, - "statusWithNs_getStatus_reason"_attr = - statusWithNs.getStatus().reason()); + "statusWithNs_getStatus_reason"_attr = statusWithNs.getStatus()); LOGV2_ERROR(22269, "Restarting the server will remove this ident."); } } @@ -397,7 +394,7 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn } const auto& toRemove = it; - LOGV2(22251, "Dropping unknown ident: {toRemove}", "toRemove"_attr = toRemove); + LOGV2(22251, "Dropping unknown ident", "ident"_attr = toRemove); WriteUnitOfWork wuow(opCtx); fassert(40591, _engine->dropIdent(opCtx, opCtx->recoveryUnit(), toRemove)); wuow.commit(); @@ -481,9 +478,10 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn LOGV2(22253, "Found index from unfinished build. Collection: {coll} ({collUUID}), index: " "{indexName}, build UUID: {buildUUID}", - "coll"_attr = coll, - "collUUID"_attr = *collUUID, - "indexName"_attr = indexName, + "Found index from unfinished build", + "namespace"_attr = coll, + "UUID"_attr = *collUUID, + "index"_attr = indexName, "buildUUID"_attr = buildUUID); // Insert in the map if a build has not already been registered. @@ -503,10 +501,9 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn // will return the index to be rebuilt. if (indexMetaData.isBackgroundSecondaryBuild && (!foundIdent || !indexMetaData.ready)) { LOGV2(22255, - "Expected background index build did not complete, rebuilding. " - "Collection: {coll} Index: {indexName}", - "coll"_attr = coll, - "indexName"_attr = indexName); + "Expected background index build did not complete, rebuilding", + "namespace"_attr = coll, + "index"_attr = indexName); ret.indexesToRebuild.push_back({entry.catalogId, coll, indexName}); continue; } @@ -519,9 +516,9 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn // should be dropped. if (!indexMetaData.ready && !indexMetaData.isBackgroundSecondaryBuild) { LOGV2(22256, - "Dropping unfinished index. Collection: {coll} Index: {indexName}", - "coll"_attr = coll, - "indexName"_attr = indexName); + "Dropping unfinished index", + "namespace"_attr = coll, + "index"_attr = indexName); // Ensure the `ident` is dropped while we have the `indexIdent` value. fassert(50713, _engine->dropIdent(opCtx, opCtx->recoveryUnit(), indexIdent)); indexesToDrop.push_back(indexName); @@ -542,7 +539,7 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn } for (auto&& temp : internalIdentsToDrop) { - LOGV2(22257, "Dropping internal ident: {temp}", "temp"_attr = temp); + LOGV2(22257, "Dropping internal ident", "ident"_attr = temp); WriteUnitOfWork wuow(opCtx); fassert(51067, _engine->dropIdent(opCtx, opCtx->recoveryUnit(), temp)); wuow.commit(); @@ -825,8 +822,8 @@ StatusWith<Timestamp> StorageEngineImpl::recoverToStableTimestamp(OperationConte catalog::openCatalog(opCtx, state); LOGV2(22259, - "recoverToStableTimestamp successful. Stable Timestamp: {swTimestamp_getValue}", - "swTimestamp_getValue"_attr = swTimestamp.getValue()); + "recoverToStableTimestamp successful", + "stableTimestamp"_attr = swTimestamp.getValue()); return {swTimestamp.getValue()}; } @@ -916,7 +913,7 @@ void StorageEngineImpl::_onMinOfCheckpointAndOldestTimestampChanged(const Timest if (auto earliestDropTimestamp = _dropPendingIdentReaper.getEarliestDropTimestamp()) { if (timestamp > *earliestDropTimestamp) { LOGV2(22260, - "Removing drop-pending idents with drop timestamps before timestamp {timestamp}", + "Removing drop-pending idents with drop timestamps before timestamp", "timestamp"_attr = timestamp); auto opCtx = cc().getOperationContext(); invariant(opCtx); @@ -1012,8 +1009,8 @@ void StorageEngineImpl::TimestampMonitor::startup() { // killed, it's fine to give up on future notifications. LOGV2(22263, "{Timestamp_monitor_is_stopping_due_to_ex_reason}", - "Timestamp_monitor_is_stopping_due_to_ex_reason"_attr = - "Timestamp monitor is stopping due to: " + ex.reason()); + "Timestamp monitor is stopping", + "reason"_attr = ex.reason()); return; } }, diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp index d9020bd56bf..b4a0fe8b640 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp @@ -51,20 +51,22 @@ Status WiredTigerGlobalOptions::store(const moe::Environment& params) { if (!wiredTigerGlobalOptions.engineConfig.empty()) { LOGV2(22293, "Engine custom option: {wiredTigerGlobalOptions_engineConfig}", - "wiredTigerGlobalOptions_engineConfig"_attr = wiredTigerGlobalOptions.engineConfig); + "Engine custom option", + "option"_attr = wiredTigerGlobalOptions.engineConfig); } if (!wiredTigerGlobalOptions.collectionConfig.empty()) { LOGV2(22294, "Collection custom option: {wiredTigerGlobalOptions_collectionConfig}", - "wiredTigerGlobalOptions_collectionConfig"_attr = - wiredTigerGlobalOptions.collectionConfig); + "Collection custom option", + "option"_attr = wiredTigerGlobalOptions.collectionConfig); } if (!wiredTigerGlobalOptions.indexConfig.empty()) { LOGV2(22295, "Index custom option: {wiredTigerGlobalOptions_indexConfig}", - "wiredTigerGlobalOptions_indexConfig"_attr = wiredTigerGlobalOptions.indexConfig); + "Index custom option", + "option"_attr = wiredTigerGlobalOptions.indexConfig); } return Status::OK(); diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp index b583fa197d2..2c9dd2d8dec 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp @@ -291,8 +291,8 @@ void WiredTigerIndex::fullValidate(OperationContext* opCtx, "in use by other operations."; LOGV2_WARNING(51781, - "Could not complete validation of {uri}. This is a transient issue as " - "the collection was actively in use by other operations.", + "Could not complete validation. This is a transient issue as " + "the collection was actively in use by other operations", "uri"_attr = _uri); fullResults->warnings.push_back(msg); } else if (err) { @@ -301,7 +301,7 @@ void WiredTigerIndex::fullValidate(OperationContext* opCtx, << "This indicates structural damage. " << "Not examining individual index entries."; LOGV2_ERROR(51782, - "verify() returned {error}. This indicates structural damage. Not " + "verify() returned an error. This indicates structural damage. Not " "examining individual index entries.", "error"_attr = wiredtiger_strerror(err)); fullResults->errors.push_back(msg); diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp index 03b206d2466..43d647e50b7 100644 --- a/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp +++ b/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp @@ -434,6 +434,7 @@ public: LOGV2(22310, "Triggering the first stable checkpoint. Initial Data: {initialData} PrevStable: " "{prevStable} CurrStable: {currStable}", + "Triggering the first stable checkpoint", "initialData"_attr = initialData, "prevStable"_attr = prevStable, "currStable"_attr = currStable); @@ -1044,7 +1045,7 @@ Status WiredTigerKVEngine::_salvageIfNeeded(const char* uri) { int rc = (session->verify)(session, uri, nullptr); if (rc == 0) { - LOGV2(22327, "Verify succeeded on uri {uri}. Not salvaging.", "uri"_attr = uri); + LOGV2(22327, "Verify succeeded. Not salvaging.", "uri"_attr = uri); return Status::OK(); } @@ -1053,7 +1054,7 @@ Status WiredTigerKVEngine::_salvageIfNeeded(const char* uri) { // lie and return OK to avoid breaking tests. This block should go away when that ticket // is resolved. LOGV2_ERROR(22356, - "Verify on {uri} failed with EBUSY. This means the collection was being " + "Verify failed with EBUSY. This means the collection was being " "accessed. No repair is necessary unless other " "errors are reported.", "uri"_attr = uri); @@ -1061,25 +1062,24 @@ Status WiredTigerKVEngine::_salvageIfNeeded(const char* uri) { } if (rc == ENOENT) { - LOGV2_WARNING( - 22350, - "Data file is missing for {uri}. Attempting to drop and re-create the collection.", - "uri"_attr = uri); + LOGV2_WARNING(22350, + "Data file is missing. Attempting to drop and re-create the collection.", + "uri"_attr = uri); return _rebuildIdent(session, uri); } - LOGV2(22328, "Verify failed on uri {uri}. Running a salvage operation.", "uri"_attr = uri); + LOGV2(22328, "Verify failed. Running a salvage operation.", "uri"_attr = uri); auto status = wtRCToStatus(session->salvage(session, uri, nullptr), "Salvage failed:"); if (status.isOK()) { return {ErrorCodes::DataModifiedByRepair, str::stream() << "Salvaged data for " << uri}; } LOGV2_WARNING(22351, - "Salvage failed for uri {uri}: {status_reason}. The file will be moved out of " + "Salvage failed. The file will be moved out of " "the way and a new ident will be created.", "uri"_attr = uri, - "status_reason"_attr = status.reason()); + "error"_attr = status); // If the data is unsalvageable, we should completely rebuild the ident. return _rebuildIdent(session, uri); @@ -1129,7 +1129,7 @@ Status WiredTigerKVEngine::_rebuildIdent(WT_SESSION* session, const char* uri) { "swMetadata_getValue"_attr = swMetadata.getValue()); return wtRCToStatus(rc); } - LOGV2(22329, "Successfully re-created {uri}.", "uri"_attr = uri); + LOGV2(22329, "Successfully re-created table", "uri"_attr = uri); return {ErrorCodes::DataModifiedByRepair, str::stream() << "Re-created empty data file for " << uri}; } |