summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorEric Milkie <milkie@10gen.com>2020-04-22 08:33:42 -0400
committerEvergreen Agent <no-reply@evergreen.mongodb.com>2020-04-22 20:28:50 +0000
commit0f3ea400ef047d93a46bc353ba7b89b1ed7cfc92 (patch)
treec1bfe3a8b53432efc1303aea2a9ecca9da8382cc /src
parent47bdd45ded2b9c16a88877c161023c3364099196 (diff)
downloadmongo-0f3ea400ef047d93a46bc353ba7b89b1ed7cfc92.tar.gz
SERVER-47397 enhanced logging for Execution owned files (part 2)
Diffstat (limited to 'src')
-rw-r--r--src/mongo/db/catalog/multi_index_block.cpp5
-rw-r--r--src/mongo/db/catalog/rename_collection.cpp44
-rw-r--r--src/mongo/db/catalog/validate_adaptor.cpp11
-rw-r--r--src/mongo/db/concurrency/deferred_writer.cpp5
-rw-r--r--src/mongo/db/read_concern_mongod.cpp5
-rw-r--r--src/mongo/db/repair_database.cpp16
-rw-r--r--src/mongo/db/repair_database_and_check_version.cpp11
-rw-r--r--src/mongo/db/storage/storage_engine_impl.cpp53
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp10
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp6
-rw-r--r--src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp20
11 files changed, 95 insertions, 91 deletions
diff --git a/src/mongo/db/catalog/multi_index_block.cpp b/src/mongo/db/catalog/multi_index_block.cpp
index ce10d03abf4..59ef2232661 100644
--- a/src/mongo/db/catalog/multi_index_block.cpp
+++ b/src/mongo/db/catalog/multi_index_block.cpp
@@ -475,8 +475,9 @@ Status MultiIndexBlock::insertAllDocumentsInCollection(OperationContext* opCtx,
LOGV2(20391,
"index build: collection scan done. scanned {n} total records in {t_seconds} seconds",
- "n"_attr = n,
- "t_seconds"_attr = t.seconds());
+ "Index build: collection scan done",
+ "totalRecords"_attr = n,
+ "duration"_attr = duration_cast<Milliseconds>(Seconds(t.seconds())));
Status ret = dumpInsertsFromBulk(opCtx);
if (!ret.isOK())
diff --git a/src/mongo/db/catalog/rename_collection.cpp b/src/mongo/db/catalog/rename_collection.cpp
index 8041004b415..eaecc1bf97b 100644
--- a/src/mongo/db/catalog/rename_collection.cpp
+++ b/src/mongo/db/catalog/rename_collection.cpp
@@ -173,12 +173,13 @@ Status renameTargetCollectionToTmp(OperationContext* opCtx,
LOGV2(20397,
"Successfully renamed the target {targetNs} ({targetUUID}) to {tmpName} so that the "
"source {sourceNs} ({sourceUUID}) could be renamed to {targetNs2}",
- "targetNs"_attr = targetNs,
- "targetUUID"_attr = targetUUID,
- "tmpName"_attr = tmpName,
- "sourceNs"_attr = sourceNs,
+ "Successfully renamed the target so that the source could be renamed",
+ "existingTargetNamespace"_attr = targetNs,
+ "existingTargetUUID"_attr = targetUUID,
+ "renamedExistingTarget"_attr = tmpName,
+ "sourceNamespace"_attr = sourceNs,
"sourceUUID"_attr = sourceUUID,
- "targetNs2"_attr = targetNs);
+ "newTargetNamespace"_attr = targetNs);
return Status::OK();
});
@@ -554,8 +555,9 @@ Status renameBetweenDBs(OperationContext* opCtx,
LOGV2(20398,
"Attempting to create temporary collection: {tmpName} with the contents of collection: "
"{source}",
- "tmpName"_attr = tmpName,
- "source"_attr = source);
+ "Attempting to create temporary collection",
+ "temporaryCollection"_attr = tmpName,
+ "sourceCollection"_attr = source);
Collection* tmpColl = nullptr;
{
@@ -590,10 +592,11 @@ Status renameBetweenDBs(OperationContext* opCtx,
LOGV2(20399,
"Unable to drop temporary collection {tmpName} while renaming from {source} to "
"{target}: {status}",
- "tmpName"_attr = tmpName,
+ "Unable to drop temporary collection while renaming",
+ "tempCollection"_attr = tmpName,
"source"_attr = source,
"target"_attr = target,
- "status"_attr = status);
+ "reason"_attr = status);
}
});
@@ -822,13 +825,13 @@ Status renameCollection(OperationContext* opCtx,
"renaming system.views collection or renaming to system.views is not allowed");
}
- const std::string dropTargetMsg =
- options.dropTarget ? " and drop " + target.toString() + "." : ".";
+ StringData dropTargetMsg = options.dropTarget ? "yes"_sd : "no"_sd;
LOGV2(20400,
"renameCollectionForCommand: rename {source} to {target}{dropTargetMsg}",
- "source"_attr = source,
- "target"_attr = target,
- "dropTargetMsg"_attr = dropTargetMsg);
+ "renameCollectionForCommand",
+ "sourceNamespace"_attr = source,
+ "targetNamespace"_attr = target,
+ "dropTarget"_attr = dropTargetMsg);
if (source.db() == target.db())
return renameCollectionWithinDB(opCtx, source, target, options);
@@ -920,16 +923,16 @@ Status renameCollectionForApplyOps(OperationContext* opCtx,
<< sourceNss.toString());
}
- const std::string dropTargetMsg =
- uuidToDrop ? " and drop " + uuidToDrop->toString() + "." : ".";
+ const std::string uuidToDropString = uuidToDrop ? uuidToDrop->toString() : "<none>";
const std::string uuidString = uuidToRename ? uuidToRename->toString() : "UUID unknown";
LOGV2(20401,
"renameCollectionForApplyOps: rename {sourceNss} ({uuidString}) to "
"{targetNss}{dropTargetMsg}",
- "sourceNss"_attr = sourceNss,
- "uuidString"_attr = uuidString,
- "targetNss"_attr = targetNss,
- "dropTargetMsg"_attr = dropTargetMsg);
+ "renameCollectionForApplyOps",
+ "sourceNamespace"_attr = sourceNss,
+ "uuid"_attr = uuidString,
+ "targetNamespace"_attr = targetNss,
+ "uuidToDrop"_attr = uuidToDropString);
if (sourceNss.db() == targetNss.db()) {
return renameCollectionWithinDBForApplyOps(
@@ -952,6 +955,7 @@ Status renameCollectionForRollback(OperationContext* opCtx,
LOGV2(20402,
"renameCollectionForRollback: rename {source} ({uuid}) to {target}.",
+ "renameCollectionForRollback",
"source"_attr = *source,
"uuid"_attr = uuid,
"target"_attr = target);
diff --git a/src/mongo/db/catalog/validate_adaptor.cpp b/src/mongo/db/catalog/validate_adaptor.cpp
index 7d1c7c66fc8..4e0219d1fcb 100644
--- a/src/mongo/db/catalog/validate_adaptor.cpp
+++ b/src/mongo/db/catalog/validate_adaptor.cpp
@@ -71,10 +71,7 @@ Status ValidateAdaptor::validateRecord(OperationContext* opCtx,
}
if (MONGO_unlikely(_validateState->extraLoggingForTest())) {
- LOGV2(46666001,
- "[validate](record) {record_id}, Value: {record_data}",
- "record_id"_attr = recordId,
- "record_data"_attr = recordBson);
+ LOGV2(46666001, "[validate]", "recordId"_attr = recordId, "recordData"_attr = recordBson);
}
const Status status = validateBSON(
@@ -328,7 +325,6 @@ void ValidateAdaptor::traverseRecordStore(OperationContext* opCtx,
// padding, but we still require that they return the unpadded record data.
if (!status.isOK() || validatedSize != static_cast<size_t>(dataSize)) {
str::stream ss;
- ss << "Document with RecordId " << record->id << " is corrupted. ";
if (!status.isOK() && validatedSize != static_cast<size_t>(dataSize)) {
ss << "Reasons: (1) " << status << "; (2) Validated size of " << validatedSize
<< " bytes does not equal the record size of " << dataSize << " bytes";
@@ -338,7 +334,10 @@ void ValidateAdaptor::traverseRecordStore(OperationContext* opCtx,
ss << "Reason: Validated size of " << validatedSize
<< " bytes does not equal the record size of " << dataSize << " bytes";
}
- LOGV2(20404, "{std_string_ss}", "std_string_ss"_attr = std::string(ss));
+ LOGV2(20404,
+ "Document corruption details",
+ "recordId"_attr = record->id,
+ "reasons"_attr = std::string(ss));
// Only log once
if (results->valid) {
diff --git a/src/mongo/db/concurrency/deferred_writer.cpp b/src/mongo/db/concurrency/deferred_writer.cpp
index bcd8bd8814c..3c1ad5639da 100644
--- a/src/mongo/db/concurrency/deferred_writer.cpp
+++ b/src/mongo/db/concurrency/deferred_writer.cpp
@@ -49,8 +49,9 @@ void DeferredWriter::_logFailure(const Status& status) {
if (TimePoint::clock::now() - _lastLogged > kLogInterval) {
LOGV2(20516,
"Unable to write to collection {nss}: {status}",
- "nss"_attr = _nss.toString(),
- "status"_attr = status.toString());
+ "Unable to write to collection",
+ "namespace"_attr = _nss.toString(),
+ "error"_attr = status);
_lastLogged = stdx::chrono::system_clock::now();
}
}
diff --git a/src/mongo/db/read_concern_mongod.cpp b/src/mongo/db/read_concern_mongod.cpp
index 205506b9f2f..b5b50da9c48 100644
--- a/src/mongo/db/read_concern_mongod.cpp
+++ b/src/mongo/db/read_concern_mongod.cpp
@@ -348,8 +348,9 @@ Status waitForReadConcernImpl(OperationContext* opCtx,
if (!status.isOK()) {
LOGV2(20990,
"Failed noop write at clusterTime: {targetClusterTime} due to {status}",
- "targetClusterTime"_attr = targetClusterTime->toString(),
- "status"_attr = status.toString());
+ "Failed noop write",
+ "targetClusterTime"_attr = targetClusterTime,
+ "error"_attr = status);
}
}
diff --git a/src/mongo/db/repair_database.cpp b/src/mongo/db/repair_database.cpp
index 1ea89de9c86..3badeddc9a6 100644
--- a/src/mongo/db/repair_database.cpp
+++ b/src/mongo/db/repair_database.cpp
@@ -96,7 +96,8 @@ Status dropUnfinishedIndexes(OperationContext* opCtx, Collection* collection) {
LOGV2(3871400,
"Dropping unfinished index '{name}' after collection was modified by "
"repair",
- "name"_attr = indexName);
+ "Dropping unfinished index after collection was modified by repair",
+ "index"_attr = indexName);
WriteUnitOfWork wuow(opCtx);
auto status = durableCatalog->removeIndex(opCtx, collection->getCatalogId(), indexName);
if (!status.isOK()) {
@@ -120,7 +121,7 @@ Status repairCollections(OperationContext* opCtx,
for (const auto& nss : colls) {
opCtx->checkForInterrupt();
- LOGV2(21027, "Repairing collection {nss}", "nss"_attr = nss);
+ LOGV2(21027, "Repairing collection", "namespace"_attr = nss);
auto collection = CollectionCatalog::get(opCtx).lookupCollectionByNamespace(opCtx, nss);
Status status = engine->repairRecordStore(opCtx, collection->getCatalogId(), nss);
@@ -173,9 +174,7 @@ Status repairCollections(OperationContext* opCtx,
return status;
}
- LOGV2(21028,
- "Collection validation results: {output_done}",
- "output_done"_attr = output.done());
+ LOGV2(21028, "Collection validation", "results"_attr = output.done());
if (!validateResults.valid) {
status = rebuildIndexesForNamespace(opCtx, nss, engine);
@@ -195,7 +194,7 @@ Status repairDatabase(OperationContext* opCtx, StorageEngine* engine, const std:
invariant(opCtx->lockState()->isW());
invariant(dbName.find('.') == std::string::npos);
- LOGV2(21029, "repairDatabase {dbName}", "dbName"_attr = dbName);
+ LOGV2(21029, "repairDatabase", "db"_attr = dbName);
BackgroundOperation::assertNoBgOpInProgForDb(dbName);
@@ -212,8 +211,9 @@ Status repairDatabase(OperationContext* opCtx, StorageEngine* engine, const std:
if (!status.isOK()) {
LOGV2_FATAL_CONTINUE(21030,
"Failed to repair database {dbName}: {status_reason}",
- "dbName"_attr = dbName,
- "status_reason"_attr = status.reason());
+ "Failed to repair database",
+ "db"_attr = dbName,
+ "error"_attr = status);
}
try {
diff --git a/src/mongo/db/repair_database_and_check_version.cpp b/src/mongo/db/repair_database_and_check_version.cpp
index f87afde7e16..bc0d4e2a4ac 100644
--- a/src/mongo/db/repair_database_and_check_version.cpp
+++ b/src/mongo/db/repair_database_and_check_version.cpp
@@ -312,8 +312,9 @@ void rebuildIndexes(OperationContext* opCtx, StorageEngine* storageEngine) {
for (const auto& indexName : entry.second.first) {
LOGV2(21004,
"Rebuilding index. Collection: {collNss} Index: {indexName}",
- "collNss"_attr = collNss,
- "indexName"_attr = indexName);
+ "Rebuilding index",
+ "namespace"_attr = collNss,
+ "index"_attr = indexName);
}
std::vector<BSONObj> indexSpecs = entry.second.second;
@@ -461,12 +462,10 @@ bool repairDatabasesAndCheckVersion(OperationContext* opCtx) {
auto repairObserver = StorageRepairObserver::get(opCtx->getServiceContext());
repairObserver->onRepairDone(opCtx);
if (repairObserver->getModifications().size() > 0) {
- LOGV2_WARNING(21018, "Modifications made by repair:");
const auto& mods = repairObserver->getModifications();
for (const auto& mod : mods) {
- LOGV2_WARNING(21019,
- " {mod_getDescription}",
- "mod_getDescription"_attr = mod.getDescription());
+ LOGV2_WARNING(
+ 21019, "repairModification", "description"_attr = mod.getDescription());
}
}
if (repairObserver->isDataInvalidated()) {
diff --git a/src/mongo/db/storage/storage_engine_impl.cpp b/src/mongo/db/storage/storage_engine_impl.cpp
index 06522ecf8ef..99007a754f3 100644
--- a/src/mongo/db/storage/storage_engine_impl.cpp
+++ b/src/mongo/db/storage/storage_engine_impl.cpp
@@ -95,9 +95,7 @@ void StorageEngineImpl::loadCatalog(OperationContext* opCtx) {
Status status = _engine->repairIdent(opCtx, catalogInfo);
if (status.code() == ErrorCodes::DataModifiedByRepair) {
- LOGV2_WARNING(22264,
- "Catalog data modified by repair: {status_reason}",
- "status_reason"_attr = status.reason());
+ LOGV2_WARNING(22264, "Catalog data modified by repair", "reason"_attr = status);
repairObserver->invalidatingModification(str::stream() << "DurableCatalog repaired: "
<< status.reason());
} else {
@@ -163,13 +161,13 @@ void StorageEngineImpl::loadCatalog(OperationContext* opCtx) {
wuow.commit();
auto orphanCollNs = statusWithNs.getValue();
LOGV2(22247,
- "Successfully created an entry in the catalog for the orphaned "
- "collection: {orphanCollNs}",
- "orphanCollNs"_attr = orphanCollNs);
+ "Successfully created an entry in the catalog for orphaned "
+ "collection",
+ "namespace"_attr = orphanCollNs);
LOGV2_WARNING(22265,
- "{orphanCollNs} does not have the _id index. Please manually "
- "build the index.",
- "orphanCollNs"_attr = orphanCollNs);
+ "Collection does not have an _id index. Please manually "
+ "build the index",
+ "namespace"_attr = orphanCollNs);
StorageRepairObserver::get(getGlobalServiceContext())
->benignModification(str::stream() << "Orphan collection created: "
@@ -183,8 +181,7 @@ void StorageEngineImpl::loadCatalog(OperationContext* opCtx) {
"Cannot create an entry in the catalog for the orphaned "
"collection ident: {ident} due to {statusWithNs_getStatus_reason}",
"ident"_attr = ident,
- "statusWithNs_getStatus_reason"_attr =
- statusWithNs.getStatus().reason());
+ "statusWithNs_getStatus_reason"_attr = statusWithNs.getStatus());
LOGV2_ERROR(22269, "Restarting the server will remove this ident.");
}
}
@@ -397,7 +394,7 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn
}
const auto& toRemove = it;
- LOGV2(22251, "Dropping unknown ident: {toRemove}", "toRemove"_attr = toRemove);
+ LOGV2(22251, "Dropping unknown ident", "ident"_attr = toRemove);
WriteUnitOfWork wuow(opCtx);
fassert(40591, _engine->dropIdent(opCtx, opCtx->recoveryUnit(), toRemove));
wuow.commit();
@@ -481,9 +478,10 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn
LOGV2(22253,
"Found index from unfinished build. Collection: {coll} ({collUUID}), index: "
"{indexName}, build UUID: {buildUUID}",
- "coll"_attr = coll,
- "collUUID"_attr = *collUUID,
- "indexName"_attr = indexName,
+ "Found index from unfinished build",
+ "namespace"_attr = coll,
+ "UUID"_attr = *collUUID,
+ "index"_attr = indexName,
"buildUUID"_attr = buildUUID);
// Insert in the map if a build has not already been registered.
@@ -503,10 +501,9 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn
// will return the index to be rebuilt.
if (indexMetaData.isBackgroundSecondaryBuild && (!foundIdent || !indexMetaData.ready)) {
LOGV2(22255,
- "Expected background index build did not complete, rebuilding. "
- "Collection: {coll} Index: {indexName}",
- "coll"_attr = coll,
- "indexName"_attr = indexName);
+ "Expected background index build did not complete, rebuilding",
+ "namespace"_attr = coll,
+ "index"_attr = indexName);
ret.indexesToRebuild.push_back({entry.catalogId, coll, indexName});
continue;
}
@@ -519,9 +516,9 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn
// should be dropped.
if (!indexMetaData.ready && !indexMetaData.isBackgroundSecondaryBuild) {
LOGV2(22256,
- "Dropping unfinished index. Collection: {coll} Index: {indexName}",
- "coll"_attr = coll,
- "indexName"_attr = indexName);
+ "Dropping unfinished index",
+ "namespace"_attr = coll,
+ "index"_attr = indexName);
// Ensure the `ident` is dropped while we have the `indexIdent` value.
fassert(50713, _engine->dropIdent(opCtx, opCtx->recoveryUnit(), indexIdent));
indexesToDrop.push_back(indexName);
@@ -542,7 +539,7 @@ StatusWith<StorageEngine::ReconcileResult> StorageEngineImpl::reconcileCatalogAn
}
for (auto&& temp : internalIdentsToDrop) {
- LOGV2(22257, "Dropping internal ident: {temp}", "temp"_attr = temp);
+ LOGV2(22257, "Dropping internal ident", "ident"_attr = temp);
WriteUnitOfWork wuow(opCtx);
fassert(51067, _engine->dropIdent(opCtx, opCtx->recoveryUnit(), temp));
wuow.commit();
@@ -825,8 +822,8 @@ StatusWith<Timestamp> StorageEngineImpl::recoverToStableTimestamp(OperationConte
catalog::openCatalog(opCtx, state);
LOGV2(22259,
- "recoverToStableTimestamp successful. Stable Timestamp: {swTimestamp_getValue}",
- "swTimestamp_getValue"_attr = swTimestamp.getValue());
+ "recoverToStableTimestamp successful",
+ "stableTimestamp"_attr = swTimestamp.getValue());
return {swTimestamp.getValue()};
}
@@ -916,7 +913,7 @@ void StorageEngineImpl::_onMinOfCheckpointAndOldestTimestampChanged(const Timest
if (auto earliestDropTimestamp = _dropPendingIdentReaper.getEarliestDropTimestamp()) {
if (timestamp > *earliestDropTimestamp) {
LOGV2(22260,
- "Removing drop-pending idents with drop timestamps before timestamp {timestamp}",
+ "Removing drop-pending idents with drop timestamps before timestamp",
"timestamp"_attr = timestamp);
auto opCtx = cc().getOperationContext();
invariant(opCtx);
@@ -1012,8 +1009,8 @@ void StorageEngineImpl::TimestampMonitor::startup() {
// killed, it's fine to give up on future notifications.
LOGV2(22263,
"{Timestamp_monitor_is_stopping_due_to_ex_reason}",
- "Timestamp_monitor_is_stopping_due_to_ex_reason"_attr =
- "Timestamp monitor is stopping due to: " + ex.reason());
+ "Timestamp monitor is stopping",
+ "reason"_attr = ex.reason());
return;
}
},
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp
index d9020bd56bf..b4a0fe8b640 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_global_options.cpp
@@ -51,20 +51,22 @@ Status WiredTigerGlobalOptions::store(const moe::Environment& params) {
if (!wiredTigerGlobalOptions.engineConfig.empty()) {
LOGV2(22293,
"Engine custom option: {wiredTigerGlobalOptions_engineConfig}",
- "wiredTigerGlobalOptions_engineConfig"_attr = wiredTigerGlobalOptions.engineConfig);
+ "Engine custom option",
+ "option"_attr = wiredTigerGlobalOptions.engineConfig);
}
if (!wiredTigerGlobalOptions.collectionConfig.empty()) {
LOGV2(22294,
"Collection custom option: {wiredTigerGlobalOptions_collectionConfig}",
- "wiredTigerGlobalOptions_collectionConfig"_attr =
- wiredTigerGlobalOptions.collectionConfig);
+ "Collection custom option",
+ "option"_attr = wiredTigerGlobalOptions.collectionConfig);
}
if (!wiredTigerGlobalOptions.indexConfig.empty()) {
LOGV2(22295,
"Index custom option: {wiredTigerGlobalOptions_indexConfig}",
- "wiredTigerGlobalOptions_indexConfig"_attr = wiredTigerGlobalOptions.indexConfig);
+ "Index custom option",
+ "option"_attr = wiredTigerGlobalOptions.indexConfig);
}
return Status::OK();
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp
index b583fa197d2..2c9dd2d8dec 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_index.cpp
@@ -291,8 +291,8 @@ void WiredTigerIndex::fullValidate(OperationContext* opCtx,
"in use by other operations.";
LOGV2_WARNING(51781,
- "Could not complete validation of {uri}. This is a transient issue as "
- "the collection was actively in use by other operations.",
+ "Could not complete validation. This is a transient issue as "
+ "the collection was actively in use by other operations",
"uri"_attr = _uri);
fullResults->warnings.push_back(msg);
} else if (err) {
@@ -301,7 +301,7 @@ void WiredTigerIndex::fullValidate(OperationContext* opCtx,
<< "This indicates structural damage. "
<< "Not examining individual index entries.";
LOGV2_ERROR(51782,
- "verify() returned {error}. This indicates structural damage. Not "
+ "verify() returned an error. This indicates structural damage. Not "
"examining individual index entries.",
"error"_attr = wiredtiger_strerror(err));
fullResults->errors.push_back(msg);
diff --git a/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp b/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp
index 03b206d2466..43d647e50b7 100644
--- a/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp
+++ b/src/mongo/db/storage/wiredtiger/wiredtiger_kv_engine.cpp
@@ -434,6 +434,7 @@ public:
LOGV2(22310,
"Triggering the first stable checkpoint. Initial Data: {initialData} PrevStable: "
"{prevStable} CurrStable: {currStable}",
+ "Triggering the first stable checkpoint",
"initialData"_attr = initialData,
"prevStable"_attr = prevStable,
"currStable"_attr = currStable);
@@ -1044,7 +1045,7 @@ Status WiredTigerKVEngine::_salvageIfNeeded(const char* uri) {
int rc = (session->verify)(session, uri, nullptr);
if (rc == 0) {
- LOGV2(22327, "Verify succeeded on uri {uri}. Not salvaging.", "uri"_attr = uri);
+ LOGV2(22327, "Verify succeeded. Not salvaging.", "uri"_attr = uri);
return Status::OK();
}
@@ -1053,7 +1054,7 @@ Status WiredTigerKVEngine::_salvageIfNeeded(const char* uri) {
// lie and return OK to avoid breaking tests. This block should go away when that ticket
// is resolved.
LOGV2_ERROR(22356,
- "Verify on {uri} failed with EBUSY. This means the collection was being "
+ "Verify failed with EBUSY. This means the collection was being "
"accessed. No repair is necessary unless other "
"errors are reported.",
"uri"_attr = uri);
@@ -1061,25 +1062,24 @@ Status WiredTigerKVEngine::_salvageIfNeeded(const char* uri) {
}
if (rc == ENOENT) {
- LOGV2_WARNING(
- 22350,
- "Data file is missing for {uri}. Attempting to drop and re-create the collection.",
- "uri"_attr = uri);
+ LOGV2_WARNING(22350,
+ "Data file is missing. Attempting to drop and re-create the collection.",
+ "uri"_attr = uri);
return _rebuildIdent(session, uri);
}
- LOGV2(22328, "Verify failed on uri {uri}. Running a salvage operation.", "uri"_attr = uri);
+ LOGV2(22328, "Verify failed. Running a salvage operation.", "uri"_attr = uri);
auto status = wtRCToStatus(session->salvage(session, uri, nullptr), "Salvage failed:");
if (status.isOK()) {
return {ErrorCodes::DataModifiedByRepair, str::stream() << "Salvaged data for " << uri};
}
LOGV2_WARNING(22351,
- "Salvage failed for uri {uri}: {status_reason}. The file will be moved out of "
+ "Salvage failed. The file will be moved out of "
"the way and a new ident will be created.",
"uri"_attr = uri,
- "status_reason"_attr = status.reason());
+ "error"_attr = status);
// If the data is unsalvageable, we should completely rebuild the ident.
return _rebuildIdent(session, uri);
@@ -1129,7 +1129,7 @@ Status WiredTigerKVEngine::_rebuildIdent(WT_SESSION* session, const char* uri) {
"swMetadata_getValue"_attr = swMetadata.getValue());
return wtRCToStatus(rc);
}
- LOGV2(22329, "Successfully re-created {uri}.", "uri"_attr = uri);
+ LOGV2(22329, "Successfully re-created table", "uri"_attr = uri);
return {ErrorCodes::DataModifiedByRepair,
str::stream() << "Re-created empty data file for " << uri};
}