/** * Copyright (C) 2018-present MongoDB, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the Server Side Public License, version 1, * as published by MongoDB, Inc. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * Server Side Public License for more details. * * You should have received a copy of the Server Side Public License * along with this program. If not, see * . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the Server Side Public License in all respects for * all of the code used other than as permitted herein. If you modify file(s) * with this exception, you may extend this exception to your version of the * file(s), but you are not obligated to do so. If you do not wish to do so, * delete this exception statement from your version. If you delete this * exception statement from all source files in the program, then also delete * it in the license file. */ #include "mongo/platform/basic.h" #include "mongo/bson/simple_bsonelement_comparator.h" #include "mongo/db/catalog/collection_catalog_entry.h" #include "mongo/db/catalog/database.h" #include "mongo/db/catalog/database_holder.h" #include "mongo/db/catalog/health_log.h" #include "mongo/db/catalog/index_catalog.h" #include "mongo/db/catalog/uuid_catalog.h" #include "mongo/db/concurrency/write_conflict_exception.h" #include "mongo/db/db_raii.h" #include "mongo/db/operation_context.h" #include "mongo/db/query/internal_plans.h" #include "mongo/db/repl/dbcheck.h" #include "mongo/db/repl/dbcheck.h" #include "mongo/db/repl/dbcheck_gen.h" #include "mongo/db/repl/oplog.h" #include "mongo/db/repl/optime.h" namespace mongo { namespace { /* * Some utilities for dealing with the expected/found documents in health log entries. */ bool operator==(const std::vector& lhs, const std::vector& rhs) { return std::equal(lhs.cbegin(), lhs.cend(), rhs.cbegin(), rhs.cend(), [](const auto& x, const auto& y) -> bool { return x.woCompare(y) == 0; }); } /** * Get whether the expected and found objects match, plus an expected/found object to report to the * health log. */ template std::pair expectedFound(const T& expected, const T& found) { auto obj = BSON("expected" << expected << "found" << found); return std::pair(expected == found, obj); } template <> std::pair expectedFound(const BSONObj& expected, const BSONObj& found) { auto obj = BSON("expected" << expected << "found" << found); return std::pair(expected.woCompare(found) == 0, obj); } /** * An overload for boost::optionals, which omits boost::none fields. */ template std::pair expectedFound(const boost::optional& expected, const boost::optional& found) { BSONObjBuilder builder; if (expected) { builder << "expected" << *expected; } if (found) { builder << "found" << *found; } auto obj = builder.obj(); if (expected && found) { return std::pair(*expected == *found, obj); } else if (expected || found) { return std::pair(false, obj); } return std::pair(true, obj); } std::string renderForHealthLog(OplogEntriesEnum op) { switch (op) { case OplogEntriesEnum::Batch: return "dbCheckBatch"; case OplogEntriesEnum::Collection: return "dbCheckCollection"; } MONGO_UNREACHABLE; } /** * Fills in the timestamp and scope, which are always the same for dbCheck's entries. */ std::unique_ptr dbCheckHealthLogEntry(const NamespaceString& nss, SeverityEnum severity, const std::string& msg, OplogEntriesEnum operation, const BSONObj& data) { auto entry = stdx::make_unique(); entry->setNss(nss); entry->setTimestamp(Date_t::now()); entry->setSeverity(severity); entry->setScope(ScopeEnum::Cluster); entry->setMsg(msg); entry->setOperation(renderForHealthLog(operation)); entry->setData(data); return entry; } } /** * Get an error message if the check fails. */ std::unique_ptr dbCheckErrorHealthLogEntry(const NamespaceString& nss, const std::string& msg, OplogEntriesEnum operation, const Status& err) { return dbCheckHealthLogEntry(nss, SeverityEnum::Error, msg, operation, BSON("success" << false << "error" << err.toString())); } /** * Get a HealthLogEntry for a dbCheck batch. */ std::unique_ptr dbCheckBatchEntry(const NamespaceString& nss, int64_t count, int64_t bytes, const std::string& expectedHash, const std::string& foundHash, const BSONKey& minKey, const BSONKey& maxKey, const repl::OpTime& optime) { auto hashes = expectedFound(expectedHash, foundHash); auto data = BSON("success" << true << "count" << count << "bytes" << bytes << "md5" << hashes.second << "minKey" << minKey.elem() << "maxKey" << maxKey.elem() << "optime" << optime); auto severity = hashes.first ? SeverityEnum::Info : SeverityEnum::Error; std::string msg = "dbCheck batch " + (hashes.first ? std::string("consistent") : std::string("inconsistent")); return dbCheckHealthLogEntry(nss, severity, msg, OplogEntriesEnum::Batch, data); } DbCheckHasher::DbCheckHasher(OperationContext* opCtx, Collection* collection, const BSONKey& start, const BSONKey& end, int64_t maxCount, int64_t maxBytes) : _opCtx(opCtx), _maxKey(end), _maxCount(maxCount), _maxBytes(maxBytes) { // Get the MD5 hasher set up. md5_init(&_state); // Get the _id index. const IndexDescriptor* desc = collection->getIndexCatalog()->findIdIndex(opCtx); uassert(ErrorCodes::IndexNotFound, "dbCheck needs _id index", desc); // Set up a simple index scan on that. _exec = InternalPlanner::indexScan(opCtx, collection, desc, start.obj(), end.obj(), BoundInclusion::kIncludeEndKeyOnly, PlanExecutor::NO_YIELD, InternalPlanner::FORWARD, InternalPlanner::IXSCAN_FETCH); } template const md5_byte_t* md5Cast(const T* ptr) { return reinterpret_cast(ptr); } void maybeAppend(md5_state_t* state, const boost::optional& uuid) { if (uuid) { md5_append(state, md5Cast(uuid->toCDR().data()), uuid->toCDR().length()); } } std::string hashCollectionInfo(const DbCheckCollectionInformation& info) { md5_state_t state; md5_init(&state); md5_append(&state, md5Cast(info.collectionName.data()), info.collectionName.size()); maybeAppend(&state, info.prev); maybeAppend(&state, info.next); for (const auto& index : info.indexes) { md5_append(&state, md5Cast(index.objdata()), index.objsize()); } md5_append(&state, md5Cast(info.options.objdata()), info.options.objsize()); md5digest digest; md5_finish(&state, digest); return digestToString(digest); } std::unique_ptr dbCheckCollectionEntry(const NamespaceString& nss, const UUID& uuid, const DbCheckCollectionInformation& expected, const DbCheckCollectionInformation& found, const repl::OpTime& optime) { auto names = expectedFound(expected.collectionName, found.collectionName); auto prevs = expectedFound(expected.prev, found.prev); auto nexts = expectedFound(expected.next, found.next); auto indices = expectedFound(expected.indexes, found.indexes); auto options = expectedFound(expected.options, found.options); bool match = names.first && prevs.first && nexts.first && indices.first && options.first; auto severity = match ? SeverityEnum::Info : SeverityEnum::Error; // Get the hash of all of the other fields. auto md5s = expectedFound(hashCollectionInfo(expected), hashCollectionInfo(found)); std::string msg = "dbCheck collection " + (match ? std::string("consistent") : std::string("inconsistent")); auto data = BSON("success" << true << "uuid" << uuid.toString() << "found" << true << "name" << names.second << "prev" << prevs.second << "next" << nexts.second << "indexes" << indices.second << "options" << options.second << "md5" << md5s.second << "optime" << optime); return dbCheckHealthLogEntry(nss, severity, msg, OplogEntriesEnum::Collection, data); } Status DbCheckHasher::hashAll(void) { BSONObj currentObj; PlanExecutor::ExecState lastState; while (PlanExecutor::ADVANCED == (lastState = _exec->getNext(¤tObj, nullptr))) { if (!currentObj.hasField("_id")) { return Status(ErrorCodes::NoSuchKey, "Document missing _id"); } // If this would put us over a limit, stop here. if (!_canHash(currentObj)) { return Status::OK(); } // Update `last` every time. _last = BSONKey::parseFromBSON(currentObj["_id"]); _bytesSeen += currentObj.objsize(); _countSeen += 1; md5_append(&_state, md5Cast(currentObj.objdata()), currentObj.objsize()); } // If we got to the end of the collection, set the last key to MaxKey. if (lastState == PlanExecutor::IS_EOF) { _last = _maxKey; } return Status::OK(); } std::string DbCheckHasher::total(void) { md5digest digest; md5_finish(&_state, digest); return digestToString(digest); } BSONKey DbCheckHasher::lastKey(void) const { return _last; } int64_t DbCheckHasher::bytesSeen(void) const { return _bytesSeen; } int64_t DbCheckHasher::docsSeen(void) const { return _countSeen; } bool DbCheckHasher::_canHash(const BSONObj& obj) { // Make sure we hash at least one document. if (_countSeen == 0) { return true; } // Check that this won't push us over our byte limit if (_bytesSeen + obj.objsize() > _maxBytes) { return false; } // or our document limit. if (_countSeen + 1 > _maxCount) { return false; } return true; } std::vector collectionIndexInfo(OperationContext* opCtx, Collection* collection) { std::vector result; std::vector names; // List the indices, const auto* cce = collection->getCatalogEntry(); invariant(cce); cce->getAllIndexes(opCtx, &names); // and get the info for each one. for (const auto& name : names) { result.push_back(cce->getIndexSpec(opCtx, name)); } auto comp = stdx::make_unique(); std::sort(result.begin(), result.end(), SimpleBSONObjComparator::LessThan()); return result; } BSONObj collectionOptions(OperationContext* opCtx, Collection* collection) { return collection->getCatalogEntry()->getCollectionOptions(opCtx).toBSON(); } AutoGetDbForDbCheck::AutoGetDbForDbCheck(OperationContext* opCtx, const NamespaceString& nss) : localLock(opCtx, "local"_sd, MODE_IX), agd(opCtx, nss.db(), MODE_S) {} AutoGetCollectionForDbCheck::AutoGetCollectionForDbCheck(OperationContext* opCtx, const NamespaceString& nss, const OplogEntriesEnum& type) : _agd(opCtx, nss), _collLock(opCtx->lockState(), nss.ns(), MODE_S) { std::string msg; _collection = _agd.getDb() ? _agd.getDb()->getCollection(opCtx, nss) : nullptr; // If the collection gets deleted after the check is launched, record that in the health log. if (!_collection) { msg = "Collection under dbCheck no longer exists"; auto entry = dbCheckHealthLogEntry(nss, SeverityEnum::Error, "dbCheck failed", type, BSON("success" << false << "error" << msg)); HealthLog::get(opCtx).log(*entry); } } namespace { Status dbCheckBatchOnSecondary(OperationContext* opCtx, const repl::OpTime& optime, const DbCheckOplogBatch& entry) { AutoGetCollectionForDbCheck agc(opCtx, entry.getNss(), entry.getType()); Collection* collection = agc.getCollection(); std::string msg = "replication consistency check"; if (!collection) { return Status::OK(); } // Set up the hasher, Status status = Status::OK(); boost::optional hasher; try { hasher.emplace(opCtx, collection, entry.getMinKey(), entry.getMaxKey()); } catch (const DBException& exception) { auto logEntry = dbCheckErrorHealthLogEntry( entry.getNss(), msg, OplogEntriesEnum::Batch, exception.toStatus()); HealthLog::get(opCtx).log(*logEntry); return Status::OK(); } // run the hasher. if (status.isOK()) { status = hasher->hashAll(); } // In case of an error, report it to the health log, if (!status.isOK()) { auto logEntry = dbCheckErrorHealthLogEntry(entry.getNss(), msg, OplogEntriesEnum::Batch, status); HealthLog::get(opCtx).log(*logEntry); return Status::OK(); } std::string expected = entry.getMd5().toString(); std::string found = hasher->total(); auto logEntry = dbCheckBatchEntry(entry.getNss(), hasher->docsSeen(), hasher->bytesSeen(), expected, found, entry.getMinKey(), hasher->lastKey(), optime); HealthLog::get(opCtx).log(*logEntry); return Status::OK(); } Status dbCheckDatabaseOnSecondary(OperationContext* opCtx, const repl::OpTime& optime, const DbCheckOplogCollection& entry) { // dbCheckCollectionResult-specific stuff. auto uuid = uassertStatusOK(UUID::parse(entry.getUuid().toString())); auto collection = UUIDCatalog::get(opCtx).lookupCollectionByUUID(uuid); if (!collection) { Status status(ErrorCodes::NamespaceNotFound, "Could not find collection for dbCheck"); auto logEntry = dbCheckErrorHealthLogEntry( entry.getNss(), "dbCheckCollection failed", OplogEntriesEnum::Collection, status); HealthLog::get(opCtx).log(*logEntry); return Status::OK(); } auto db = collection->ns().db(); AutoGetDb agd(opCtx, db, MODE_X); DbCheckCollectionInformation expected; DbCheckCollectionInformation found; expected.collectionName = entry.getNss().coll().toString(); found.collectionName = collection->ns().coll().toString(); // found/expected previous UUID, expected.prev = entry.getPrev(); found.prev = UUIDCatalog::get(opCtx).prev(db, uuid); // found/expected next UUID, expected.next = entry.getNext(); found.next = UUIDCatalog::get(opCtx).next(db, uuid); // found/expected indices, expected.indexes = entry.getIndexes(); found.indexes = collectionIndexInfo(opCtx, collection); // and found/expected collection options. expected.options = entry.getOptions(); found.options = collectionOptions(opCtx, collection); auto hle = dbCheckCollectionEntry(entry.getNss(), uuid, expected, found, optime); HealthLog::get(opCtx).log(*hle); return Status::OK(); } } namespace repl { /* * The corresponding command run on the secondary. */ Status dbCheckOplogCommand(OperationContext* opCtx, const char* ns, const BSONElement& ui, BSONObj& cmd, const repl::OpTime& optime, const repl::OplogEntry& entry, OplogApplication::Mode mode) { auto type = OplogEntries_parse(IDLParserErrorContext("type"), cmd.getStringField("type")); IDLParserErrorContext ctx("o"); switch (type) { case OplogEntriesEnum::Batch: { auto invocation = DbCheckOplogBatch::parse(ctx, cmd); return dbCheckBatchOnSecondary(opCtx, optime, invocation); } case OplogEntriesEnum::Collection: { auto invocation = DbCheckOplogCollection::parse(ctx, cmd); return dbCheckDatabaseOnSecondary(opCtx, optime, invocation); } } MONGO_UNREACHABLE; } } // namespace repl } // namespace mongo