/** * Copyright (C) 2018-present MongoDB, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the Server Side Public License, version 1, * as published by MongoDB, Inc. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * Server Side Public License for more details. * * You should have received a copy of the Server Side Public License * along with this program. If not, see * . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the Server Side Public License in all respects for * all of the code used other than as permitted herein. If you modify file(s) * with this exception, you may extend this exception to your version of the * file(s), but you are not obligated to do so. If you do not wish to do so, * delete this exception statement from your version. If you delete this * exception statement from all source files in the program, then also delete * it in the license file. */ #define MONGO_LOG_DEFAULT_COMPONENT ::mongo::logger::LogComponent::kSharding #include "mongo/platform/basic.h" #include "mongo/s/commands/strategy.h" #include "mongo/base/data_cursor.h" #include "mongo/base/init.h" #include "mongo/base/status.h" #include "mongo/bson/util/bson_extract.h" #include "mongo/bson/util/builder.h" #include "mongo/db/audit.h" #include "mongo/db/auth/action_type.h" #include "mongo/db/auth/authorization_session.h" #include "mongo/db/commands.h" #include "mongo/db/commands/test_commands_enabled.h" #include "mongo/db/curop.h" #include "mongo/db/handle_request_response.h" #include "mongo/db/initialize_operation_session_info.h" #include "mongo/db/lasterror.h" #include "mongo/db/logical_clock.h" #include "mongo/db/logical_session_id_helpers.h" #include "mongo/db/logical_time_validator.h" #include "mongo/db/matcher/extensions_callback_noop.h" #include "mongo/db/namespace_string.h" #include "mongo/db/operation_time_tracker.h" #include "mongo/db/ops/write_ops.h" #include "mongo/db/query/find_common.h" #include "mongo/db/query/getmore_request.h" #include "mongo/db/query/query_request.h" #include "mongo/db/stats/counters.h" #include "mongo/db/transaction_validation.h" #include "mongo/db/views/resolved_view.h" #include "mongo/rpc/factory.h" #include "mongo/rpc/get_status_from_command_result.h" #include "mongo/rpc/metadata/logical_time_metadata.h" #include "mongo/rpc/metadata/tracking_metadata.h" #include "mongo/rpc/op_msg.h" #include "mongo/rpc/op_msg_rpc_impls.h" #include "mongo/s/cannot_implicitly_create_collection_info.h" #include "mongo/s/catalog_cache.h" #include "mongo/s/client/parallel.h" #include "mongo/s/client/shard_connection.h" #include "mongo/s/client/shard_registry.h" #include "mongo/s/cluster_commands_helpers.h" #include "mongo/s/commands/cluster_explain.h" #include "mongo/s/grid.h" #include "mongo/s/query/cluster_cursor_manager.h" #include "mongo/s/query/cluster_find.h" #include "mongo/s/stale_exception.h" #include "mongo/s/transaction_router.h" #include "mongo/util/fail_point_service.h" #include "mongo/util/log.h" #include "mongo/util/mongoutils/str.h" #include "mongo/util/scopeguard.h" #include "mongo/util/timer.h" namespace mongo { namespace { const auto kOperationTime = "operationTime"_sd; /** * Extract and process metadata from the command request body. */ Status processCommandMetadata(OperationContext* opCtx, const BSONObj& cmdObj) { ReadPreferenceSetting::get(opCtx) = uassertStatusOK(ReadPreferenceSetting::fromContainingBSON(cmdObj)); auto logicalClock = LogicalClock::get(opCtx); invariant(logicalClock); auto logicalTimeMetadata = rpc::LogicalTimeMetadata::readFromMetadata(cmdObj); if (!logicalTimeMetadata.isOK()) { return logicalTimeMetadata.getStatus(); } auto logicalTimeValidator = LogicalTimeValidator::get(opCtx); const auto& signedTime = logicalTimeMetadata.getValue().getSignedTime(); // No need to check proof is no time is given. if (signedTime.getTime() == LogicalTime::kUninitialized) { return Status::OK(); } if (!LogicalTimeValidator::isAuthorizedToAdvanceClock(opCtx)) { auto advanceClockStatus = logicalTimeValidator->validate(opCtx, signedTime); if (!advanceClockStatus.isOK()) { return advanceClockStatus; } } return logicalClock->advanceClusterTime(signedTime.getTime()); } /** * Append required fields to command response. */ void appendRequiredFieldsToResponse(OperationContext* opCtx, BSONObjBuilder* responseBuilder) { auto validator = LogicalTimeValidator::get(opCtx); if (validator->shouldGossipLogicalTime()) { auto now = LogicalClock::get(opCtx)->getClusterTime(); // Add operationTime. auto operationTime = OperationTimeTracker::get(opCtx)->getMaxOperationTime(); if (operationTime != LogicalTime::kUninitialized) { responseBuilder->append(kOperationTime, operationTime.asTimestamp()); } else if (now != LogicalTime::kUninitialized) { // If we don't know the actual operation time, use the cluster time instead. This is // safe but not optimal because we can always return a later operation time than actual. responseBuilder->append(kOperationTime, now.asTimestamp()); } // Add $clusterTime. if (LogicalTimeValidator::isAuthorizedToAdvanceClock(opCtx)) { SignedLogicalTime dummySignedTime(now, TimeProofService::TimeProof(), 0); rpc::LogicalTimeMetadata(dummySignedTime).writeToMetadata(responseBuilder); } else { auto currentTime = validator->signLogicalTime(opCtx, now); rpc::LogicalTimeMetadata(currentTime).writeToMetadata(responseBuilder); } } } /** * Invokes the given command and aborts the transaction on any non-retryable errors. */ void invokeInTransactionRouter(OperationContext* opCtx, CommandInvocation* invocation, TransactionRouter* txnRouter, rpc::ReplyBuilderInterface* result) { // No-op if the transaction is not running with snapshot read concern. txnRouter->setDefaultAtClusterTime(opCtx); try { invocation->run(opCtx, result); } catch (const DBException& e) { if (ErrorCodes::isSnapshotError(e.code()) || ErrorCodes::isNeedRetargettingError(e.code()) || e.code() == ErrorCodes::StaleDbVersion) { // Don't abort on possibly retryable errors. throw; } txnRouter->implicitlyAbortTransaction(opCtx); throw; } } /** * Throws NoSuchTransaction if canRetry is false. */ void handleCanRetryInTransaction(OperationContext* opCtx, TransactionRouter* txnRouter, bool canRetry, const DBException& ex) { if (!canRetry) { uasserted(ErrorCodes::NoSuchTransaction, str::stream() << "Transaction " << opCtx->getTxnNumber() << " was aborted after " << kMaxNumStaleVersionRetries << " failed retries. The latest attempt failed with: " << ex.toStatus()); } } void execCommandClient(OperationContext* opCtx, CommandInvocation* invocation, const OpMsgRequest& request, rpc::ReplyBuilderInterface* result) { const Command* c = invocation->definition(); ON_BLOCK_EXIT([opCtx, &result] { auto body = result->getBodyBuilder(); appendRequiredFieldsToResponse(opCtx, &body); }); const auto dbname = request.getDatabase(); uassert(ErrorCodes::IllegalOperation, "Can't use 'local' database through mongos", dbname != NamespaceString::kLocalDb); uassert(ErrorCodes::InvalidNamespace, str::stream() << "Invalid database name: '" << dbname << "'", NamespaceString::validDBName(dbname, NamespaceString::DollarInDbNameBehavior::Allow)); StringMap topLevelFields; for (auto&& element : request.body) { StringData fieldName = element.fieldNameStringData(); if (fieldName == "help" && element.type() == Bool && element.Bool()) { std::stringstream help; help << "help for: " << c->getName() << " " << c->help(); auto body = result->getBodyBuilder(); body.append("help", help.str()); CommandHelpers::appendSimpleCommandStatus(body, true, ""); return; } uassert(ErrorCodes::FailedToParse, str::stream() << "Parsed command object contains duplicate top level key: " << fieldName, topLevelFields[fieldName]++ == 0); } try { invocation->checkAuthorization(opCtx, request); } catch (const DBException& e) { auto body = result->getBodyBuilder(); CommandHelpers::appendCommandStatusNoThrow(body, e.toStatus()); return; } c->incrementCommandsExecuted(); if (c->shouldAffectCommandCounter()) { globalOpCounters.gotCommand(); } auto wcResult = uassertStatusOK(WriteConcernOptions::extractWCFromCommand(request.body)); bool supportsWriteConcern = invocation->supportsWriteConcern(); if (!supportsWriteConcern && !wcResult.usedDefault) { // This command doesn't do writes so it should not be passed a writeConcern. // If we did not use the default writeConcern, one was provided when it shouldn't have // been by the user. auto body = result->getBodyBuilder(); CommandHelpers::appendCommandStatusNoThrow( body, Status(ErrorCodes::InvalidOptions, "Command does not support writeConcern")); return; } if (TransactionRouter::get(opCtx)) { validateWriteConcernForTransaction(wcResult, c->getName()); } auto& readConcernArgs = repl::ReadConcernArgs::get(opCtx); if (readConcernArgs.getLevel() == repl::ReadConcernLevel::kSnapshotReadConcern) { uassert(ErrorCodes::InvalidOptions, "read concern snapshot is only supported in a multi-statement transaction", TransactionRouter::get(opCtx)); } // attach tracking rpc::TrackingMetadata trackingMetadata; trackingMetadata.initWithOperName(c->getName()); rpc::TrackingMetadata::get(opCtx) = trackingMetadata; auto metadataStatus = processCommandMetadata(opCtx, request.body); if (!metadataStatus.isOK()) { auto body = result->getBodyBuilder(); CommandHelpers::appendCommandStatusNoThrow(body, metadataStatus); return; } auto txnRouter = TransactionRouter::get(opCtx); if (!supportsWriteConcern) { if (txnRouter) { invokeInTransactionRouter(opCtx, invocation, txnRouter, result); } else { invocation->run(opCtx, result); } } else { // Change the write concern while running the command. const auto oldWC = opCtx->getWriteConcern(); ON_BLOCK_EXIT([&] { opCtx->setWriteConcern(oldWC); }); opCtx->setWriteConcern(wcResult); if (txnRouter) { invokeInTransactionRouter(opCtx, invocation, txnRouter, result); } else { invocation->run(opCtx, result); } } auto body = result->getBodyBuilder(); bool ok = CommandHelpers::extractOrAppendOk(body); if (!ok) { c->incrementCommandsFailed(); if (auto txnRouter = TransactionRouter::get(opCtx)) { txnRouter->implicitlyAbortTransaction(opCtx); } } } MONGO_FAIL_POINT_DEFINE(doNotRefreshShardsOnRetargettingError); /** * Executes the command for the given request, and appends the result to replyBuilder * and error labels, if any, to errorBuilder. */ void runCommand(OperationContext* opCtx, const OpMsgRequest& request, const NetworkOp opType, rpc::ReplyBuilderInterface* replyBuilder, BSONObjBuilder* errorBuilder) { auto const commandName = request.getCommandName(); auto const command = CommandHelpers::findCommand(commandName); if (!command) { auto builder = replyBuilder->getBodyBuilder(); ON_BLOCK_EXIT([opCtx, &builder] { appendRequiredFieldsToResponse(opCtx, &builder); }); CommandHelpers::appendCommandStatusNoThrow( builder, {ErrorCodes::CommandNotFound, str::stream() << "no such cmd: " << commandName}); globalCommandRegistry()->incrementUnknownCommands(); return; } CommandHelpers::uassertShouldAttemptParse(opCtx, command, request); // Parse the 'maxTimeMS' command option, and use it to set a deadline for the operation on // the OperationContext. Be sure to do this as soon as possible so that further processing by // subsequent code has the deadline available. The 'maxTimeMS' option unfortunately has a // different meaning for a getMore command, where it is used to communicate the maximum time to // wait for new inserts on tailable cursors, not as a deadline for the operation. // TODO SERVER-34277 Remove the special handling for maxTimeMS for getMores. This will // require introducing a new 'max await time' parameter for getMore, and eventually banning // maxTimeMS altogether on a getMore command. uassert(ErrorCodes::InvalidOptions, "no such command option $maxTimeMs; use maxTimeMS instead", request.body[QueryRequest::queryOptionMaxTimeMS].eoo()); const int maxTimeMS = uassertStatusOK( QueryRequest::parseMaxTimeMS(request.body[QueryRequest::cmdOptionMaxTimeMS])); if (maxTimeMS > 0 && command->getLogicalOp() != LogicalOp::opGetMore) { opCtx->setDeadlineAfterNowBy(Milliseconds{maxTimeMS}, ErrorCodes::MaxTimeMSExpired); } opCtx->checkForInterrupt(); // May trigger maxTimeAlwaysTimeOut fail point. auto invocation = command->parse(opCtx, request); // Set the logical optype, command object and namespace as soon as we identify the command. If // the command does not define a fully-qualified namespace, set CurOp to the generic command // namespace db.$cmd. std::string ns = invocation->ns().toString(); auto nss = (request.getDatabase() == ns ? NamespaceString(ns, "$cmd") : NamespaceString(ns)); // Fill out all currentOp details. CurOp::get(opCtx)->setGenericOpRequestDetails(opCtx, nss, command, request.body, opType); auto& readConcernArgs = repl::ReadConcernArgs::get(opCtx); auto readConcernParseStatus = readConcernArgs.initialize(request.body); if (!readConcernParseStatus.isOK()) { auto builder = replyBuilder->getBodyBuilder(); CommandHelpers::appendCommandStatusNoThrow(builder, readConcernParseStatus); return; } if (readConcernArgs.getLevel() == repl::ReadConcernLevel::kSnapshotReadConcern) { uassert(ErrorCodes::InvalidOptions, str::stream() << "read concern snapshot is not supported on mongos for the command " << commandName, invocation->supportsReadConcern(readConcernArgs.getLevel())); uassert(ErrorCodes::InvalidOptions, "read concern snapshot is not supported with atClusterTime on mongos", !readConcernArgs.getArgsAtClusterTime()); } boost::optional scopedSession; auto osi = initializeOperationSessionInfo(opCtx, request.body, command->requiresAuth(), true, true); try { if (osi.getAutocommit()) { scopedSession.emplace(opCtx); auto txnRouter = TransactionRouter::get(opCtx); invariant(txnRouter); auto txnNumber = opCtx->getTxnNumber(); invariant(txnNumber); auto startTxnSetting = osi.getStartTransaction(); bool startTransaction = startTxnSetting ? *startTxnSetting : false; uassertStatusOK(CommandHelpers::canUseTransactions(nss.db(), command->getName())); txnRouter->beginOrContinueTxn(opCtx, *txnNumber, startTransaction); } for (int tries = 0;; ++tries) { // Try kMaxNumStaleVersionRetries times. On the last try, exceptions are rethrown. bool canRetry = tries < kMaxNumStaleVersionRetries - 1; if (tries > 0) { // Re-parse before retrying in case the process of run()-ning the // invocation could affect the parsed result. invocation = command->parse(opCtx, request); invariant(invocation->ns().toString() == ns, "unexpected change of namespace when retrying"); } replyBuilder->reset(); try { execCommandClient(opCtx, invocation.get(), request, replyBuilder); return; } catch (const ExceptionForCat& ex) { const auto staleNs = [&] { if (auto staleInfo = ex.extraInfo()) { return staleInfo->getNss(); } else if (auto implicitCreateInfo = ex.extraInfo()) { return implicitCreateInfo->getNss(); } else { throw; } }(); // Send setShardVersion on this thread's versioned connections to shards (to support // commands that use the legacy (ShardConnection) versioning protocol). // // Versioned connections are a legacy concept, which is never used from code running // under a transaction (see the invariant inside ShardConnection). Because of this, // the retargeting error could not have come from a ShardConnection, so we don't // need to reset the connection's in-memory state. if (!MONGO_FAIL_POINT(doNotRefreshShardsOnRetargettingError) && !TransactionRouter::get(opCtx)) { ShardConnection::checkMyConnectionVersions(opCtx, staleNs.ns()); } Grid::get(opCtx)->catalogCache()->invalidateShardedCollection(staleNs); // Update transaction tracking state for a possible retry. Throws and aborts the // transaction if it cannot continue. if (auto txnRouter = TransactionRouter::get(opCtx)) { auto abortGuard = MakeGuard([&] { txnRouter->implicitlyAbortTransaction(opCtx); }); handleCanRetryInTransaction(opCtx, txnRouter, canRetry, ex); txnRouter->onStaleShardOrDbError(commandName); abortGuard.Dismiss(); } if (canRetry) { continue; } throw; } catch (const ExceptionFor& ex) { // Mark database entry in cache as stale. Grid::get(opCtx)->catalogCache()->onStaleDatabaseVersion(ex->getDb(), ex->getVersionReceived()); // Update transaction tracking state for a possible retry. Throws and aborts the // transaction if it cannot continue. if (auto txnRouter = TransactionRouter::get(opCtx)) { auto abortGuard = MakeGuard([&] { txnRouter->implicitlyAbortTransaction(opCtx); }); handleCanRetryInTransaction(opCtx, txnRouter, canRetry, ex); txnRouter->onStaleShardOrDbError(commandName); abortGuard.Dismiss(); } if (canRetry) { continue; } throw; } catch (const ExceptionForCat& ex) { // Simple retry on any type of snapshot error. // Update transaction tracking state for a possible retry. Throws and aborts the // transaction if it cannot continue. if (auto txnRouter = TransactionRouter::get(opCtx)) { auto abortGuard = MakeGuard([&] { txnRouter->implicitlyAbortTransaction(opCtx); }); handleCanRetryInTransaction(opCtx, txnRouter, canRetry, ex); txnRouter->onSnapshotError(); abortGuard.Dismiss(); } if (canRetry) { continue; } throw; } MONGO_UNREACHABLE; } } catch (const DBException& e) { command->incrementCommandsFailed(); LastError::get(opCtx->getClient()).setLastError(e.code(), e.reason()); auto errorLabels = getErrorLabels(osi, command->getName(), e.code()); errorBuilder->appendElements(errorLabels); throw; } } } // namespace DbResponse Strategy::queryOp(OperationContext* opCtx, const NamespaceString& nss, DbMessage* dbm) { globalOpCounters.gotQuery(); const QueryMessage q(*dbm); const auto upconvertedQuery = upconvertQueryEntry(q.query, nss, q.ntoreturn, q.ntoskip); // Set the upconverted query as the CurOp command object. CurOp::get(opCtx)->setGenericOpRequestDetails( opCtx, nss, nullptr, upconvertedQuery, dbm->msg().operation()); Client* const client = opCtx->getClient(); AuthorizationSession* const authSession = AuthorizationSession::get(client); Status status = authSession->checkAuthForFind(nss, false); audit::logQueryAuthzCheck(client, nss, q.query, status.code()); uassertStatusOK(status); LOG(3) << "query: " << q.ns << " " << redact(q.query) << " ntoreturn: " << q.ntoreturn << " options: " << q.queryOptions; if (q.queryOptions & QueryOption_Exhaust) { uasserted(18526, str::stream() << "The 'exhaust' query option is invalid for mongos queries: " << nss.ns() << " " << q.query.toString()); } // Determine the default read preference mode based on the value of the slaveOk flag. const auto defaultReadPref = q.queryOptions & QueryOption_SlaveOk ? ReadPreference::SecondaryPreferred : ReadPreference::PrimaryOnly; ReadPreferenceSetting::get(opCtx) = uassertStatusOK(ReadPreferenceSetting::fromContainingBSON(q.query, defaultReadPref)); const boost::intrusive_ptr expCtx; auto canonicalQuery = uassertStatusOK( CanonicalQuery::canonicalize(opCtx, q, expCtx, ExtensionsCallbackNoop(), MatchExpressionParser::kAllowAllSpecialFeatures)); const QueryRequest& queryRequest = canonicalQuery->getQueryRequest(); // Handle query option $maxTimeMS (not used with commands). if (queryRequest.getMaxTimeMS() > 0) { uassert(50749, "Illegal attempt to set operation deadline within DBDirectClient", !opCtx->getClient()->isInDirectClient()); opCtx->setDeadlineAfterNowBy(Milliseconds{queryRequest.getMaxTimeMS()}, ErrorCodes::MaxTimeMSExpired); } opCtx->checkForInterrupt(); // May trigger maxTimeAlwaysTimeOut fail point. // If the $explain flag was set, we must run the operation on the shards as an explain command // rather than a find command. if (queryRequest.isExplain()) { const BSONObj findCommand = queryRequest.asFindCommand(); // We default to allPlansExecution verbosity. const auto verbosity = ExplainOptions::Verbosity::kExecAllPlans; BSONObjBuilder explainBuilder; Strategy::explainFind(opCtx, findCommand, queryRequest, verbosity, ReadPreferenceSetting::get(opCtx), &explainBuilder); BSONObj explainObj = explainBuilder.done(); return replyToQuery(explainObj); } // Do the work to generate the first batch of results. This blocks waiting to get responses from // the shard(s). std::vector batch; // 0 means the cursor is exhausted. Otherwise we assume that a cursor with the returned id can // be retrieved via the ClusterCursorManager. CursorId cursorId; try { cursorId = ClusterFind::runQuery( opCtx, *canonicalQuery, ReadPreferenceSetting::get(opCtx), &batch); } catch (const ExceptionFor&) { uasserted(40247, "OP_QUERY not supported on views"); } // Fill out the response buffer. int numResults = 0; OpQueryReplyBuilder reply; for (auto&& obj : batch) { obj.appendSelfToBufBuilder(reply.bufBuilderForResults()); numResults++; } return DbResponse{reply.toQueryReply(0, // query result flags numResults, 0, // startingFrom cursorId)}; } DbResponse Strategy::clientCommand(OperationContext* opCtx, const Message& m) { auto reply = rpc::makeReplyBuilder(rpc::protocolForMessage(m)); BSONObjBuilder errorBuilder; bool propagateException = false; try { // Parse. OpMsgRequest request = [&] { try { return rpc::opMsgRequestFromAnyProtocol(m); } catch (const DBException& ex) { // If this error needs to fail the connection, propagate it out. if (ErrorCodes::isConnectionFatalMessageParseError(ex.code())) propagateException = true; LOG(1) << "Exception thrown while parsing command " << causedBy(redact(ex)); throw; } }(); // Execute. std::string db = request.getDatabase().toString(); try { LOG(3) << "Command begin db: " << db << " msg id: " << m.header().getId(); runCommand(opCtx, request, m.operation(), reply.get(), &errorBuilder); LOG(3) << "Command end db: " << db << " msg id: " << m.header().getId(); } catch (const DBException& ex) { LOG(1) << "Exception thrown while processing command on " << db << " msg id: " << m.header().getId() << causedBy(redact(ex)); // Record the exception in CurOp. CurOp::get(opCtx)->debug().errInfo = ex.toStatus(); throw; } } catch (const DBException& ex) { if (propagateException) { throw; } reply->reset(); auto bob = reply->getBodyBuilder(); CommandHelpers::appendCommandStatusNoThrow(bob, ex.toStatus()); appendRequiredFieldsToResponse(opCtx, &bob); bob.appendElements(errorBuilder.obj()); } if (OpMsg::isFlagSet(m, OpMsg::kMoreToCome)) { return {}; // Don't reply. } return DbResponse{reply->done()}; } void Strategy::commandOp(OperationContext* opCtx, const std::string& db, const BSONObj& command, const std::string& versionedNS, const BSONObj& targetingQuery, const BSONObj& targetingCollation, std::vector* results) { QuerySpec qSpec(db + ".$cmd", command, BSONObj(), 0, 1, 0); ParallelSortClusteredCursor cursor( qSpec, CommandInfo(versionedNS, targetingQuery, targetingCollation)); // Initialize the cursor cursor.init(opCtx); std::set shardIds; cursor.getQueryShardIds(shardIds); for (const ShardId& shardId : shardIds) { CommandResult result; result.shardTargetId = shardId; result.target = fassert(34417, ConnectionString::parse(cursor.getShardCursor(shardId)->originalHost())); result.result = cursor.getShardCursor(shardId)->peekFirst().getOwned(); results->push_back(result); } } DbResponse Strategy::getMore(OperationContext* opCtx, const NamespaceString& nss, DbMessage* dbm) { const int ntoreturn = dbm->pullInt(); uassert( 34424, str::stream() << "Invalid ntoreturn for OP_GET_MORE: " << ntoreturn, ntoreturn >= 0); const long long cursorId = dbm->pullInt64(); globalOpCounters.gotGetMore(); // TODO: Handle stale config exceptions here from coll being dropped or sharded during op for // now has same semantics as legacy request. auto statusGetDb = Grid::get(opCtx)->catalogCache()->getDatabase(opCtx, nss.db()); if (statusGetDb == ErrorCodes::NamespaceNotFound) { return replyToQuery(ResultFlag_CursorNotFound, nullptr, 0, 0); } uassertStatusOK(statusGetDb); boost::optional batchSize; if (ntoreturn) { batchSize = ntoreturn; } GetMoreRequest getMoreRequest(nss, cursorId, batchSize, boost::none, boost::none, boost::none); // Set the upconverted getMore as the CurOp command object. CurOp::get(opCtx)->setGenericOpRequestDetails( opCtx, nss, nullptr, getMoreRequest.toBSON(), dbm->msg().operation()); auto cursorResponse = ClusterFind::runGetMore(opCtx, getMoreRequest); if (cursorResponse == ErrorCodes::CursorNotFound) { return replyToQuery(ResultFlag_CursorNotFound, nullptr, 0, 0); } uassertStatusOK(cursorResponse.getStatus()); // Build the response document. BufBuilder buffer(FindCommon::kInitReplyBufferSize); int numResults = 0; for (const auto& obj : cursorResponse.getValue().getBatch()) { buffer.appendBuf((void*)obj.objdata(), obj.objsize()); ++numResults; } return replyToQuery(0, buffer.buf(), buffer.len(), numResults, cursorResponse.getValue().getNumReturnedSoFar().value_or(0), cursorResponse.getValue().getCursorId()); } void Strategy::killCursors(OperationContext* opCtx, DbMessage* dbm) { const int numCursors = dbm->pullInt(); massert(34425, str::stream() << "Invalid killCursors message. numCursors: " << numCursors << ", message size: " << dbm->msg().dataSize() << ".", dbm->msg().dataSize() == 8 + (8 * numCursors)); uassert(28794, str::stream() << "numCursors must be between 1 and 29999. numCursors: " << numCursors << ".", numCursors >= 1 && numCursors < 30000); globalOpCounters.gotOp(dbKillCursors, false); ConstDataCursor cursors(dbm->getArray(numCursors)); Client* const client = opCtx->getClient(); ClusterCursorManager* const manager = Grid::get(opCtx)->getCursorManager(); for (int i = 0; i < numCursors; ++i) { const CursorId cursorId = cursors.readAndAdvance>(); boost::optional nss = manager->getNamespaceForCursorId(cursorId); if (!nss) { LOG(3) << "Can't find cursor to kill. Cursor id: " << cursorId << "."; continue; } auto authzSession = AuthorizationSession::get(client); auto authChecker = [&authzSession, &nss](UserNameIterator userNames) -> Status { return authzSession->checkAuthForKillCursors(*nss, userNames); }; auto authzStatus = manager->checkAuthForKillCursors(opCtx, *nss, cursorId, authChecker); audit::logKillCursorsAuthzCheck(client, *nss, cursorId, authzStatus.code()); if (!authzStatus.isOK()) { LOG(3) << "Not authorized to kill cursor. Namespace: '" << *nss << "', cursor id: " << cursorId << "."; continue; } Status killCursorStatus = manager->killCursor(opCtx, *nss, cursorId); if (!killCursorStatus.isOK()) { LOG(3) << "Can't find cursor to kill. Namespace: '" << *nss << "', cursor id: " << cursorId << "."; continue; } LOG(3) << "Killed cursor. Namespace: '" << *nss << "', cursor id: " << cursorId << "."; } } void Strategy::writeOp(OperationContext* opCtx, DbMessage* dbm) { const auto& msg = dbm->msg(); rpc::OpMsgReplyBuilder reply; BSONObjBuilder errorBuilder; runCommand(opCtx, [&]() { switch (msg.operation()) { case dbInsert: { return InsertOp::parseLegacy(msg).serialize({}); } case dbUpdate: { return UpdateOp::parseLegacy(msg).serialize({}); } case dbDelete: { return DeleteOp::parseLegacy(msg).serialize({}); } default: MONGO_UNREACHABLE; } }(), msg.operation(), &reply, &errorBuilder); // built objects are ignored } void Strategy::explainFind(OperationContext* opCtx, const BSONObj& findCommand, const QueryRequest& qr, ExplainOptions::Verbosity verbosity, const ReadPreferenceSetting& readPref, BSONObjBuilder* out) { const auto explainCmd = ClusterExplain::wrapAsExplain(findCommand, verbosity); long long millisElapsed; std::vector shardResponses; for (int tries = 0;; ++tries) { bool canRetry = tries < 4; // Fifth try (i.e. try #4) is the last one. // We will time how long it takes to run the commands on the shards. Timer timer; try { const auto routingInfo = uassertStatusOK( Grid::get(opCtx)->catalogCache()->getCollectionRoutingInfo(opCtx, qr.nss())); shardResponses = scatterGatherVersionedTargetByRoutingTable(opCtx, qr.nss().db(), qr.nss(), routingInfo, explainCmd, readPref, Shard::RetryPolicy::kIdempotent, qr.getFilter(), qr.getCollation()); millisElapsed = timer.millis(); break; } catch (const ExceptionForCat& ex) { const auto staleNs = [&] { if (auto staleInfo = ex.extraInfo()) { return staleInfo->getNss(); } else if (auto implicitCreateInfo = ex.extraInfo()) { return implicitCreateInfo->getNss(); } else { throw; } }(); // Send setShardVersion on this thread's versioned connections to shards (to support // commands that use the legacy (ShardConnection) versioning protocol). // // Versioned connections are a legacy concept, which is never used from code running // under a transaction (see the invariant inside ShardConnection). Because of this, the // retargeting error could not have come from a ShardConnection, so we don't need to // reset the connection's in-memory state. if (!MONGO_FAIL_POINT(doNotRefreshShardsOnRetargettingError) && !TransactionRouter::get(opCtx)) { ShardConnection::checkMyConnectionVersions(opCtx, staleNs.ns()); } Grid::get(opCtx)->catalogCache()->invalidateShardedCollection(staleNs); if (canRetry) { continue; } throw; } catch (const ExceptionFor& ex) { // Mark database entry in cache as stale. Grid::get(opCtx)->catalogCache()->onStaleDatabaseVersion(ex->getDb(), ex->getVersionReceived()); if (canRetry) { continue; } throw; } catch (const ExceptionForCat&) { // Simple retry on any type of snapshot error. if (canRetry) { continue; } throw; } } const char* mongosStageName = ClusterExplain::getStageNameForReadOp(shardResponses.size(), findCommand); uassertStatusOK( ClusterExplain::buildExplainResult(opCtx, ClusterExplain::downconvert(opCtx, shardResponses), mongosStageName, millisElapsed, out)); } } // namespace mongo