/** * Copyright (C) 2018-present MongoDB, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the Server Side Public License, version 1, * as published by MongoDB, Inc. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * Server Side Public License for more details. * * You should have received a copy of the Server Side Public License * along with this program. If not, see * . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the Server Side Public License in all respects for * all of the code used other than as permitted herein. If you modify file(s) * with this exception, you may extend this exception to your version of the * file(s), but you are not obligated to do so. If you do not wish to do so, * delete this exception statement from your version. If you delete this * exception statement from all source files in the program, then also delete * it in the license file. */ #include "mongo/platform/basic.h" #include "mongo/db/pipeline/document_source.h" #include "mongo/db/exec/document_value/value.h" #include "mongo/db/feature_compatibility_version_documentation.h" #include "mongo/db/matcher/expression_algo.h" #include "mongo/db/pipeline/document_source_add_fields.h" #include "mongo/db/pipeline/document_source_group.h" #include "mongo/db/pipeline/document_source_internal_shard_filter.h" #include "mongo/db/pipeline/document_source_match.h" #include "mongo/db/pipeline/document_source_project.h" #include "mongo/db/pipeline/document_source_replace_root.h" #include "mongo/db/pipeline/document_source_sample.h" #include "mongo/db/pipeline/document_source_sequential_document_cache.h" #include "mongo/db/pipeline/expression_context.h" #include "mongo/db/pipeline/field_path.h" #include "mongo/db/query/plan_summary_stats_visitor.h" #include "mongo/logv2/log.h" #include "mongo/util/string_map.h" #define MONGO_LOGV2_DEFAULT_COMPONENT ::mongo::logv2::LogComponent::kQuery namespace mongo { using Parser = DocumentSource::Parser; using boost::intrusive_ptr; using std::list; using std::string; using std::vector; DocumentSource::DocumentSource(const StringData stageName, const intrusive_ptr& pCtx) : pSource(nullptr), pExpCtx(pCtx), _commonStats(stageName.rawData()) { if (pExpCtx->shouldCollectDocumentSourceExecStats()) { _commonStats.executionTime.emplace(0); } } namespace { struct ParserRegistration { Parser parser; boost::optional featureFlag; }; // Used to keep track of which DocumentSources are registered under which name. static StringMap parserMap; } // namespace void accumulatePipelinePlanSummaryStats(const Pipeline& pipeline, PlanSummaryStats& planSummaryStats) { auto visitor = PlanSummaryStatsVisitor(planSummaryStats); for (auto&& source : pipeline.getSources()) { if (auto specificStats = source->getSpecificStats()) { specificStats->acceptVisitor(&visitor); } } } void DocumentSource::registerParser(string name, Parser parser, boost::optional featureFlag) { auto it = parserMap.find(name); massert(28707, str::stream() << "Duplicate document source (" << name << ") registered.", it == parserMap.end()); parserMap[name] = {parser, featureFlag}; } void DocumentSource::registerParser(string name, SimpleParser simpleParser, boost::optional featureFlag) { Parser parser = [simpleParser = std::move(simpleParser)]( BSONElement stageSpec, const intrusive_ptr& expCtx) -> list> { return {simpleParser(std::move(stageSpec), expCtx)}; }; return registerParser(std::move(name), std::move(parser), std::move(featureFlag)); } bool DocumentSource::hasQuery() const { return false; } BSONObj DocumentSource::getQuery() const { MONGO_UNREACHABLE; } list> DocumentSource::parse( const intrusive_ptr& expCtx, BSONObj stageObj) { uassert(16435, "A pipeline stage specification object must contain exactly one field.", stageObj.nFields() == 1); BSONElement stageSpec = stageObj.firstElement(); auto stageName = stageSpec.fieldNameStringData(); // Get the registered parser and call that. auto it = parserMap.find(stageName); uassert(16436, str::stream() << "Unrecognized pipeline stage name: '" << stageName << "'", it != parserMap.end()); uassert( ErrorCodes::QueryFeatureNotAllowed, str::stream() << stageName << " is not allowed in the current feature compatibility version. See " << feature_compatibility_version_documentation::kCompatibilityLink << " for more information.", !expCtx->maxFeatureCompatibilityVersion || !it->second.featureFlag || it->second.featureFlag->isEnabledOnVersion(*expCtx->maxFeatureCompatibilityVersion)); return it->second.parser(stageSpec, expCtx); } intrusive_ptr DocumentSource::optimize() { return this; } namespace { /** * Verifies whether or not a $group is able to swap with a succeeding $match stage. While ordinarily * $group can swap with a $match, it cannot if the following $match has an $exists predicate on _id, * and the $group has exactly one field as the $group key. This is because every document will have * an _id field following such a $group stage, including those whose group key was missing before * the $group. As an example, the following optimization would be incorrect as the post-optimization * pipeline would handle documents that had nullish _id fields differently. Thus, given such a * $group and $match, this function would return false. * {$group: {_id: "$x"}} * {$match: {_id: {$exists: true}} * ----> * {$match: {x: {$exists: true}} * {$group: {_id: "$x"}} */ bool groupMatchSwapVerified(const DocumentSourceMatch& nextMatch, const DocumentSourceGroup& thisGroup) { if (thisGroup.getIdFields().size() != 1) { return true; } return !expression::hasExistencePredicateOnPath(*(nextMatch.getMatchExpression()), "_id"_sd); } } // namespace bool DocumentSource::pushMatchBefore(Pipeline::SourceContainer::iterator itr, Pipeline::SourceContainer* container) { auto nextMatch = dynamic_cast((*std::next(itr)).get()); auto thisGroup = dynamic_cast(this); if (constraints().canSwapWithMatch && nextMatch && !nextMatch->isTextQuery() && (!thisGroup || groupMatchSwapVerified(*nextMatch, *thisGroup))) { // We're allowed to swap with a $match and the stage after us is a $match. Furthermore, the // $match does not contain a text search predicate, which we do not attempt to optimize // because such a $match must already be the first stage in the pipeline. We can attempt to // swap the $match or part of the $match before ourselves. auto splitMatch = DocumentSourceMatch::splitMatchByModifiedFields(nextMatch, getModifiedPaths()); invariant(splitMatch.first || splitMatch.second); if (splitMatch.first) { // At least part of the $match can be moved before this stage. Erase the original $match // and put the independent part before this stage. If splitMatch.second is not null, // then there is a new $match stage to insert after ourselves which is dependent on the // modified fields. LOGV2_DEBUG( 5943503, 5, "Swapping all or part of a $match stage in front of another stage: ", "matchMovingBefore"_attr = redact(splitMatch.first->serializeToBSONForDebug()), "thisStage"_attr = redact(serializeToBSONForDebug()), "matchLeftAfter"_attr = redact( splitMatch.second ? splitMatch.second->serializeToBSONForDebug() : BSONObj())); container->erase(std::next(itr)); container->insert(itr, std::move(splitMatch.first)); if (splitMatch.second) { container->insert(std::next(itr), std::move(splitMatch.second)); } return true; } } return false; } bool DocumentSource::pushSampleBefore(Pipeline::SourceContainer::iterator itr, Pipeline::SourceContainer* container) { auto nextSample = dynamic_cast((*std::next(itr)).get()); if (constraints().canSwapWithSkippingOrLimitingStage && nextSample) { container->insert(itr, std::move(nextSample)); container->erase(std::next(itr)); return true; } return false; } BSONObj DocumentSource::serializeToBSONForDebug() const { std::vector serialized; serializeToArray(serialized, ExplainOptions::Verbosity::kQueryPlanner); if (serialized.empty()) { LOGV2_DEBUG(5943501, 5, "warning: stage did not serialize to anything as it was trying to be printed " "for debugging"); return BSONObj(); } if (serialized.size() > 1) { LOGV2_DEBUG(5943502, 5, "stage serialized to multiple stages. Ignoring all but the first"); } return serialized[0].getDocument().toBson(); } bool DocumentSource::pushSingleDocumentTransformBefore(Pipeline::SourceContainer::iterator itr, Pipeline::SourceContainer* container) { auto singleDocTransform = dynamic_cast((*std::next(itr)).get()); if (constraints().canSwapWithSingleDocTransform && singleDocTransform) { LOGV2_DEBUG(5943500, 5, "Swapping a single document transform stage in front of another stage: ", "singleDocTransform"_attr = redact(singleDocTransform->serializeToBSONForDebug()), "thisStage"_attr = redact(serializeToBSONForDebug())); container->insert(itr, std::move(singleDocTransform)); container->erase(std::next(itr)); return true; } return false; } Pipeline::SourceContainer::iterator DocumentSource::optimizeAt( Pipeline::SourceContainer::iterator itr, Pipeline::SourceContainer* container) { invariant(*itr == this); // Attempt to swap 'itr' with a subsequent stage, if applicable. if (attemptToPushStageBefore(itr, container)) { // The stage before the pushed before stage may be able to optimize further, if there is // such a stage. return std::prev(itr) == container->begin() ? std::prev(itr) : std::prev(std::prev(itr)); } return doOptimizeAt(itr, container); } void DocumentSource::serializeToArray(vector& array, SerializationOptions opts) const { Value entry = serialize(opts); if (!entry.missing()) { array.push_back(entry); } } namespace { std::list> throwOnParse( BSONElement spec, const boost::intrusive_ptr& expCtx) { uasserted(6047400, "Search stages are only allowed on MongoDB Atlas"); } std::unique_ptr throwOnParseLite(NamespaceString nss, const BSONElement& spec) { uasserted(6047401, "Search stages are only allowed on MongoDB Atlas"); } } // namespace MONGO_INITIALIZER_GROUP(BeginDocumentSourceRegistration, ("default"), ("EndDocumentSourceRegistration")) // Any remaining work on the parserMap should be done before finishing DocumentSource Registration. MONGO_INITIALIZER_WITH_PREREQUISITES(EndDocumentSourceRegistration, ("BeginDocumentSourceRegistration")) (InitializerContext*) { auto stageName = "$search"_sd; auto searchIt = parserMap.find(stageName); // If the $search stage has not been registered at this point, register a parser that errors // with a useful error message on parsing a search stage. if (searchIt == parserMap.end()) { LiteParsedDocumentSource::registerParser("$search", throwOnParseLite, AllowedWithApiStrict::kAlways, AllowedWithClientType::kAny); DocumentSource::registerParser("$search", throwOnParse, boost::none); LiteParsedDocumentSource::registerParser("$searchMeta", throwOnParseLite, AllowedWithApiStrict::kAlways, AllowedWithClientType::kAny); DocumentSource::registerParser("$searchMeta", throwOnParse, boost::none); } } } // namespace mongo