/** * Copyright (C) 2018-present MongoDB, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the Server Side Public License, version 1, * as published by MongoDB, Inc. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * Server Side Public License for more details. * * You should have received a copy of the Server Side Public License * along with this program. If not, see * . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the Server Side Public License in all respects for * all of the code used other than as permitted herein. If you modify file(s) * with this exception, you may extend this exception to your version of the * file(s), but you are not obligated to do so. If you do not wish to do so, * delete this exception statement from your version. If you delete this * exception statement from all source files in the program, then also delete * it in the license file. */ #pragma once #include #include #include #include #include "mongo/db/pipeline/document_source.h" #include "mongo/db/pipeline/lite_parsed_document_source.h" #include "mongo/db/pipeline/lite_parsed_pipeline.h" #include "mongo/db/pipeline/pipeline.h" #include "mongo/db/pipeline/tee_buffer.h" namespace mongo { class BSONElement; class TeeBuffer; class DocumentSourceTeeConsumer; class ExpressionContext; class NamespaceString; /** * A $facet stage contains multiple sub-pipelines. Each input to the $facet stage will feed into * each of the sub-pipelines. The $facet stage is blocking, and outputs only one document, * containing an array of results for each sub-pipeline. * * For example, {$facet: {facetA: [{$skip: 1}], facetB: [{$limit: 1}]}} would describe a $facet * stage which will produce a document like the following: * {facetA: [], facetB: []}. */ class DocumentSourceFacet final : public DocumentSource { public: static constexpr StringData kStageName = "$facet"_sd; static constexpr StringData kTeeConsumerStageName = "$internalFacetTeeConsumer"_sd; struct FacetPipeline { FacetPipeline(std::string name, std::unique_ptr pipeline) : name(std::move(name)), pipeline(std::move(pipeline)) {} std::string name; std::unique_ptr pipeline; }; class LiteParsed final : public LiteParsedDocumentSourceNestedPipelines { public: static std::unique_ptr parse(const NamespaceString& nss, const BSONElement& spec); LiteParsed(std::string parseTimeName, std::vector pipelines) : LiteParsedDocumentSourceNestedPipelines( std::move(parseTimeName), boost::none, std::move(pipelines)) {} PrivilegeVector requiredPrivileges(bool isMongos, bool bypassDocumentValidation) const override final { PrivilegeVector requiredPrivileges; for (auto&& pipeline : _pipelines) { Privilege::addPrivilegesToPrivilegeVector( &requiredPrivileges, pipeline.requiredPrivileges(isMongos, bypassDocumentValidation)); } return requiredPrivileges; } }; static boost::intrusive_ptr createFromBson( BSONElement elem, const boost::intrusive_ptr& pExpCtx); static boost::intrusive_ptr create( std::vector facetPipelines, const boost::intrusive_ptr& expCtx, size_t bufferSizeBytes = internalQueryFacetBufferSizeBytes.load(), size_t maxOutputDocBytes = internalQueryFacetMaxOutputDocSizeBytes.load()); /** * Optimizes inner pipelines. */ boost::intrusive_ptr optimize() final; /** * Takes a union of all sub-pipelines, and adds them to 'deps'. */ DepsTracker::State getDependencies(DepsTracker* deps) const final; void addVariableRefs(std::set* refs) const final; const char* getSourceName() const final { return DocumentSourceFacet::kStageName.rawData(); } /** * Sets 'source' as the source of '_teeBuffer'. */ void setSource(DocumentSource* source) final; /** * The $facet stage must be run on the merging shard. * * TODO SERVER-24154: Should be smarter about splitting so that parts of the sub-pipelines can * potentially be run in parallel on multiple shards. */ boost::optional distributedPlanLogic() final { // {shardsStage, mergingStage, sortPattern} return DistributedPlanLogic{nullptr, this, boost::none}; } const std::vector& getFacetPipelines() const { return _facets; } auto& getFacetPipelines() { return _facets; } // The following are overridden just to forward calls to sub-pipelines. void addInvolvedCollections(stdx::unordered_set* involvedNssSet) const final; void detachFromOperationContext() final; void reattachToOperationContext(OperationContext* opCtx) final; bool validateOperationContext(const OperationContext* opCtx) const final; StageConstraints constraints(Pipeline::SplitState pipeState) const final; bool usedDisk() final; const SpecificStats* getSpecificStats() const final { return &_stats; } protected: /** * Blocking call. Will consume all input and produces one output document. */ GetNextResult doGetNext() final; void doDispose() final; private: DocumentSourceFacet(std::vector facetPipelines, const boost::intrusive_ptr& expCtx, size_t bufferSizeBytes, size_t maxOutputDocBytes); Value serialize(SerializationOptions opts = SerializationOptions()) const final override; boost::intrusive_ptr _teeBuffer; std::vector _facets; const size_t _maxOutputDocSizeBytes; bool _done = false; DocumentSourceFacetStats _stats; }; } // namespace mongo