/** * Copyright (C) 2018-present MongoDB, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the Server Side Public License, version 1, * as published by MongoDB, Inc. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * Server Side Public License for more details. * * You should have received a copy of the Server Side Public License * along with this program. If not, see * . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the Server Side Public License in all respects for * all of the code used other than as permitted herein. If you modify file(s) * with this exception, you may extend this exception to your version of the * file(s), but you are not obligated to do so. If you do not wish to do so, * delete this exception statement from your version. If you delete this * exception statement from all source files in the program, then also delete * it in the license file. */ #pragma once #include #include #include "mongo/db/exec/plan_stats.h" #include "mongo/db/query/canonical_query.h" #include "mongo/db/query/index_tag.h" #include "mongo/db/query/lru_key_value.h" #include "mongo/db/query/plan_cache_indexability.h" #include "mongo/db/query/query_planner_params.h" #include "mongo/platform/atomic_word.h" #include "mongo/platform/mutex.h" #include "mongo/util/container_size_helper.h" namespace mongo { /** * Represents the "key" used in the PlanCache mapping from query shape -> query plan. */ class PlanCacheKey { public: PlanCacheKey(CanonicalQuery::QueryShapeString shapeString, std::string indexabilityString) { _lengthOfStablePart = shapeString.size(); _key = std::move(shapeString); _key += indexabilityString; } CanonicalQuery::QueryShapeString getStableKey() const { return std::string(_key, 0, _lengthOfStablePart); } StringData getStableKeyStringData() const { return StringData(_key.c_str(), _lengthOfStablePart); } /** * Return the "unstable" portion of the key, which may vary across catalog changes. */ StringData getUnstablePart() const { return StringData(_key.c_str() + _lengthOfStablePart, _key.size() - _lengthOfStablePart); } StringData stringData() const { return _key; } const std::string& toString() const { return _key; } bool operator==(const PlanCacheKey& other) const { return other._key == _key && other._lengthOfStablePart == _lengthOfStablePart; } bool operator!=(const PlanCacheKey& other) const { return !(*this == other); } private: // Key is broken into two parts: // | // Combined, the two parts make up the plan cache key. We store them in one std::string so that // we can easily/cheaply extract the stable key. std::string _key; // How long the "stable key" is. size_t _lengthOfStablePart; }; std::ostream& operator<<(std::ostream& stream, const PlanCacheKey& key); StringBuilder& operator<<(StringBuilder& builder, const PlanCacheKey& key); class PlanCacheKeyHasher { public: std::size_t operator()(const PlanCacheKey& k) const { return std::hash{}(k.toString()); } }; namespace plan_ranker { struct PlanRankingDecision; } struct QuerySolution; struct QuerySolutionNode; /** * A PlanCacheIndexTree is the meaty component of the data * stored in SolutionCacheData. It is a tree structure with * index tags that indicates to the access planner which indices * it should try to use. * * How a PlanCacheIndexTree is created: * The query planner tags a match expression with indices. It * then uses the tagged tree to create a PlanCacheIndexTree, * using QueryPlanner::cacheDataFromTaggedTree. The PlanCacheIndexTree * is isomorphic to the tagged match expression, and has matching * index tags. * * How a PlanCacheIndexTree is used: * When the query planner is planning from the cache, it uses * the PlanCacheIndexTree retrieved from the cache in order to * recreate index assignments. Specifically, a raw MatchExpression * is tagged according to the index tags in the PlanCacheIndexTree. * This is done by QueryPlanner::tagAccordingToCache. */ struct PlanCacheIndexTree { /** * An OrPushdown is the cached version of an OrPushdownTag::Destination. It indicates that this * node is a predicate that can be used inside of a sibling indexed OR, to tighten index bounds * or satisfy the first field in the index. */ struct OrPushdown { uint64_t estimateObjectSizeInBytes() const { return // Add size of each element in 'route' vector. container_size_helper::estimateObjectSizeInBytes(route) + // Subtract static size of 'identifier' since it is already included in // 'sizeof(*this)'. (indexEntryId.estimateObjectSizeInBytes() - sizeof(indexEntryId)) + // Add size of the object. sizeof(*this); } IndexEntry::Identifier indexEntryId; size_t position; bool canCombineBounds; std::deque route; }; PlanCacheIndexTree() : entry(nullptr), index_pos(0), canCombineBounds(true) {} ~PlanCacheIndexTree() { for (std::vector::const_iterator it = children.begin(); it != children.end(); ++it) { delete *it; } } /** * Clone 'ie' and set 'this->entry' to be the clone. */ void setIndexEntry(const IndexEntry& ie); /** * Make a deep copy. */ PlanCacheIndexTree* clone() const; /** * For debugging. */ std::string toString(int indents = 0) const; uint64_t estimateObjectSizeInBytes() const { return // Recursively add size of each element in 'children' vector. container_size_helper::estimateObjectSizeInBytes( children, [](const auto& child) { return child->estimateObjectSizeInBytes(); }, true) + // Add size of each element in 'orPushdowns' vector. container_size_helper::estimateObjectSizeInBytes( orPushdowns, [](const auto& orPushdown) { return orPushdown.estimateObjectSizeInBytes(); }, false) + // Add size of 'entry' if present. (entry ? entry->estimateObjectSizeInBytes() : 0) + // Add size of the object. sizeof(*this); } // Children owned here. std::vector children; // Owned here. std::unique_ptr entry; size_t index_pos; // The value for this member is taken from the IndexTag of the corresponding match expression // and is used to ensure that bounds are correctly intersected and/or compounded when a query is // planned from the plan cache. bool canCombineBounds; std::vector orPushdowns; }; /** * Data stored inside a QuerySolution which can subsequently be * used to create a cache entry. When this data is retrieved * from the cache, it is sufficient to reconstruct the original * QuerySolution. */ struct SolutionCacheData { SolutionCacheData() : tree(nullptr), solnType(USE_INDEX_TAGS_SOLN), wholeIXSolnDir(1), indexFilterApplied(false) {} // Make a deep copy. SolutionCacheData* clone() const; // For debugging. std::string toString() const; uint64_t estimateObjectSizeInBytes() const { return (tree ? tree->estimateObjectSizeInBytes() : 0) + sizeof(*this); } // Owned here. If 'wholeIXSoln' is false, then 'tree' // can be used to tag an isomorphic match expression. If 'wholeIXSoln' // is true, then 'tree' is used to store the relevant IndexEntry. // If 'collscanSoln' is true, then 'tree' should be NULL. std::unique_ptr tree; enum SolutionType { // Indicates that the plan should use // the index as a proxy for a collection // scan (e.g. using index to provide sort). WHOLE_IXSCAN_SOLN, // The cached plan is a collection scan. COLLSCAN_SOLN, // Build the solution by using 'tree' // to tag the match expression. USE_INDEX_TAGS_SOLN } solnType; // The direction of the index scan used as // a proxy for a collection scan. Used only // for WHOLE_IXSCAN_SOLN. int wholeIXSolnDir; // True if index filter was applied. bool indexFilterApplied; }; class PlanCacheEntry; /** * Information returned from a get(...) query. */ class CachedSolution { private: CachedSolution(const CachedSolution&) = delete; CachedSolution& operator=(const CachedSolution&) = delete; public: CachedSolution(const PlanCacheKey& key, const PlanCacheEntry& entry); ~CachedSolution(); // Owned here. std::vector plannerData; // Key used to provide feedback on the entry. PlanCacheKey key; // For debugging. std::string toString() const; // We are extracting just enough information from the canonical // query. We could clone the canonical query but the following // items are all that is displayed to the user. BSONObj query; BSONObj sort; BSONObj projection; BSONObj collation; // The number of work cycles taken to decide on a winning plan when the plan was first // cached. size_t decisionWorks; }; /** * Used by the cache to track entries and their performance over time. * Also used by the plan cache commands to display plan cache state. */ class PlanCacheEntry { public: /** * Create a new PlanCacheEntry. * Grabs any planner-specific data required from the solutions. */ static std::unique_ptr create( const std::vector& solutions, std::unique_ptr decision, const CanonicalQuery& query, uint32_t queryHash, uint32_t planCacheKey, Date_t timeOfCreation, bool isActive, size_t works); ~PlanCacheEntry(); /** * Make a deep copy. */ std::unique_ptr clone() const; // For debugging. std::string toString() const; // // Planner data // // Data provided to the planner to allow it to recreate the solutions this entry // represents. Each SolutionCacheData is fully owned here, so in order to return // it from the cache a deep copy is made and returned inside CachedSolution. const std::vector> plannerData; // TODO: Do we really want to just hold a copy of the CanonicalQuery? For now we just // extract the data we need. // // Used by the plan cache commands to display an example query // of the appropriate shape. const BSONObj query; const BSONObj sort; const BSONObj projection; const BSONObj collation; const Date_t timeOfCreation; // Hash of the PlanCacheKey. Intended as an identifier for the query shape in logs and other // diagnostic output. const uint32_t queryHash; // Hash of the "stable" PlanCacheKey, which is the same regardless of what indexes are around. const uint32_t planCacheKey; // // Performance stats // // Information that went into picking the winning plan and also why the other plans lost. const std::unique_ptr decision; // Whether or not the cache entry is active. Inactive cache entries should not be used for // planning. bool isActive = false; // The number of "works" required for a plan to run on this shape before it becomes // active. This value is also used to determine the number of works necessary in order to // trigger a replan. Running a query of the same shape while this cache entry is inactive may // cause this value to be increased. size_t works = 0; /** * Tracks the approximate cumulative size of the plan cache entries across all the collections. */ inline static Counter64 planCacheTotalSizeEstimateBytes; private: /** * All arguments constructor. */ PlanCacheEntry(std::vector> plannerData, const BSONObj& query, const BSONObj& sort, const BSONObj& projection, const BSONObj& collation, Date_t timeOfCreation, uint32_t queryHash, uint32_t planCacheKey, std::unique_ptr decision, bool isActive, size_t works); // Ensure that PlanCacheEntry is non-copyable. PlanCacheEntry(const PlanCacheEntry&) = delete; PlanCacheEntry& operator=(const PlanCacheEntry&) = delete; uint64_t _estimateObjectSizeInBytes() const; // The total runtime size of the current object in bytes. This is the deep size, obtained by // recursively following references to all owned objects. const uint64_t _entireObjectSize; }; /** * Caches the best solution to a query. Aside from the (CanonicalQuery -> QuerySolution) * mapping, the cache contains information on why that mapping was made and statistics on the * cache entry's actual performance on subsequent runs. */ class PlanCache { private: PlanCache(const PlanCache&) = delete; PlanCache& operator=(const PlanCache&) = delete; public: // We have three states for a cache entry to be in. Rather than just 'present' or 'not // present', we use a notion of 'inactive entries' as a way of remembering how performant our // original solution to the query was. This information is useful to prevent much slower // queries from putting their plans in the cache immediately, which could cause faster queries // to run with a sub-optimal plan. Since cache entries must go through the "vetting" process of // being inactive, we protect ourselves from the possibility of simply adding a cache entry // with a very high works value which will never be evicted. enum CacheEntryState { // There is no cache entry for the given query shape. kNotPresent, // There is a cache entry for the given query shape, but it is inactive, meaning that it // should not be used when planning. kPresentInactive, // There is a cache entry for the given query shape, and it is active. kPresentActive, }; /** * Encapsulates the value returned from a call to get(). */ struct GetResult { CacheEntryState state; std::unique_ptr cachedSolution; }; /** * We don't want to cache every possible query. This function * encapsulates the criteria for what makes a canonical query * suitable for lookup/inclusion in the cache. */ static bool shouldCacheQuery(const CanonicalQuery& query); /** * If omitted, namespace set to empty string. */ PlanCache(); PlanCache(size_t size); ~PlanCache(); /** * Record solutions for query. Best plan is first element in list. * Each query in the cache will have more than 1 plan because we only * add queries which are considered by the multi plan runner (which happens * only when the query planner generates multiple candidate plans). Callers are responsible * for passing the current time so that the time the plan cache entry was created is stored * in the plan cache. * * 'worksGrowthCoefficient' specifies what multiplier to use when growing the 'works' value of * an inactive cache entry. If boost::none is provided, the function will use * 'internalQueryCacheWorksGrowthCoefficient'. * * If the mapping was set successfully, returns Status::OK(), even if it evicted another entry. */ Status set(const CanonicalQuery& query, const std::vector& solns, std::unique_ptr why, Date_t now, boost::optional worksGrowthCoefficient = boost::none); /** * Set a cache entry back to the 'inactive' state. Rather than completely evicting an entry * when the associated plan starts to perform poorly, we deactivate it, so that plans which * perform even worse than the one already in the cache may not easily take its place. */ void deactivate(const CanonicalQuery& query); /** * Look up the cached data access for the provided 'query'. Used by the query planner * to shortcut planning. * * The return value will provide the "state" of the cache entry, as well as the CachedSolution * for the query (if there is one). */ GetResult get(const CanonicalQuery& query) const; /** * Look up the cached data access for the provided PlanCacheKey. Circumvents the recalculation * of a plan cache key. * * The return value will provide the "state" of the cache entry, as well as the CachedSolution * for the query (if there is one). */ GetResult get(const PlanCacheKey& key) const; /** * If the cache entry exists and is active, return a CachedSolution. If the cache entry is * inactive, log a message and return a nullptr. If no cache entry exists, return a nullptr. */ std::unique_ptr getCacheEntryIfActive(const PlanCacheKey& key) const; /** * Remove the entry corresponding to 'ck' from the cache. Returns Status::OK() if the plan * was present and removed and an error status otherwise. */ Status remove(const CanonicalQuery& canonicalQuery); /** * Remove *all* cached plans. Does not clear index information. */ void clear(); /** * Get the cache key corresponding to the given canonical query. The query need not already * be cached. * * This is provided in the public API simply as a convenience for consumers who need some * description of query shape (e.g. index filters). * * Callers must hold the collection lock when calling this method. */ PlanCacheKey computeKey(const CanonicalQuery&) const; /** * Returns a copy of a cache entry, looked up by CanonicalQuery. * * If there is no entry in the cache for the 'query', returns an error Status. */ StatusWith> getEntry(const CanonicalQuery& cq) const; /** * Returns a vector of all cache entries. * Used by planCacheListQueryShapes and index_filter_commands_test.cpp. */ std::vector> getAllEntries() const; /** * Returns number of entries in cache. Includes inactive entries. * Used for testing. */ size_t size() const; /** * Updates internal state kept about the collection's indexes. Must be called when the set * of indexes on the associated collection have changed. * * Callers must hold the collection lock in exclusive mode when calling this method. */ void notifyOfIndexUpdates(const std::vector& indexCores); /** * Iterates over the plan cache. For each entry, serializes the PlanCacheEntry according to * 'serializationFunc'. Returns a vector of all serialized entries which match 'filterFunc'. */ std::vector getMatchingStats( const std::function& serializationFunc, const std::function& filterFunc) const; private: struct NewEntryState { bool shouldBeCreated = false; bool shouldBeActive = false; }; NewEntryState getNewEntryState(const CanonicalQuery& query, uint32_t queryHash, uint32_t planCacheKey, PlanCacheEntry* oldEntry, size_t newWorks, double growthCoefficient); LRUKeyValue _cache; // Protects _cache. mutable Mutex _cacheMutex = MONGO_MAKE_LATCH("PlanCache::_cacheMutex"); // Holds computed information about the collection's indexes. Used for generating plan // cache keys. // // Concurrent access is synchronized by the collection lock. Multiple concurrent readers // are allowed. PlanCacheIndexabilityState _indexabilityState; }; } // namespace mongo