// @file chunk.h /** * Copyright (C) 2008 10gen Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU Affero General Public License, version 3, * as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU Affero General Public License for more details. * * You should have received a copy of the GNU Affero General Public License * along with this program. If not, see . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the GNU Affero General Public License in all respects * for all of the code used other than as permitted herein. If you modify * file(s) with this exception, you may extend this exception to your * version of the file(s), but you are not obligated to do so. If you do not * wish to do so, delete this exception statement from your version. If you * delete this exception statement from all source files in the program, * then also delete it in the license file. */ #pragma once #include "mongo/base/string_data.h" #include "mongo/bson/util/atomic_int.h" #include "mongo/client/distlock.h" #include "mongo/s/chunk_version.h" #include "mongo/s/shard.h" #include "mongo/s/shardkey.h" #include "mongo/util/concurrency/ticketholder.h" namespace mongo { class DBConfig; class Chunk; class ChunkRange; class ChunkManager; class ChunkObjUnitTest; typedef shared_ptr ChunkPtr; // key is max for each Chunk or ChunkRange typedef map ChunkMap; typedef map,BSONObjCmp> ChunkRangeMap; typedef shared_ptr ChunkManagerPtr; /** config.chunks { ns : "alleyinsider.fs.chunks" , min : {} , max : {} , server : "localhost:30001" } x is in a shard iff min <= x < max */ class Chunk : boost::noncopyable { public: Chunk( const ChunkManager * info , BSONObj from); Chunk( const ChunkManager * info , const BSONObj& min, const BSONObj& max, const Shard& shard, ChunkVersion lastmod = ChunkVersion() ); // // serialization support // void serialize(BSONObjBuilder& to, ChunkVersion myLastMod=ChunkVersion(0,OID())); // // chunk boundary support // const BSONObj& getMin() const { return _min; } const BSONObj& getMax() const { return _max; } // if min/max key is pos/neg infinity bool minIsInf() const; bool maxIsInf() const; // Returns true if this chunk contains the given point, and false otherwise // // Note: this function takes an extracted *key*, not an original document // (the point may be computed by, say, hashing a given field or projecting // to a subset of fields). bool containsPoint( const BSONObj& point ) const; string genID() const; static string genID( const string& ns , const BSONObj& min ); // // chunk version support // void appendShortVersion( const char * name , BSONObjBuilder& b ) const; ChunkVersion getLastmod() const { return _lastmod; } void setLastmod( ChunkVersion v ) { _lastmod = v; } // // split support // long getBytesWritten() const { return _dataWritten; } // Const since _dataWritten is mutable and a heuristic // TODO: Split data tracking and chunk information void setBytesWritten( long bytesWritten ) const { _dataWritten = bytesWritten; } /** * if the amount of data written nears the max size of a shard * then we check the real size, and if its too big, we split * @return if something was split */ bool splitIfShould( long dataWritten ) const; /** * Splits this chunk at a non-specificed split key to be chosen by the mongod holding this chunk. * * @param force if set to true, will split the chunk regardless if the split is really necessary size wise * if set to false, will only split if the chunk has reached the currently desired maximum size * @param res the object containing details about the split execution * @return splitPoint if found a key and split successfully, else empty BSONObj */ BSONObj singleSplit( bool force , BSONObj& res ) const; /** * Splits this chunk at the given key (or keys) * * @param splitPoints the vector of keys that should be used to divide this chunk * @param res the object containing details about the split execution * @return if the split was successful */ bool multiSplit( const vector& splitPoints , BSONObj& res ) const; /** * Asks the mongod holding this chunk to find a key that approximately divides this chunk in two * * @param medianKey the key that divides this chunk, if there is one, or empty */ void pickMedianKey( BSONObj& medianKey ) const; /** * @param splitPoints vector to be filled in * @param chunkSize chunk size to target in bytes * @param maxPoints limits the number of split points that are needed, zero is max (optional) * @param maxObjs limits the number of objects in each chunk, zero is as max (optional) */ void pickSplitVector( vector& splitPoints , int chunkSize , int maxPoints = 0, int maxObjs = 0) const; // // migration support // /** * Issues a migrate request for this chunk * * @param to shard to move this chunk to * @param chunSize maximum number of bytes beyond which the migrate should no go trhough * @param secondaryThrottle whether during migrate all writes should block for repl * @param waitForDelete whether chunk move should wait for cleanup or return immediately * @param maxTimeMS max time for the migrate request * @param res the object containing details about the migrate execution * @return true if move was successful */ bool moveAndCommit(const Shard& to, long long chunkSize, bool secondaryThrottle, bool waitForDelete, int maxTimeMS, BSONObj& res) const; /** * @return size of shard in bytes * talks to mongod to do this */ long getPhysicalSize() const; /** * marks this chunk as a jumbo chunk * that means the chunk will be inelligble for migrates */ void markAsJumbo() const; bool isJumbo() const { return _jumbo; } /** * Attempt to refresh maximum chunk size from config. */ static void refreshChunkSize(); /** * sets MaxChunkSize * 1 <= newMaxChunkSize <= 1024 * @return true if newMaxChunkSize is valid and was set */ static bool setMaxChunkSizeSizeMB( int newMaxChunkSize ); // // public constants // static int MaxChunkSize; static int MaxObjectPerChunk; static bool ShouldAutoSplit; // // accessors and helpers // string toString() const; friend ostream& operator << (ostream& out, const Chunk& c) { return (out << c.toString()); } // chunk equality is determined by comparing the min and max bounds of the chunk bool operator==(const Chunk& s) const; bool operator!=(const Chunk& s) const { return ! ( *this == s ); } string getns() const; Shard getShard() const { return _shard; } const ChunkManager* getManager() const { return _manager; } private: // main shard info const ChunkManager * _manager; BSONObj _min; BSONObj _max; Shard _shard; ChunkVersion _lastmod; mutable bool _jumbo; // transient stuff mutable long _dataWritten; // methods, etc.. /** Returns the highest or lowest existing value in the shard-key space. * Warning: this assumes that the shard key is not "special"- that is, the shardKeyPattern * is simply an ordered list of ascending/descending field names. Examples: * {a : 1, b : -1} is not special. {a : "hashed"} is. * * if sort 1, return lowest key * if sort -1, return highest key * will return empty object if have none */ BSONObj _getExtremeKey( int sort ) const; /** initializes _dataWritten with a random value so that a mongos restart wouldn't cause delay in splitting */ static int mkDataWritten(); ShardKeyPattern skey() const; }; class ChunkRange { public: const ChunkManager* getManager() const { return _manager; } Shard getShard() const { return _shard; } const BSONObj& getMin() const { return _min; } const BSONObj& getMax() const { return _max; } // clones of Chunk methods // Returns true if this ChunkRange contains the given point, and false otherwise // // Note: this function takes an extracted *key*, not an original document // (the point may be computed by, say, hashing a given field or projecting // to a subset of fields). bool containsPoint( const BSONObj& point ) const; ChunkRange(ChunkMap::const_iterator begin, const ChunkMap::const_iterator end) : _manager(begin->second->getManager()) , _shard(begin->second->getShard()) , _min(begin->second->getMin()) , _max(boost::prior(end)->second->getMax()) { verify( begin != end ); DEV while (begin != end) { verify(begin->second->getManager() == _manager); verify(begin->second->getShard() == _shard); ++begin; } } // Merge min and max (must be adjacent ranges) ChunkRange(const ChunkRange& min, const ChunkRange& max) : _manager(min.getManager()) , _shard(min.getShard()) , _min(min.getMin()) , _max(max.getMax()) { verify(min.getShard() == max.getShard()); verify(min.getManager() == max.getManager()); verify(min.getMax() == max.getMin()); } friend ostream& operator<<(ostream& out, const ChunkRange& cr) { return (out << "ChunkRange(min=" << cr._min << ", max=" << cr._max << ", shard=" << cr._shard <<")"); } private: const ChunkManager* _manager; const Shard _shard; const BSONObj _min; const BSONObj _max; }; class ChunkRangeManager { public: const ChunkRangeMap& ranges() const { return _ranges; } void clear() { _ranges.clear(); } void reloadAll(const ChunkMap& chunks); // Slow operation -- wrap with DEV void assertValid() const; ChunkRangeMap::const_iterator upper_bound(const BSONObj& o) const { return _ranges.upper_bound(o); } ChunkRangeMap::const_iterator lower_bound(const BSONObj& o) const { return _ranges.lower_bound(o); } private: // assumes nothing in this range exists in _ranges void _insertRange(ChunkMap::const_iterator begin, const ChunkMap::const_iterator end); ChunkRangeMap _ranges; }; /* config.sharding { ns: 'alleyinsider.fs.chunks' , key: { ts : 1 } , shards: [ { min: 1, max: 100, server: a } , { min: 101, max: 200 , server : b } ] } */ class ChunkManager { public: typedef map ShardVersionMap; // Loads a new chunk manager from a collection document ChunkManager( const BSONObj& collDoc ); // Creates an empty chunk manager for the namespace ChunkManager( const string& ns, const ShardKeyPattern& pattern, bool unique ); // Updates a chunk manager based on an older manager ChunkManager( ChunkManagerPtr oldManager ); string getns() const { return _ns; } const ShardKeyPattern& getShardKey() const { return _key; } bool hasShardKey( const BSONObj& obj ) const; bool isUnique() const { return _unique; } /** * this is just an increasing number of how many ChunkManagers we have so we know if something has been updated */ unsigned long long getSequenceNumber() const { return _sequenceNumber; } // // After constructor is invoked, we need to call loadExistingRanges. If this is a new // sharded collection, we can call createFirstChunks first. // // Creates new chunks based on info in chunk manager void createFirstChunks( const string& config, const Shard& primary, const vector* initPoints, const vector* initShards ); // Loads existing ranges based on info in chunk manager void loadExistingRanges( const string& config ); // Helpers for load void calcInitSplitsAndShards( const Shard& primary, const vector* initPoints, const vector* initShards, vector* splitPoints, vector* shards ) const; // // Methods to use once loaded / created // int numChunks() const { return _chunkMap.size(); } /** Given a document, returns the chunk which contains that document. * This works by extracting the shard key part of the given document, then * calling findIntersectingChunk() on the extracted key. * * See also the description for findIntersectingChunk(). */ ChunkPtr findChunkForDoc( const BSONObj& doc ) const; /** Given a key that has been extracted from a document, returns the * chunk that contains that key. * * For instance, to locate the chunk for document {a : "foo" , b : "bar"} * when the shard key is {a : "hashed"}, you can call * findChunkForDoc() on {a : "foo" , b : "bar"}, or * findIntersectingChunk() on {a : hash("foo") } */ ChunkPtr findIntersectingChunk( const BSONObj& point ) const; ChunkPtr findChunkOnServer( const Shard& shard ) const; void getShardsForQuery( set& shards , const BSONObj& query ) const; void getAllShards( set& all ) const; /** @param shards set to the shards covered by the interval [min, max], see SERVER-4791 */ void getShardsForRange( set& shards, const BSONObj& min, const BSONObj& max ) const; ChunkMap getChunkMap() const { return _chunkMap; } /** * Returns true if, for this shard, the chunks are identical in both chunk managers */ bool compatibleWith( const ChunkManager& other, const Shard& shard ) const; bool compatibleWith( ChunkManagerPtr other, const Shard& shard ) const { if( ! other ) return false; return compatibleWith( *other, shard ); } bool compatibleWith( const Chunk& other ) const; bool compatibleWith( ChunkPtr other ) const { if( ! other ) return false; return compatibleWith( *other ); } string toString() const; ChunkVersion getVersion( const StringData& shardName ) const; ChunkVersion getVersion( const Shard& shard ) const; ChunkVersion getVersion() const; void getInfo( BSONObjBuilder& b ) const; /** * @param me - so i don't get deleted before i'm done */ void drop( ChunkManagerPtr me ) const; void _printChunks() const; int getCurrentDesiredChunkSize() const; ChunkManagerPtr reload(bool force=true) const; // doesn't modify self! void markMinorForReload( ChunkVersion majorVersion ) const; void getMarkedMinorVersions( set& minorVersions ) const; private: // helpers for loading // returns true if load was consistent bool _load( const string& config, ChunkMap& chunks, set& shards, ShardVersionMap& shardVersions, ChunkManagerPtr oldManager); static bool _isValid(const ChunkMap& chunks); // end helpers // All members should be const for thread-safety const string _ns; const ShardKeyPattern _key; const bool _unique; const ChunkMap _chunkMap; const ChunkRangeManager _chunkRanges; const set _shards; const ShardVersionMap _shardVersions; // max version per shard // max version of any chunk ChunkVersion _version; // the previous manager this was based on // cleared after loading chunks ChunkManagerPtr _oldManager; mutable mutex _mutex; // only used with _nsLock const unsigned long long _sequenceNumber; // // Split Heuristic info // class SplitHeuristics { public: SplitHeuristics() : _splitTickets( maxParallelSplits ), _staleMinorSetMutex( "SplitHeuristics::staleMinorSet" ), _staleMinorCount( 0 ) {} void markMinorForReload( const string& ns, ChunkVersion majorVersion ); void getMarkedMinorVersions( set& minorVersions ); TicketHolder _splitTickets; mutex _staleMinorSetMutex; // mutex protects below int _staleMinorCount; set _staleMinorSet; // Test whether we should split once data * splitTestFactor > chunkSize (approximately) static const int splitTestFactor = 5; // Maximum number of parallel threads requesting a split static const int maxParallelSplits = 5; // The idea here is that we're over-aggressive on split testing by a factor of // splitTestFactor, so we can safely wait until we get to splitTestFactor invalid splits // before changing. Unfortunately, we also potentially over-request the splits by a // factor of maxParallelSplits, but since the factors are identical it works out // (for now) for parallel or sequential oversplitting. // TODO: Make splitting a separate thread with notifications? static const int staleMinorReloadThreshold = maxParallelSplits; }; mutable SplitHeuristics _splitHeuristics; // // End split heuristics // friend class Chunk; friend class ChunkRangeManager; // only needed for CRM::assertValid() static AtomicUInt NextSequenceNumber; /** Just for testing */ friend class TestableChunkManager; ChunkManager(); }; // like BSONObjCmp. for use as an STL comparison functor // key-order in "order" argument must match key-order in shardkey class ChunkCmp { public: ChunkCmp( const BSONObj &order = BSONObj() ) : _cmp( order ) {} bool operator()( const Chunk &l, const Chunk &r ) const { return _cmp(l.getMin(), r.getMin()); } bool operator()( const ptr l, const ptr r ) const { return operator()(*l, *r); } // Also support ChunkRanges bool operator()( const ChunkRange &l, const ChunkRange &r ) const { return _cmp(l.getMin(), r.getMin()); } bool operator()( const shared_ptr l, const shared_ptr r ) const { return operator()(*l, *r); } private: BSONObjCmp _cmp; }; /* struct chunk_lock { chunk_lock( const Chunk* c ){ } Chunk _c; }; */ inline string Chunk::genID() const { return genID(_manager->getns(), _min); } bool setShardVersion( DBClientBase & conn, const string& ns, ChunkVersion version, ChunkManagerPtr manager, bool authoritative, BSONObj& result ); } // namespace mongo