1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
|
/**
* Copyright (C) 2014 MongoDB Inc.
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU Affero General Public License, version 3,
* as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Affero General Public License for more details.
*
* You should have received a copy of the GNU Affero General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*
* As a special exception, the copyright holders give permission to link the
* code of portions of this program with the OpenSSL library under certain
* conditions as described in each individual source file and distribute
* linked combinations including the program with the OpenSSL library. You
* must comply with the GNU Affero General Public License in all respects for
* all of the code used other than as permitted herein. If you modify file(s)
* with this exception, you may extend this exception to your version of the
* file(s), but you are not obligated to do so. If you do not wish to do so,
* delete this exception statement from your version. If you delete this
* exception statement from all source files in the program, then also delete
* it in the license file.
*/
#pragma once
#include <boost/next_prior.hpp>
#include <boost/shared_ptr.hpp>
#include <map>
#include <string>
#include <vector>
#include "mongo/s/chunk.h"
namespace mongo {
class CanonicalQuery;
class ChunkManager;
class CollectionType;
struct QuerySolutionNode;
typedef boost::shared_ptr<ChunkManager> ChunkManagerPtr;
// The key for the map is max for each Chunk or ChunkRange
typedef std::map<BSONObj, boost::shared_ptr<const Chunk>, BSONObjCmp> ChunkMap;
class ChunkRange {
public:
ChunkRange(ChunkMap::const_iterator begin, const ChunkMap::const_iterator end);
// Merge min and max (must be adjacent ranges)
ChunkRange(const ChunkRange& min, const ChunkRange& max);
const ChunkManager* getManager() const { return _manager; }
Shard getShard() const { return _shard; }
const BSONObj& getMin() const { return _min; }
const BSONObj& getMax() const { return _max; }
// clones of Chunk methods
// Returns true if this ChunkRange contains the given shard key, and false otherwise
//
// Note: this function takes an extracted *key*, not an original document
// (the point may be computed by, say, hashing a given field or projecting
// to a subset of fields).
bool containsKey( const BSONObj& shardKey ) const;
std::string toString() const;
private:
const ChunkManager* _manager;
const Shard _shard;
const BSONObj _min;
const BSONObj _max;
};
typedef std::map<BSONObj, boost::shared_ptr<ChunkRange>, BSONObjCmp> ChunkRangeMap;
class ChunkRangeManager {
public:
const ChunkRangeMap& ranges() const { return _ranges; }
void clear() { _ranges.clear(); }
void reloadAll(const ChunkMap& chunks);
// Slow operation -- wrap with DEV
void assertValid() const;
ChunkRangeMap::const_iterator upper_bound(const BSONObj& o) const { return _ranges.upper_bound(o); }
ChunkRangeMap::const_iterator lower_bound(const BSONObj& o) const { return _ranges.lower_bound(o); }
private:
// assumes nothing in this range exists in _ranges
void _insertRange(ChunkMap::const_iterator begin, const ChunkMap::const_iterator end);
ChunkRangeMap _ranges;
};
/* config.sharding
{ ns: 'alleyinsider.fs.chunks' ,
key: { ts : 1 } ,
shards: [ { min: 1, max: 100, server: a } , { min: 101, max: 200 , server : b } ]
}
*/
class ChunkManager {
public:
typedef std::map<std::string, ChunkVersion> ShardVersionMap;
// Loads a new chunk manager from a collection document
explicit ChunkManager(const CollectionType& coll);
// Creates an empty chunk manager for the namespace
ChunkManager( const std::string& ns, const ShardKeyPattern& pattern, bool unique );
std::string getns() const { return _ns; }
const ShardKeyPattern& getShardKeyPattern() const { return _keyPattern; }
bool isUnique() const { return _unique; }
/**
* this is just an increasing number of how many ChunkManagers we have so we know if something has been updated
*/
unsigned long long getSequenceNumber() const { return _sequenceNumber; }
//
// After constructor is invoked, we need to call loadExistingRanges. If this is a new
// sharded collection, we can call createFirstChunks first.
//
// Creates new chunks based on info in chunk manager
void createFirstChunks(const Shard& primary,
const std::vector<BSONObj>* initPoints,
const std::vector<Shard>* initShards);
// Loads existing ranges based on info in chunk manager
void loadExistingRanges(const ChunkManager* oldManager);
// Helpers for load
void calcInitSplitsAndShards( const Shard& primary,
const std::vector<BSONObj>* initPoints,
const std::vector<Shard>* initShards,
std::vector<BSONObj>* splitPoints,
std::vector<Shard>* shards ) const;
//
// Methods to use once loaded / created
//
int numChunks() const { return _chunkMap.size(); }
/**
* Given a key that has been extracted from a document, returns the
* chunk that contains that key.
*
* For instance, to locate the chunk for document {a : "foo" , b : "bar"}
* when the shard key is {a : "hashed"}, you can call
* findIntersectingChunk() on {a : hash("foo") }
*/
ChunkPtr findIntersectingChunk( const BSONObj& shardKey ) const;
void getShardsForQuery( std::set<Shard>& shards , const BSONObj& query ) const;
void getAllShards( std::set<Shard>& all ) const;
/** @param shards set to the shards covered by the interval [min, max], see SERVER-4791 */
void getShardsForRange( std::set<Shard>& shards, const BSONObj& min, const BSONObj& max ) const;
// Transforms query into bounds for each field in the shard key
// for example :
// Key { a: 1, b: 1 },
// Query { a : { $gte : 1, $lt : 2 },
// b : { $gte : 3, $lt : 4 } }
// => Bounds { a : [1, 2), b : [3, 4) }
static IndexBounds getIndexBoundsForQuery(const BSONObj& key, const CanonicalQuery* canonicalQuery);
// Collapse query solution tree.
//
// If it has OR node, the result could be a superset of the index bounds generated.
// Since to give a single IndexBounds, this gives the union of bounds on each field.
// for example:
// OR: { a: (0, 1), b: (0, 1) },
// { a: (2, 3), b: (2, 3) }
// => { a: (0, 1), (2, 3), b: (0, 1), (2, 3) }
static IndexBounds collapseQuerySolution( const QuerySolutionNode* node );
const ChunkMap& getChunkMap() const { return _chunkMap; }
/**
* Returns true if, for this shard, the chunks are identical in both chunk managers
*/
bool compatibleWith(const ChunkManager& other, const std::string& shard) const;
std::string toString() const;
ChunkVersion getVersion(const std::string& shardName) const;
ChunkVersion getVersion() const;
void _printChunks() const;
int getCurrentDesiredChunkSize() const;
ChunkManagerPtr reload(bool force=true) const; // doesn't modify self!
void markMinorForReload( ChunkVersion majorVersion ) const;
void getMarkedMinorVersions( std::set<ChunkVersion>& minorVersions ) const;
private:
// helpers for loading
// returns true if load was consistent
bool _load(ChunkMap& chunks,
std::set<Shard>& shards,
ShardVersionMap* shardVersions,
const ChunkManager* oldManager);
static bool _isValid(const ChunkMap& chunks);
// end helpers
// All members should be const for thread-safety
const std::string _ns;
const ShardKeyPattern _keyPattern;
const bool _unique;
const ChunkMap _chunkMap;
const ChunkRangeManager _chunkRanges;
const std::set<Shard> _shards;
const ShardVersionMap _shardVersions; // max version per shard
// max version of any chunk
ChunkVersion _version;
mutable mutex _mutex; // only used with _nsLock
const unsigned long long _sequenceNumber;
//
// Split Heuristic info
//
class SplitHeuristics {
public:
SplitHeuristics()
: _splitTickets(maxParallelSplits) {
}
TicketHolder _splitTickets;
// Test whether we should split once data * splitTestFactor > chunkSize (approximately)
static const int splitTestFactor = 5;
// Maximum number of parallel threads requesting a split
static const int maxParallelSplits = 5;
// The idea here is that we're over-aggressive on split testing by a factor of
// splitTestFactor, so we can safely wait until we get to splitTestFactor invalid splits
// before changing. Unfortunately, we also potentially over-request the splits by a
// factor of maxParallelSplits, but since the factors are identical it works out
// (for now) for parallel or sequential oversplitting.
// TODO: Make splitting a separate thread with notifications?
static const int staleMinorReloadThreshold = maxParallelSplits;
};
mutable SplitHeuristics _splitHeuristics;
//
// End split heuristics
//
friend class Chunk;
friend class ChunkRangeManager; // only needed for CRM::assertValid()
static AtomicUInt32 NextSequenceNumber;
friend class TestableChunkManager;
};
} // namespace mongo
|