1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
|
// @file d_logic.h
/*
* Copyright (C) 2010 10gen Inc.
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU Affero General Public License, version 3,
* as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU Affero General Public License for more details.
*
* You should have received a copy of the GNU Affero General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*
* As a special exception, the copyright holders give permission to link the
* code of portions of this program with the OpenSSL library under certain
* conditions as described in each individual source file and distribute
* linked combinations including the program with the OpenSSL library. You
* must comply with the GNU Affero General Public License in all respects
* for all of the code used other than as permitted herein. If you modify
* file(s) with this exception, you may extend this exception to your
* version of the file(s), but you are not obligated to do so. If you do not
* wish to do so, delete this exception statement from your version. If you
* delete this exception statement from all source files in the program,
* then also delete it in the license file.
*/
#pragma once
#include "mongo/pch.h"
#include "mongo/db/jsobj.h"
#include "mongo/s/collection_metadata.h"
#include "mongo/s/chunk_version.h"
#include "mongo/util/concurrency/ticketholder.h"
#include "mongo/util/net/message.h"
namespace mongo {
class Database;
class DiskLoc;
// --------------
// --- global state ---
// --------------
class ShardingState {
public:
ShardingState();
bool enabled() const { return _enabled; }
const string& getConfigServer() const { return _configServer; }
void enable( const string& server );
// Initialize sharding state and begin authenticating outgoing connections and handling
// shard versions. If this is not run before sharded operations occur auth will not work
// and versions will not be tracked.
static void initialize(const string& server);
void gotShardName( const string& name );
bool setShardName( const string& name ); // Same as above, does not throw
string getShardName() { scoped_lock lk(_mutex); return _shardName; }
/** Reverts back to a state where this mongod is not sharded. */
void resetShardingState();
// versioning support
bool hasVersion( const string& ns );
bool hasVersion( const string& ns , ChunkVersion& version );
const ChunkVersion getVersion( const string& ns ) const;
/**
* If the metadata for 'ns' at this shard is at or above the requested version,
* 'reqShardVersion', returns OK and fills in 'latestShardVersion' with the latest shard
* version. The latter is always greater or equal than 'reqShardVersion' if in the same
* epoch.
*
* Otherwise, falls back to refreshMetadataNow.
*
* This call blocks if there are more than N threads
* currently refreshing metadata. (N is the number of
* tickets in ShardingState::_configServerTickets,
* currently 3.)
*
* Locking Note:
* + Must NOT be called with the write lock because this call may go into the network,
* and deadlocks may occur with shard-as-a-config. Therefore, nothing here guarantees
* that 'latestShardVersion' is indeed the current one on return.
*/
Status refreshMetadataIfNeeded( const string& ns,
const ChunkVersion& reqShardVersion,
ChunkVersion* latestShardVersion );
/**
* Refreshes collection metadata by asking the config server for the latest information.
* Starts a new config server request.
*
* Locking Notes:
* + Must NOT be called with the write lock because this call may go into the network,
* and deadlocks may occur with shard-as-a-config. Therefore, nothing here guarantees
* that 'latestShardVersion' is indeed the current one on return.
*
* + Because this call must not be issued with the DBLock held, by the time the config
* server sent us back the collection metadata information, someone else may have
* updated the previously stored collection metadata. There are cases when one can't
* tell which of updated or loaded metadata are the freshest. There are also cases where
* the data coming from configs do not correspond to a consistent snapshot.
* In these cases, return RemoteChangeDetected. (This usually means this call needs to
* be issued again, at caller discretion)
*
* @return OK if remote metadata successfully loaded (may or may not have been installed)
* @return RemoteChangeDetected if something changed while reloading and we may retry
* @return !OK if something else went wrong during reload
* @return latestShardVersion the version that is now stored for this collection
*/
Status refreshMetadataNow( const string& ns, ChunkVersion* latestShardVersion );
void appendInfo( BSONObjBuilder& b );
// querying support
bool needCollectionMetadata( const string& ns ) const;
CollectionMetadataPtr getCollectionMetadata( const string& ns );
// chunk migrate and split support
/**
* Creates and installs a new chunk metadata for a given collection by "forgetting" about
* one of its chunks. The new metadata uses the provided version, which has to be higher
* than the current metadata's shard version.
*
* One exception: if the forgotten chunk is the last one in this shard for the collection,
* version has to be 0.
*
* If it runs successfully, clients need to grab the new version to access the collection.
*
* LOCKING NOTE:
* Only safe to do inside the
*
* @param ns the collection
* @param min max the chunk to eliminate from the current metadata
* @param version at which the new metadata should be at
*/
void donateChunk( const string& ns , const BSONObj& min , const BSONObj& max , ChunkVersion version );
/**
* Creates and installs new chunk metadata for a given collection by reclaiming a previously
* donated chunk. The previous metadata's shard version has to be provided.
*
* If it runs successfully, clients that became stale by the previous donateChunk will be
* able to access the collection again.
*
* Note: If a migration has aborted but not yet unregistered a pending chunk, replacing the
* metadata may leave the chunk as pending - this is not dangerous and should be rare, but
* will require a stepdown to fully recover.
*
* @param ns the collection
* @param prevMetadata the previous metadata before we donated a chunk
*/
void undoDonateChunk( const string& ns, CollectionMetadataPtr prevMetadata );
/**
* Remembers a chunk range between 'min' and 'max' as a range which will have data migrated
* into it. This data can then be protected against cleanup of orphaned data.
*
* Overlapping pending ranges will be removed, so it is only safe to use this when you know
* your metadata view is definitive, such as at the start of a migration.
*
* @return false with errMsg if the range is owned by this shard
*/
bool notePending( const string& ns,
const BSONObj& min,
const BSONObj& max,
const OID& epoch,
string* errMsg );
/**
* Stops tracking a chunk range between 'min' and 'max' that previously was having data
* migrated into it. This data is no longer protected against cleanup of orphaned data.
*
* To avoid removing pending ranges of other operations, ensure that this is only used when
* a migration is still active.
* TODO: Because migrations may currently be active when a collection drops, an epoch is
* necessary to ensure the pending metadata change is still applicable.
*
* @return false with errMsg if the range is owned by the shard or the epoch of the metadata
* has changed
*/
bool forgetPending( const string& ns,
const BSONObj& min,
const BSONObj& max,
const OID& epoch,
string* errMsg );
/**
* Creates and installs a new chunk metadata for a given collection by splitting one of its
* chunks in two or more. The version for the first split chunk should be provided. The
* subsequent chunks' version would be the latter with the minor portion incremented.
*
* The effect on clients will depend on the version used. If the major portion is the same
* as the current shards, clients shouldn't perceive the split.
*
* @param ns the collection
* @param min max the chunk that should be split
* @param splitKeys point in which to split
* @param version at which the new metadata should be at
*/
void splitChunk( const string& ns , const BSONObj& min , const BSONObj& max , const vector<BSONObj>& splitKeys ,
ChunkVersion version );
/**
* Creates and installs a new chunk metadata for a given collection by merging a range of
* chunks ['minKey', 'maxKey') into a single chunk with version 'mergedVersion'.
* The current metadata must overlap the range completely and minKey and maxKey must not
* divide an existing chunk.
*
* The merged chunk version must have a greater version than the current shard version,
* and if it has a greater major version clients will need to reload metadata.
*
* @param ns the collection
* @param minKey maxKey the range which should be merged
* @param newShardVersion the shard version the newly merged chunk should have
*/
void mergeChunks( const string& ns,
const BSONObj& minKey,
const BSONObj& maxKey,
ChunkVersion mergedVersion );
bool inCriticalMigrateSection();
/**
* @return true if we are NOT in the critical section
*/
bool waitTillNotInCriticalSection( int maxSecondsToWait );
/**
* TESTING ONLY
* Uninstalls the metadata for a given collection.
*/
void resetMetadata( const string& ns );
private:
/**
* Refreshes collection metadata by asking the config server for the latest information.
* May or may not be based on a requested version.
*/
Status doRefreshMetadata( const string& ns,
const ChunkVersion& reqShardVersion,
bool useRequestedVersion,
ChunkVersion* latestShardVersion );
bool _enabled;
string _configServer;
string _shardName;
// protects state below
mutable mongo::mutex _mutex;
// protects accessing the config server
// Using a ticket holder so we can have multiple redundant tries at any given time
mutable TicketHolder _configServerTickets;
// Map from a namespace into the metadata we need for each collection on this shard
typedef map<string,CollectionMetadataPtr> CollectionMetadataMap;
CollectionMetadataMap _collMetadata;
};
extern ShardingState shardingState;
/**
* one per connection from mongos
* holds version state for each namespace
*/
class ShardedConnectionInfo {
public:
ShardedConnectionInfo();
const OID& getID() const { return _id; }
bool hasID() const { return _id.isSet(); }
void setID( const OID& id );
const ChunkVersion getVersion( const string& ns ) const;
void setVersion( const string& ns , const ChunkVersion& version );
static ShardedConnectionInfo* get( bool create );
static void reset();
static void addHook();
bool inForceVersionOkMode() const {
return _forceVersionOk;
}
void enterForceVersionOkMode() { _forceVersionOk = true; }
void leaveForceVersionOkMode() { _forceVersionOk = false; }
private:
OID _id;
bool _forceVersionOk; // if this is true, then chunk version #s aren't check, and all ops are allowed
typedef map<string,ChunkVersion> NSVersionMap;
NSVersionMap _versions;
static boost::thread_specific_ptr<ShardedConnectionInfo> _tl;
};
struct ShardForceVersionOkModeBlock {
ShardForceVersionOkModeBlock() {
info = ShardedConnectionInfo::get( false );
if ( info )
info->enterForceVersionOkMode();
}
~ShardForceVersionOkModeBlock() {
if ( info )
info->leaveForceVersionOkMode();
}
ShardedConnectionInfo * info;
};
// -----------------
// --- core ---
// -----------------
unsigned long long extractVersion( BSONElement e , string& errmsg );
/**
* @return true if we have any shard info for the ns
*/
bool haveLocalShardingInfo( const string& ns );
/**
* @return true if the current threads shard version is ok, or not in sharded version
* Also returns an error message and the Config/ChunkVersions causing conflicts
*/
bool shardVersionOk( const string& ns,
string& errmsg,
ChunkVersion& received,
ChunkVersion& wanted );
/**
* @return true if we took care of the message and nothing else should be done
*/
struct DbResponse;
bool _handlePossibleShardedMessage( Message &m, DbResponse * dbresponse );
/** What does this do? document please? */
inline bool handlePossibleShardedMessage( Message &m, DbResponse * dbresponse ) {
if( !shardingState.enabled() )
return false;
return _handlePossibleShardedMessage(m, dbresponse);
}
/**
* If a migration for the chunk in 'ns' where 'obj' lives is occurring, save this log entry
* if it's relevant. The entries saved here are later transferred to the receiving side of
* the migration. A relevant entry is an insertion, a deletion, or an update.
*/
void logOpForSharding( const char * opstr,
const char * ns,
const BSONObj& obj,
BSONObj * patt,
const BSONObj* fullObj,
bool forMigrateCleanup );
void aboutToDeleteForSharding( const StringData& ns, const Database* db , const DiskLoc& dl );
}
|