// repl.cpp /* TODO PAIRING _ on a syncexception, don't allow going back to master state? */ /** * Copyright (C) 2008 10gen Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU Affero General Public License, version 3, * as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU Affero General Public License for more details. * * You should have received a copy of the GNU Affero General Public License * along with this program. If not, see . */ /* Collections we use: local.sources - indicates what sources we pull from as a "slave", and the last update of each local.oplog.$main - our op log as "master" local.dbinfo. local.pair.startup - can contain a special value indicating for a pair that we have the master copy. used when replacing other half of the pair which has permanently failed. local.pair.sync - { initialsynccomplete: 1 } */ #include "stdafx.h" #include "jsobj.h" #include "../util/goodies.h" #include "repl.h" #include "../util/message.h" #include "../client/dbclient.h" #include "../client/connpool.h" #include "pdfile.h" #include "query.h" #include "db.h" #include "commands.h" #include "security.h" #include "cmdline.h" namespace mongo { ReplSettings replSettings; void ensureHaveIdIndex(const char *ns); /* if 1 sync() is running */ int syncing = 0; /* if true replace our peer in a replication pair -- don't worry about if his local.oplog.$main is empty. */ bool replacePeer = false; /* "dead" means something really bad happened like replication falling completely out of sync. when non-null, we are dead and the string is informational */ const char *replAllDead = 0; time_t lastForcedResync = 0; IdTracker &idTracker = *( new IdTracker() ); } // namespace mongo #include "replset.h" namespace mongo { PairSync *pairSync = new PairSync(); bool getInitialSyncCompleted() { return pairSync->initialSyncCompleted(); } /* --- ReplPair -------------------------------- */ ReplPair *replPair = 0; /* output by the web console */ const char *replInfo = ""; struct ReplInfo { ReplInfo(const char *msg) { replInfo = msg; } ~ReplInfo() { replInfo = "?"; } }; void ReplPair::setMaster(int n, const char *_comment ) { if ( n == State_Master && !getInitialSyncCompleted() ) return; info = _comment; if ( n != state && !cmdLine.quiet ) log() << "pair: setting master=" << n << " was " << state << '\n'; state = n; } /* peer unreachable, try our arbiter */ void ReplPair::arbitrate() { ReplInfo r("arbitrate"); if ( arbHost == "-" ) { // no arbiter. we are up, let's assume partner is down and network is not partitioned. setMasterLocked(State_Master, "remote unreachable"); return; } auto_ptr conn( newClientConnection() ); string errmsg; if ( !conn->connect(arbHost.c_str(), errmsg) ) { log() << "repl: cantconn arbiter " << errmsg << endl; setMasterLocked(State_CantArb, "can't connect to arb"); return; } negotiate( conn.get(), "arbiter" ); } /* --------------------------------------------- */ class CmdReplacePeer : public Command { public: virtual bool slaveOk() { return true; } virtual bool adminOnly() { return true; } virtual bool logTheOp() { return false; } virtual LockType locktype(){ return WRITE; } CmdReplacePeer() : Command("replacepeer") { } virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool fromRepl) { if ( replPair == 0 ) { errmsg = "not paired"; return false; } if ( !getInitialSyncCompleted() ) { errmsg = "not caught up cannot replace peer"; return false; } if ( syncing < 0 ) { errmsg = "replacepeer already invoked"; return false; } Timer t; while ( 1 ) { if ( syncing == 0 || t.millis() > 20000 ) break; { dbtemprelease t; sleepmillis(10); } } if ( syncing ) { assert( syncing > 0 ); errmsg = "timeout waiting for sync() to finish"; return false; } { ReplSource::SourceVector sources; ReplSource::loadAll(sources); if ( sources.size() != 1 ) { errmsg = "local.sources.count() != 1, cannot replace peer"; return false; } } { Helpers::emptyCollection("local.sources"); BSONObj o = fromjson("{\"replacepeer\":1}"); Helpers::putSingleton("local.pair.startup", o); } syncing = -1; replAllDead = "replacepeer invoked -- adjust local.sources hostname then restart this db process"; result.append("info", "adjust local.sources hostname; db restart now required"); return true; } } cmdReplacePeer; class CmdForceDead : public Command { public: virtual bool slaveOk() { return true; } virtual bool adminOnly() { return true; } virtual bool logTheOp() { return false; } virtual LockType locktype(){ return WRITE; } CmdForceDead() : Command("forcedead") { } virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool fromRepl) { replAllDead = "forced by command"; return true; } } cmdForceDead; /* operator requested resynchronization of replication (on the slave). { resync : 1 } */ class CmdResync : public Command { public: virtual bool slaveOk() { return true; } virtual bool adminOnly() { return true; } virtual bool logTheOp() { return false; } virtual LockType locktype(){ return WRITE; } CmdResync() : Command("resync") { } virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool fromRepl) { if ( cmdObj.getBoolField( "force" ) ) { if ( !waitForSyncToFinish( errmsg ) ) return false; replAllDead = "resync forced"; } if ( !replAllDead ) { errmsg = "not dead, no need to resync"; return false; } if ( !waitForSyncToFinish( errmsg ) ) return false; ReplSource::forceResyncDead( "client" ); result.append( "info", "triggered resync for all sources" ); return true; } bool waitForSyncToFinish( string &errmsg ) const { // Wait for slave thread to finish syncing, so sources will be be // reloaded with new saved state on next pass. Timer t; while ( 1 ) { if ( syncing == 0 || t.millis() > 20000 ) break; { dbtemprelease t; sleepmillis(10); } } if ( syncing ) { errmsg = "timeout waiting for sync() to finish"; return false; } return true; } } cmdResync; bool anyReplEnabled(){ return replPair || replSettings.slave || replSettings.master; } void appendReplicationInfo( BSONObjBuilder& result , bool authed , int level ){ if ( replAllDead ) { result.append("ismaster", 0.0); if( authed ) { if ( replPair ) result.append("remote", replPair->remote); result.append("info", replAllDead); } } else if ( replPair ) { result.append("ismaster", replPair->state); if( authed ) { result.append("remote", replPair->remote); if ( !replPair->info.empty() ) result.append("info", replPair->info); } } else { result.append("ismaster", replSettings.slave ? 0 : 1); result.append("msg", "not paired"); } if ( level ){ BSONObjBuilder sources( result.subarrayStart( "sources" ) ); readlock lk( "local.sources" ); Client::Context ctx( "local.sources" ); auto_ptr c = findTableScan("local.sources", BSONObj()); int n = 0; while ( c->ok() ){ BSONObj s = c->current(); BSONObjBuilder bb; bb.append( s["host"] ); string sourcename = s["source"].valuestr(); if ( sourcename != "main" ) bb.append( s["source"] ); { BSONElement e = s["syncedTo"]; BSONObjBuilder t( bb.subobjStart( "syncedTo" ) ); t.appendDate( "time" , e.timestampTime() ); t.append( "inc" , e.timestampInc() ); t.done(); } if ( level > 1 ){ dbtemprelease unlock; ScopedDbConnection conn( s["host"].valuestr() ); BSONObj first = conn->findOne( (string)"local.oplog.$" + sourcename , Query().sort( BSON( "$natural" << 1 ) ) ); BSONObj last = conn->findOne( (string)"local.oplog.$" + sourcename , Query().sort( BSON( "$natural" << -1 ) ) ); bb.appendDate( "masterFirst" , first["ts"].timestampTime() ); bb.appendDate( "masterLast" , last["ts"].timestampTime() ); double lag = (double) (last["ts"].timestampTime() - s["syncedTo"].timestampTime()); bb.append( "lagSeconds" , lag / 1000 ); conn.done(); } sources.append( BSONObjBuilder::numStr( n++ ) , bb.obj() ); c->advance(); } sources.done(); } } class CmdIsMaster : public Command { public: virtual bool requiresAuth() { return false; } virtual bool slaveOk() { return true; } virtual LockType locktype(){ return NONE; } CmdIsMaster() : Command("ismaster") { } virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool /*fromRepl*/) { /* currently request to arbiter is (somewhat arbitrarily) an ismaster request that is not authenticated. we allow unauthenticated ismaster but we aren't as verbose informationally if one is not authenticated for admin db to be safe. */ bool authed = cc().getAuthenticationInfo()->isAuthorizedReads("admin"); appendReplicationInfo( result , authed ); return true; } } cmdismaster; class CmdIsInitialSyncComplete : public Command { public: virtual bool requiresAuth() { return false; } virtual bool slaveOk() { return true; } virtual LockType locktype(){ return WRITE; } CmdIsInitialSyncComplete() : Command( "isinitialsynccomplete" ) {} virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool /*fromRepl*/) { result.appendBool( "initialsynccomplete", getInitialSyncCompleted() ); return true; } } cmdisinitialsynccomplete; /* negotiate who is master -1=not set (probably means we just booted) 0=was slave 1=was master remote,local -> new remote,local !1,1 -> 0,1 1,!1 -> 1,0 -1,-1 -> dominant->1, nondom->0 0,0 -> dominant->1, nondom->0 1,1 -> dominant->1, nondom->0 { negotiatemaster:1, i_was:, your_name: } returns: { ok:1, you_are:..., i_am:... } */ class CmdNegotiateMaster : public Command { public: CmdNegotiateMaster() : Command("negotiatemaster") { } virtual bool slaveOk() { return true; } virtual bool adminOnly() { return true; } virtual LockType locktype(){ return WRITE; } virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool) { if ( replPair == 0 ) { massert( 10383 , "Another mongod instance believes incorrectly that this node is its peer", !cmdObj.getBoolField( "fromArbiter" ) ); // assume that we are an arbiter and should forward the request string host = cmdObj.getStringField("your_name"); int port = cmdObj.getIntField( "your_port" ); if ( port == INT_MIN ) { errmsg = "no port specified"; problem() << errmsg << endl; return false; } stringstream ss; ss << host << ":" << port; string remote = ss.str(); BSONObj ret; { dbtemprelease t; auto_ptr conn( new DBClientConnection() ); if ( !conn->connect( remote.c_str(), errmsg ) ) { result.append( "you_are", ReplPair::State_Master ); return true; } BSONObjBuilder forwardCommand; forwardCommand.appendElements( cmdObj ); forwardCommand.appendBool( "fromArbiter", true ); ret = conn->findOne( "admin.$cmd", forwardCommand.done() ); } BSONObjIterator i( ret ); while( i.moreWithEOO() ) { BSONElement e = i.next(); if ( e.eoo() ) break; if ( e.fieldName() != string( "ok" ) ) result.append( e ); } return ( ret.getIntField("ok") == 1 ); } int was = cmdObj.getIntField("i_was"); string myname = cmdObj.getStringField("your_name"); if ( myname.empty() || was < -3 ) { errmsg = "your_name/i_was not specified"; return false; } int N = ReplPair::State_Negotiating; int M = ReplPair::State_Master; int S = ReplPair::State_Slave; if ( !replPair->dominant( myname ) ) { result.append( "you_are", N ); result.append( "i_am", replPair->state ); return true; } int me, you; if ( !getInitialSyncCompleted() || ( replPair->state != M && was == M ) ) { me=S; you=M; } else { me=M; you=S; } replPair->setMaster( me, "CmdNegotiateMaster::run()" ); result.append("you_are", you); result.append("i_am", me); return true; } } cmdnegotiatemaster; int ReplPair::negotiate(DBClientConnection *conn, string method) { BSONObjBuilder b; b.append("negotiatemaster",1); b.append("i_was", state); b.append("your_name", remoteHost); b.append("your_port", remotePort); BSONObj cmd = b.done(); BSONObj res = conn->findOne("admin.$cmd", cmd); if ( res.getIntField("ok") != 1 ) { string message = method + " negotiate failed"; problem() << message << ": " << res.toString() << '\n'; setMasterLocked(State_Confused, message.c_str()); return State_Confused; } int x = res.getIntField("you_are"); int remote = res.getIntField("i_am"); // State_Negotiating means the remote node is not dominant and cannot // choose who is master. if ( x != State_Slave && x != State_Master && x != State_Negotiating ) { problem() << method << " negotiate: bad you_are value " << res.toString() << endl; } else if ( x != State_Negotiating ) { string message = method + " negotiation"; setMasterLocked(x, message.c_str()); } return remote; } struct TestOpTime { TestOpTime() { OpTime t; for ( int i = 0; i < 10; i++ ) { OpTime s = OpTime::now(); assert( s != t ); t = s; } OpTime q = t; assert( q == t ); assert( !(q != t) ); } } testoptime; /* --------------------------------------------------------------*/ ReplSource::ReplSource() { replacing = false; nClonedThisPass = 0; paired = false; } ReplSource::ReplSource(BSONObj o) : nClonedThisPass(0) { replacing = false; paired = false; only = o.getStringField("only"); hostName = o.getStringField("host"); _sourceName = o.getStringField("source"); uassert( 10118 , "'host' field not set in sources collection object", !hostName.empty() ); uassert( 10119 , "only source='main' allowed for now with replication", sourceName() == "main" ); BSONElement e = o.getField("syncedTo"); if ( !e.eoo() ) { uassert( 10120 , "bad sources 'syncedTo' field value", e.type() == Date || e.type() == Timestamp ); OpTime tmp( e.date() ); syncedTo = tmp; } BSONObj dbsObj = o.getObjectField("dbsNextPass"); if ( !dbsObj.isEmpty() ) { BSONObjIterator i(dbsObj); while ( 1 ) { BSONElement e = i.next(); if ( e.eoo() ) break; addDbNextPass.insert( e.fieldName() ); } } dbsObj = o.getObjectField("incompleteCloneDbs"); if ( !dbsObj.isEmpty() ) { BSONObjIterator i(dbsObj); while ( 1 ) { BSONElement e = i.next(); if ( e.eoo() ) break; incompleteCloneDbs.insert( e.fieldName() ); } } _lastSavedLocalTs = OpTime( o.getField( "localLogTs" ).date() ); } /* Turn our C++ Source object into a BSONObj */ BSONObj ReplSource::jsobj() { BSONObjBuilder b; b.append("host", hostName); b.append("source", sourceName()); if ( !only.empty() ) b.append("only", only); if ( !syncedTo.isNull() ) b.appendTimestamp("syncedTo", syncedTo.asDate()); b.appendTimestamp("localLogTs", _lastSavedLocalTs.asDate()); BSONObjBuilder dbsNextPassBuilder; int n = 0; for ( set::iterator i = addDbNextPass.begin(); i != addDbNextPass.end(); i++ ) { n++; dbsNextPassBuilder.appendBool(i->c_str(), 1); } if ( n ) b.append("dbsNextPass", dbsNextPassBuilder.done()); BSONObjBuilder incompleteCloneDbsBuilder; n = 0; for ( set::iterator i = incompleteCloneDbs.begin(); i != incompleteCloneDbs.end(); i++ ) { n++; incompleteCloneDbsBuilder.appendBool(i->c_str(), 1); } if ( n ) b.append("incompleteCloneDbs", incompleteCloneDbsBuilder.done()); return b.obj(); } void ReplSource::save() { BSONObjBuilder b; assert( !hostName.empty() ); b.append("host", hostName); // todo: finish allowing multiple source configs. // this line doesn't work right when source is null, if that is allowed as it is now: //b.append("source", _sourceName); BSONObj pattern = b.done(); BSONObj o = jsobj(); log( 1 ) << "Saving repl source: " << o << endl; { OpDebug debug; Client::Context ctx("local.sources"); UpdateResult res = updateObjects("local.sources", o, pattern, true/*upsert for pair feature*/, false,false,debug); assert( ! res.mod ); assert( res.num == 1 ); } if ( replacing ) { /* if we were in "replace" mode, we now have synced up with the replacement, so turn that off. */ replacing = false; wassert( replacePeer ); replacePeer = false; Helpers::emptyCollection("local.pair.startup"); } } static void addSourceToList(ReplSource::SourceVector &v, ReplSource& s, const BSONObj &spec, ReplSource::SourceVector &old) { if ( !s.syncedTo.isNull() ) { // Don't reuse old ReplSource if there was a forced resync. for ( ReplSource::SourceVector::iterator i = old.begin(); i != old.end(); ) { if ( s == **i ) { v.push_back(*i); old.erase(i); return; } i++; } } v.push_back( shared_ptr< ReplSource >( new ReplSource( s ) ) ); } /* we reuse our existing objects so that we can keep our existing connection and cursor in effect. */ void ReplSource::loadAll(SourceVector &v) { Client::Context ctx("local.sources"); SourceVector old = v; v.clear(); bool gotPairWith = false; if ( !cmdLine.source.empty() ) { // --source specified. // check that no items are in sources other than that // add if missing auto_ptr c = findTableScan("local.sources", BSONObj()); int n = 0; while ( c->ok() ) { n++; ReplSource tmp(c->current()); if ( tmp.hostName != cmdLine.source ) { log() << "repl: --source " << cmdLine.source << " != " << tmp.hostName << " from local.sources collection" << endl; log() << "repl: terminating mongod after 30 seconds" << endl; sleepsecs(30); dbexit( EXIT_REPLICATION_ERROR ); } if ( tmp.only != cmdLine.only ) { log() << "--only " << cmdLine.only << " != " << tmp.only << " from local.sources collection" << endl; log() << "terminating after 30 seconds" << endl; sleepsecs(30); dbexit( EXIT_REPLICATION_ERROR ); } c->advance(); } uassert( 10002 , "local.sources collection corrupt?", n<2 ); if ( n == 0 ) { // source missing. add. ReplSource s; s.hostName = cmdLine.source; s.only = cmdLine.only; s.save(); } } else { try { massert( 10384 , "--only requires use of --source", cmdLine.only.empty()); } catch ( ... ) { dbexit( EXIT_BADOPTIONS ); } } if ( replPair ) { const string &remote = replPair->remote; // --pairwith host specified. if ( replSettings.fastsync ) { Helpers::emptyCollection( "local.sources" ); // ignore saved sources } // check that no items are in sources other than that // add if missing auto_ptr c = findTableScan("local.sources", BSONObj()); int n = 0; while ( c->ok() ) { n++; ReplSource tmp(c->current()); if ( tmp.hostName != remote ) { log() << "pairwith " << remote << " != " << tmp.hostName << " from local.sources collection" << endl; log() << "terminating after 30 seconds" << endl; sleepsecs(30); dbexit( EXIT_REPLICATION_ERROR ); } c->advance(); } uassert( 10122 , "local.sources collection corrupt?", n<2 ); if ( n == 0 ) { // source missing. add. ReplSource s; s.hostName = remote; s.save(); } } auto_ptr c = findTableScan("local.sources", BSONObj()); while ( c->ok() ) { ReplSource tmp(c->current()); if ( replPair && tmp.hostName == replPair->remote && tmp.sourceName() == "main" ) { gotPairWith = true; tmp.paired = true; if ( replacePeer ) { // peer was replaced -- start back at the beginning. tmp.syncedTo = OpTime(); tmp.replacing = true; } } if ( ( !replPair && tmp.syncedTo.isNull() ) || ( replPair && replSettings.fastsync ) ) { DBDirectClient c; if ( c.exists( "local.oplog.$main" ) ) { BSONObj op = c.findOne( "local.oplog.$main", Query().sort( BSON( "$natural" << -1 ) ) ); if ( !op.isEmpty() ) { tmp.syncedTo = op[ "ts" ].date(); tmp._lastSavedLocalTs = op[ "ts" ].date(); } } } addSourceToList(v, tmp, c->current(), old); c->advance(); } if ( !gotPairWith && replPair ) { /* add the --pairwith server */ shared_ptr< ReplSource > s( new ReplSource() ); s->paired = true; s->hostName = replPair->remote; s->replacing = replacePeer; v.push_back(s); } } BSONObj opTimeQuery = fromjson("{\"getoptime\":1}"); bool ReplSource::throttledForceResyncDead( const char *requester ) { if ( time( 0 ) - lastForcedResync > 600 ) { forceResyncDead( requester ); lastForcedResync = time( 0 ); return true; } return false; } void ReplSource::forceResyncDead( const char *requester ) { if ( !replAllDead ) return; SourceVector sources; ReplSource::loadAll(sources); for( SourceVector::iterator i = sources.begin(); i != sources.end(); ++i ) { (*i)->forceResync( requester ); } replAllDead = 0; } void ReplSource::forceResync( const char *requester ) { BSONObj info; { dbtemprelease t; connect(); bool ok = conn->runCommand( "admin", BSON( "listDatabases" << 1 ), info ); massert( 10385 , "Unable to get database list", ok ); } BSONObjIterator i( info.getField( "databases" ).embeddedObject() ); while( i.moreWithEOO() ) { BSONElement e = i.next(); if ( e.eoo() ) break; string name = e.embeddedObject().getField( "name" ).valuestr(); if ( !e.embeddedObject().getBoolField( "empty" ) ) { if ( name != "local" ) { if ( only.empty() || only == name ) { resyncDrop( name.c_str(), requester ); } } } } syncedTo = OpTime(); addDbNextPass.clear(); save(); } string ReplSource::resyncDrop( const char *db, const char *requester ) { log() << "resync: dropping database " << db << endl; string dummyns = string( db ) + "."; Client::Context ctx(dummyns); assert( cc().database()->name == db ); dropDatabase(dummyns.c_str()); return dummyns; } /* grab initial copy of a database from the master */ bool ReplSource::resync(string db) { string dummyNs = resyncDrop( db.c_str(), "internal" ); Client::Context ctx( dummyNs ); { log() << "resync: cloning database " << db << " to get an initial copy" << endl; ReplInfo r("resync: cloning a database"); string errmsg; bool ok = cloneFrom(hostName.c_str(), errmsg, cc().database()->name, false, /*slaveok*/ true, /*replauth*/ true, /*snapshot*/false); if ( !ok ) { problem() << "resync of " << db << " from " << hostName << " failed " << errmsg << endl; throw SyncException(); } } log() << "resync: done with initial clone for db: " << db << endl; return true; } void ReplSource::applyOperation(const BSONObj& op) { log( 6 ) << "applying op: " << op << endl; OpDebug debug; BSONObj o = op.getObjectField("o"); const char *ns = op.getStringField("ns"); // operation type -- see logOp() comments for types const char *opType = op.getStringField("op"); try { if ( *opType == 'i' ) { const char *p = strchr(ns, '.'); if ( p && strcmp(p, ".system.indexes") == 0 ) { // updates aren't allowed for indexes -- so we will do a regular insert. if index already // exists, that is ok. theDataFileMgr.insert(ns, (void*) o.objdata(), o.objsize()); } else { // do upserts for inserts as we might get replayed more than once BSONElement _id; if( !o.getObjectID(_id) ) { /* No _id. This will be very slow. */ Timer t; updateObjects(ns, o, o, true, false, false , debug ); if( t.millis() >= 2 ) { RARELY OCCASIONALLY log() << "warning, repl doing slow updates (no _id field) for " << ns << endl; } } else { BSONObjBuilder b; b.append(_id); /* erh 10/16/2009 - this is probably not relevant any more since its auto-created, but not worth removing */ RARELY ensureHaveIdIndex(ns); // otherwise updates will be slow updateObjects(ns, o, b.done(), true, false, false , debug ); } } } else if ( *opType == 'u' ) { RARELY ensureHaveIdIndex(ns); // otherwise updates will be super slow updateObjects(ns, o, op.getObjectField("o2"), op.getBoolField("b"), false, false , debug ); } else if ( *opType == 'd' ) { if ( opType[1] == 0 ) deleteObjects(ns, o, op.getBoolField("b")); else assert( opType[1] == 'b' ); // "db" advertisement } else if ( *opType == 'n' ) { // no op } else { BufBuilder bb; BSONObjBuilder ob; assert( *opType == 'c' ); _runCommands(ns, o, bb, ob, true, 0); } } catch ( UserException& e ) { log() << "sync: caught user assertion " << e << " while applying op: " << op << endl;; } catch ( DBException& e ) { log() << "sync: caught db exception " << e << " while applying op: " << op << endl;; } } /* local.$oplog.main is of the form: { ts: ..., op: , ns: ..., o: , o2: , b: } ... see logOp() comments. */ void ReplSource::sync_pullOpLog_applyOperation(BSONObj& op, OpTime *localLogTail) { log( 6 ) << "processing op: " << op << endl; // skip no-op if ( op.getStringField( "op" )[ 0 ] == 'n' ) return; char clientName[MaxDatabaseLen]; const char *ns = op.getStringField("ns"); nsToDatabase(ns, clientName); if ( *ns == '.' ) { problem() << "skipping bad op in oplog: " << op.toString() << endl; return; } else if ( *ns == 0 ) { problem() << "halting replication, bad op in oplog:\n " << op.toString() << endl; replAllDead = "bad object in oplog"; throw SyncException(); } if ( !only.empty() && only != clientName ) return; dblock lk; if ( localLogTail && replPair && replPair->state == ReplPair::State_Master ) { updateSetsWithLocalOps( *localLogTail, true ); // allow unlocking updateSetsWithLocalOps( *localLogTail, false ); // don't allow unlocking or conversion to db backed storage } if ( replAllDead ) { // hmmm why is this check here and not at top of this function? does it get set between top and here? log() << "replAllDead, throwing SyncException: " << replAllDead << endl; throw SyncException(); } Client::Context ctx( ns ); bool empty = ctx.db()->isEmpty(); bool incompleteClone = incompleteCloneDbs.count( clientName ) != 0; log( 6 ) << "ns: " << ns << ", justCreated: " << ctx.justCreated() << ", empty: " << empty << ", incompleteClone: " << incompleteClone << endl; // always apply admin command command // this is a bit hacky -- the semantics of replication/commands aren't well specified if ( strcmp( clientName, "admin" ) == 0 && *op.getStringField( "op" ) == 'c' ) { applyOperation( op ); return; } if ( ctx.justCreated() || empty || incompleteClone ) { // we must add to incomplete list now that setClient has been called incompleteCloneDbs.insert( clientName ); if ( nClonedThisPass ) { /* we only clone one database per pass, even if a lot need done. This helps us avoid overflowing the master's transaction log by doing too much work before going back to read more transactions. (Imagine a scenario of slave startup where we try to clone 100 databases in one pass.) */ addDbNextPass.insert( clientName ); } else { if ( incompleteClone ) { log() << "An earlier initial clone of '" << clientName << "' did not complete, now resyncing." << endl; } save(); Client::Context ctx(ns); nClonedThisPass++; resync(ctx.db()->name); addDbNextPass.erase(clientName); incompleteCloneDbs.erase( clientName ); } save(); } else { bool mod; if ( replPair && replPair->state == ReplPair::State_Master ) { BSONObj id = idForOp( op, mod ); if ( !idTracker.haveId( ns, id ) ) { applyOperation( op ); } else if ( idTracker.haveModId( ns, id ) ) { log( 6 ) << "skipping operation matching mod id object " << op << endl; BSONObj existing; if ( Helpers::findOne( ns, id, existing ) ) logOp( "i", ns, existing ); } else { log( 6 ) << "skipping operation matching changed id object " << op << endl; } } else { applyOperation( op ); } addDbNextPass.erase( clientName ); } } BSONObj ReplSource::idForOp( const BSONObj &op, bool &mod ) { mod = false; const char *opType = op.getStringField( "op" ); BSONObj o = op.getObjectField( "o" ); switch( opType[ 0 ] ) { case 'i': { BSONObjBuilder idBuilder; BSONElement id; if ( !o.getObjectID( id ) ) return BSONObj(); idBuilder.append( id ); return idBuilder.obj(); } case 'u': { BSONObj o2 = op.getObjectField( "o2" ); if ( strcmp( o2.firstElement().fieldName(), "_id" ) != 0 ) return BSONObj(); if ( o.firstElement().fieldName()[ 0 ] == '$' ) mod = true; return o2; } case 'd': { if ( opType[ 1 ] != '\0' ) return BSONObj(); // skip "db" op type return o; } default: break; } return BSONObj(); } void ReplSource::updateSetsWithOp( const BSONObj &op, bool mayUnlock ) { if ( mayUnlock ) { idTracker.mayUpgradeStorage(); } bool mod; BSONObj id = idForOp( op, mod ); if ( !id.isEmpty() ) { const char *ns = op.getStringField( "ns" ); // Since our range of local ops may not be the same as our peer's // range of unapplied ops, it is always necessary to rewrite objects // to the oplog after a mod update. if ( mod ) idTracker.haveModId( ns, id, true ); idTracker.haveId( ns, id, true ); } } void ReplSource::syncToTailOfRemoteLog() { string _ns = ns(); BSONObjBuilder b; if ( !only.empty() ) { b.appendRegex("ns", string("^") + only); } BSONObj last = conn->findOne( _ns.c_str(), Query( b.done() ).sort( BSON( "$natural" << -1 ) ) ); if ( !last.isEmpty() ) { BSONElement ts = last.getField( "ts" ); massert( 10386 , "non Date ts found", ts.type() == Date || ts.type() == Timestamp ); syncedTo = OpTime( ts.date() ); } } OpTime ReplSource::nextLastSavedLocalTs() const { Client::Context ctx( "local.oplog.$main" ); auto_ptr< Cursor > c = findTableScan( "local.oplog.$main", BSON( "$natural" << -1 ) ); if ( c->ok() ) return OpTime( c->current().getField( "ts" ).date() ); return OpTime(); } void ReplSource::setLastSavedLocalTs( const OpTime &nextLocalTs ) { _lastSavedLocalTs = nextLocalTs; log( 3 ) << "updated _lastSavedLocalTs to: " << _lastSavedLocalTs << endl; } void ReplSource::resetSlave() { massert( 10387 , "request to kill slave replication falied", conn->simpleCommand( "admin", 0, "forcedead" ) ); syncToTailOfRemoteLog(); { dblock lk; setLastSavedLocalTs( nextLastSavedLocalTs() ); save(); cursor.reset(); } } bool ReplSource::updateSetsWithLocalOps( OpTime &localLogTail, bool mayUnlock ) { Client::Context ctx( "local.oplog.$main" ); auto_ptr< Cursor > localLog = findTableScan( "local.oplog.$main", BSON( "$natural" << -1 ) ); OpTime newTail; for( ; localLog->ok(); localLog->advance() ) { BSONObj op = localLog->current(); OpTime ts( localLog->current().getField( "ts" ).date() ); if ( newTail.isNull() ) { newTail = ts; } if ( !( localLogTail < ts ) ) break; updateSetsWithOp( op, mayUnlock ); if ( mayUnlock ) { RARELY { dbtemprelease t; } } } if ( !localLogTail.isNull() && !localLog->ok() ) { // local log filled up idTracker.reset(); dbtemprelease t; resetSlave(); massert( 10388 , "local master log filled, forcing slave resync", false ); } if ( !newTail.isNull() ) localLogTail = newTail; return true; } /* slave: pull some data from the master's oplog note: not yet in db mutex at this point. */ bool ReplSource::sync_pullOpLog(int& nApplied) { string ns = string("local.oplog.$") + sourceName(); log(2) << "repl: sync_pullOpLog " << ns << " syncedTo:" << syncedTo.toStringLong() << '\n'; bool tailing = true; DBClientCursor *c = cursor.get(); if ( c && c->isDead() ) { log() << "repl: old cursor isDead, initiating a new one\n"; c = 0; } if ( replPair && replPair->state == ReplPair::State_Master ) { dblock lk; idTracker.reset(); } OpTime localLogTail = _lastSavedLocalTs; bool initial = syncedTo.isNull(); if ( c == 0 || initial ) { if ( initial ) { // Important to grab last oplog timestamp before listing databases. syncToTailOfRemoteLog(); BSONObj info; bool ok = conn->runCommand( "admin", BSON( "listDatabases" << 1 ), info ); massert( 10389 , "Unable to get database list", ok ); BSONObjIterator i( info.getField( "databases" ).embeddedObject() ); while( i.moreWithEOO() ) { BSONElement e = i.next(); if ( e.eoo() ) break; string name = e.embeddedObject().getField( "name" ).valuestr(); if ( !e.embeddedObject().getBoolField( "empty" ) ) { if ( name != "local" ) { if ( only.empty() || only == name ) { log( 2 ) << "adding to 'addDbNextPass': " << name << endl; addDbNextPass.insert( name ); } } } } dblock lk; save(); } BSONObjBuilder q; q.appendDate("$gte", syncedTo.asDate()); BSONObjBuilder query; query.append("ts", q.done()); if ( !only.empty() ) { // note we may here skip a LOT of data table scanning, a lot of work for the master. query.appendRegex("ns", string("^") + only); } BSONObj queryObj = query.done(); // queryObj = { ts: { $gte: syncedTo } } log(2) << "repl: " << ns << ".find(" << queryObj.toString() << ')' << '\n'; cursor = conn->query( ns.c_str(), queryObj, 0, 0, 0, QueryOption_CursorTailable | QueryOption_SlaveOk | QueryOption_OplogReplay | QueryOption_AwaitData ); c = cursor.get(); tailing = false; } else { log(2) << "repl: tailing=true\n"; } if ( c == 0 ) { problem() << "repl: dbclient::query returns null (conn closed?)" << endl; resetConnection(); return false; } // show any deferred database creates from a previous pass { set::iterator i = addDbNextPass.begin(); if ( i != addDbNextPass.end() ) { BSONObjBuilder b; b.append("ns", *i + '.'); b.append("op", "db"); BSONObj op = b.done(); sync_pullOpLog_applyOperation(op, 0); } } if ( !c->more() ) { if ( tailing ) { log(2) << "repl: tailing & no new activity\n"; } else { log() << "repl: " << ns << " oplog is empty\n"; } { dblock lk; OpTime nextLastSaved = nextLastSavedLocalTs(); { dbtemprelease t; if ( !c->more() ) { setLastSavedLocalTs( nextLastSaved ); } } save(); } return true; } OpTime nextOpTime; { BSONObj op = c->next(); BSONElement ts = op.getField("ts"); if ( ts.type() != Date && ts.type() != Timestamp ) { string err = op.getStringField("$err"); if ( !err.empty() ) { problem() << "repl: $err reading remote oplog: " + err << '\n'; massert( 10390 , "got $err reading remote oplog", false ); } else { problem() << "repl: bad object read from remote oplog: " << op.toString() << '\n'; massert( 10391 , "repl: bad object read from remote oplog", false); } } if ( replPair && replPair->state == ReplPair::State_Master ) { OpTime nextOpTime( ts.date() ); if ( !tailing && !initial && nextOpTime != syncedTo ) { log() << "remote slave log filled, forcing slave resync" << endl; resetSlave(); return true; } dblock lk; updateSetsWithLocalOps( localLogTail, true ); } nextOpTime = OpTime( ts.date() ); log(2) << "repl: first op time received: " << nextOpTime.toString() << '\n'; if ( tailing || initial ) { if ( initial ) log(1) << "repl: initial run\n"; else assert( syncedTo < nextOpTime ); c->putBack( op ); // op will be processed in the loop below nextOpTime = OpTime(); // will reread the op below } else if ( nextOpTime != syncedTo ) { // didn't get what we queried for - error Nullstream& l = log(); l << "repl: nextOpTime " << nextOpTime.toStringLong() << ' '; if ( nextOpTime < syncedTo ) l << ""; l << " syncedTo " << syncedTo.toStringLong() << '\n'; log() << "repl: time diff: " << (nextOpTime.getSecs() - syncedTo.getSecs()) << "sec\n"; log() << "repl: tailing: " << tailing << '\n'; log() << "repl: data too stale, halting replication" << endl; replInfo = replAllDead = "data too stale halted replication"; assert( syncedTo < nextOpTime ); throw SyncException(); } else { /* t == syncedTo, so the first op was applied previously. */ } } // apply operations { int n = 0; time_t saveLast = time(0); while ( 1 ) { /* from a.s.: I think the idea here is that we can establish a sync point between the local op log and the remote log with the following steps: 1) identify most recent op in local log -- call it O 2) ask "does nextOpTime reflect the tail of the remote op log?" (in other words, is more() false?) - If yes, all subsequent ops after nextOpTime in the remote log must have occurred after O. If no, we can't establish a sync point. Note that we can't do step (2) followed by step (1) because if we do so ops may be added to both machines between steps (2) and (1) and we can't establish a sync point. (In particular, between (2) and (1) an op may be added to the remote log before a different op is added to the local log. In this case, the newest remote op will have occurred after nextOpTime but before O.) Now, for performance reasons we don't want to have to identify the most recent op in the local log every time we call c->more() because in performance sensitive situations more() will be true most of the time. So we do: 0) more()? 1) find most recent op in local log 2) more()? */ if ( !c->more() ) { dblock lk; OpTime nextLastSaved = nextLastSavedLocalTs(); { dbtemprelease t; if ( c->more() ) { continue; } else { setLastSavedLocalTs( nextLastSaved ); } } syncedTo = nextOpTime; save(); // note how far we are synced up to now log() << "repl: applied " << n << " operations" << endl; nApplied = n; log() << "repl: end sync_pullOpLog syncedTo: " << syncedTo.toStringLong() << endl; break; } OCCASIONALLY if( n > 0 && ( n > 100000 || time(0) - saveLast > 60 ) ) { // periodically note our progress, in case we are doing a lot of work and crash dblock lk; syncedTo = nextOpTime; // can't update local log ts since there are pending operations from our peer save(); log() << "repl: checkpoint applied " << n << " operations" << endl; log() << "repl: syncedTo: " << syncedTo.toStringLong() << endl; saveLast = time(0); n = 0; } BSONObj op = c->next(); BSONElement ts = op.getField("ts"); assert( ts.type() == Date || ts.type() == Timestamp ); OpTime last = nextOpTime; OpTime tmp( ts.date() ); nextOpTime = tmp; if ( !( last < nextOpTime ) ) { problem() << "sync error: last " << last.toString() << " >= nextOpTime " << nextOpTime.toString() << endl; uassert( 10123 , "bad 'ts' value in sources", false); } if ( replSettings.slavedelay && ( unsigned( time( 0 ) ) < nextOpTime.getSecs() + replSettings.slavedelay ) ) { c->putBack( op ); _sleepAdviceTime = nextOpTime.getSecs() + replSettings.slavedelay + 1; dblock lk; if ( n > 0 ) { syncedTo = last; save(); } log() << "repl: applied " << n << " operations" << endl; log() << "repl: syncedTo: " << syncedTo.toStringLong() << endl; log() << "waiting until: " << _sleepAdviceTime << " to continue" << endl; break; } sync_pullOpLog_applyOperation(op, &localLogTail); n++; } } return true; } BSONObj userReplQuery = fromjson("{\"user\":\"repl\"}"); bool replAuthenticate(DBClientConnection *conn) { if( ! cc().isAdmin() ){ log() << "replauthenticate: requires admin permissions, failing\n"; return false; } BSONObj user; { dblock lk; Client::Context ctxt("local."); if( !Helpers::findOne("local.system.users", userReplQuery, user) ) { // try the first user is local if( !Helpers::getSingleton("local.system.users", user) ) { if( noauth ) return true; // presumably we are running a --noauth setup all around. log() << "replauthenticate: no user in local.system.users to use for authentication\n"; return false; } } } string u = user.getStringField("user"); string p = user.getStringField("pwd"); massert( 10392 , "bad user object? [1]", !u.empty()); massert( 10393 , "bad user object? [2]", !p.empty()); string err; if( !conn->auth("local", u.c_str(), p.c_str(), err, false) ) { log() << "replauthenticate: can't authenticate to master server, user:" << u << endl; return false; } return true; } bool ReplSource::connect() { if ( conn.get() == 0 ) { conn = auto_ptr(new DBClientConnection()); string errmsg; ReplInfo r("trying to connect to sync source"); if ( !conn->connect(hostName.c_str(), errmsg) || !replAuthenticate(conn.get()) ) { resetConnection(); log() << "repl: " << errmsg << endl; return false; } } return true; } /* note: not yet in mutex at this point. returns true if everything happy. return false if you want to reconnect. */ bool ReplSource::sync(int& nApplied) { _sleepAdviceTime = 0; ReplInfo r("sync"); if ( !cmdLine.quiet ) log() << "repl: " << sourceName() << '@' << hostName << endl; nClonedThisPass = 0; // FIXME Handle cases where this db isn't on default port, or default port is spec'd in hostName. if ( (string("localhost") == hostName || string("127.0.0.1") == hostName) && cmdLine.port == CmdLine::DefaultDBPort ) { log() << "repl: can't sync from self (localhost). sources configuration may be wrong." << endl; sleepsecs(5); return false; } if ( !connect() ) { log(4) << "repl: can't connect to sync source" << endl; if ( replPair && paired ) { assert( startsWith(hostName.c_str(), replPair->remoteHost.c_str()) ); replPair->arbitrate(); } return false; } if ( paired ) { int remote = replPair->negotiate(conn.get(), "direct"); int nMasters = ( remote == ReplPair::State_Master ) + ( replPair->state == ReplPair::State_Master ); if ( getInitialSyncCompleted() && nMasters != 1 ) { log() << ( nMasters == 0 ? "no master" : "two masters" ) << ", deferring oplog pull" << endl; return true; } } /* // get current mtime at the server. BSONObj o = conn->findOne("admin.$cmd", opTimeQuery); BSONElement e = o.getField("optime"); if( e.eoo() ) { log() << "repl: failed to get cur optime from master" << endl; log() << " " << o.toString() << endl; return false; } uassert( 10124 , e.type() == Date ); OpTime serverCurTime; serverCurTime.asDate() = e.date(); */ return sync_pullOpLog(nApplied); } /* -- Logging of operations -------------------------------------*/ // cached copies of these...so don't rename them NamespaceDetails *localOplogMainDetails = 0; Database *localOplogDB = 0; void logOp(const char *opstr, const char *ns, const BSONObj& obj, BSONObj *patt, bool *b) { if ( replSettings.master ) { _logOp(opstr, ns, "local.oplog.$main", obj, patt, b, OpTime::now()); char cl[ 256 ]; nsToDatabase( ns, cl ); } NamespaceDetailsTransient &t = NamespaceDetailsTransient::get_w( ns ); if ( t.cllEnabled() ) { try { _logOp(opstr, ns, t.cllNS().c_str(), obj, patt, b, OpTime::now()); } catch ( const DBException & ) { t.cllInvalidate(); } } } /* we write to local.opload.$main: { ts : ..., op: ..., ns: ..., o: ... } ts: an OpTime timestamp op: "i" insert "u" update "d" delete "c" db cmd "db" declares presence of a database (ns is set to the db name + '.') "n" no op bb: if not null, specifies a boolean to pass along to the other side as b: param. used for "justOne" or "upsert" flags on 'd', 'u' first: true when set, indicates this is the first thing we have logged for this database. thus, the slave does not need to copy down all the data when it sees this. */ void _logOp(const char *opstr, const char *ns, const char *logNS, const BSONObj& obj, BSONObj *o2, bool *bb, const OpTime &ts ) { if ( strncmp(ns, "local.", 6) == 0 ) return; DEV assertInWriteLock(); Client::Context context; /* we jump through a bunch of hoops here to avoid copying the obj buffer twice -- instead we do a single copy to the destination position in the memory mapped file. */ BSONObjBuilder b; b.appendTimestamp("ts", ts.asDate()); b.append("op", opstr); b.append("ns", ns); if ( bb ) b.appendBool("b", *bb); if ( o2 ) b.append("o2", *o2); BSONObj partial = b.done(); int posz = partial.objsize(); int len = posz + obj.objsize() + 1 + 2 /*o:*/; Record *r; if ( strncmp( logNS, "local.", 6 ) == 0 ) { // For now, assume this is olog main if ( localOplogMainDetails == 0 ) { Client::Context ctx("local."); localOplogDB = ctx.db(); localOplogMainDetails = nsdetails(logNS); } Client::Context ctx( "" , localOplogDB ); r = theDataFileMgr.fast_oplog_insert(localOplogMainDetails, logNS, len); } else { Client::Context ctx( logNS ); assert( nsdetails( logNS ) ); r = theDataFileMgr.fast_oplog_insert( nsdetails( logNS ), logNS, len); } char *p = r->data; memcpy(p, partial.objdata(), posz); *((unsigned *)p) += obj.objsize() + 1 + 2; p += posz - 1; *p++ = (char) Object; *p++ = 'o'; *p++ = 0; memcpy(p, obj.objdata(), obj.objsize()); p += obj.objsize(); *p = EOO; if ( logLevel >= 6 ) { BSONObj temp(r); log( 6 ) << "logging op:" << temp << endl; } } /* --------------------------------------------------------------*/ /* TODO: _ source has autoptr to the cursor _ reuse that cursor when we can */ /* returns: # of seconds to sleep before next pass 0 = no sleep recommended 1 = special sentinel indicating adaptive sleep recommended */ int _replMain(ReplSource::SourceVector& sources, int& nApplied) { { ReplInfo r("replMain load sources"); dblock lk; ReplSource::loadAll(sources); } if ( sources.empty() ) { /* replication is not configured yet (for --slave) in local.sources. Poll for config it every 20 seconds. */ return 20; } int sleepAdvice = 1; for ( ReplSource::SourceVector::iterator i = sources.begin(); i != sources.end(); i++ ) { ReplSource *s = i->get(); bool ok = false; try { ok = s->sync(nApplied); bool moreToSync = s->haveMoreDbsToSync(); if( !ok ) { sleepAdvice = 3; } else if( moreToSync ) { sleepAdvice = 0; } else if ( s->sleepAdvice() ) { sleepAdvice = s->sleepAdvice(); } if ( ok && !moreToSync /*&& !s->syncedTo.isNull()*/ ) { pairSync->setInitialSyncCompletedLocking(); } } catch ( const SyncException& ) { log() << "caught SyncException" << endl; return 10; } catch ( AssertionException& e ) { if ( e.severe() ) { log() << "replMain AssertionException " << e.what() << endl; return 60; } else { log() << "repl: AssertionException " << e.what() << '\n'; } replInfo = "replMain caught AssertionException"; } catch ( const DBException& e ) { log() << "repl: DBException " << e.what() << endl; replInfo = "replMain caught DBException"; } catch ( const std::exception &e ) { log() << "repl: std::exception " << e.what() << endl; replInfo = "replMain caught std::exception"; } catch ( ... ) { log() << "unexpected exception during replication. replication will halt" << endl; replAllDead = "caught unexpected exception during replication"; } if ( !ok ) s->resetConnection(); } return sleepAdvice; } void replMain() { ReplSource::SourceVector sources; while ( 1 ) { int s = 0; { dblock lk; if ( replAllDead ) { if ( !replSettings.autoresync || !ReplSource::throttledForceResyncDead( "auto" ) ) break; } assert( syncing == 0 ); // i.e., there is only one sync thread running. we will want to change/fix this. syncing++; } try { int nApplied = 0; s = _replMain(sources, nApplied); if( s == 1 ) { if( nApplied == 0 ) s = 2; else if( nApplied > 100 ) { // sleep very little - just enought that we aren't truly hammering master sleepmillis(75); s = 0; } } } catch (...) { out() << "caught exception in _replMain" << endl; s = 4; } { dblock lk; assert( syncing == 1 ); syncing--; } if ( s ) { stringstream ss; ss << "repl: sleep " << s << "sec before next pass"; string msg = ss.str(); if ( ! cmdLine.quiet ) log() << msg << endl; ReplInfo r(msg.c_str()); sleepsecs(s); } } } int debug_stop_repl = 0; void replSlaveThread() { sleepsecs(1); Client::initThread("replslave"); { dblock lk; cc().getAuthenticationInfo()->authorize("admin"); BSONObj obj; if ( Helpers::getSingleton("local.pair.startup", obj) ) { // should be: {replacepeer:1} replacePeer = true; pairSync->setInitialSyncCompleted(); // we are the half that has all the data } } while ( 1 ) { try { replMain(); if ( debug_stop_repl ) break; sleepsecs(5); } catch ( AssertionException& ) { ReplInfo r("Assertion in replSlaveThread(): sleeping 5 minutes before retry"); problem() << "Assertion in replSlaveThread(): sleeping 5 minutes before retry" << endl; sleepsecs(300); } } } void tempThread() { while ( 1 ) { out() << dbMutex.info().isLocked() << endl; sleepmillis(100); } } void createOplog() { dblock lk; const char * ns = "local.oplog.$main"; Client::Context ctx(ns); if ( nsdetails( ns ) ) return; /* create an oplog collection, if it doesn't yet exist. */ BSONObjBuilder b; double sz; if ( cmdLine.oplogSize != 0 ) sz = (double)cmdLine.oplogSize; else { sz = 50.0 * 1000 * 1000; if ( sizeof(int *) >= 8 ) { sz = 990.0 * 1000 * 1000; boost::intmax_t free = freeSpace(); //-1 if call not supported. double fivePct = free * 0.05; if ( fivePct > sz ) sz = fivePct; } } log() << "******\n"; log() << "creating replication oplog of size: " << (int)( sz / ( 1024 * 1024 ) ) << "MB (use --oplogSize to change)\n"; log() << "******" << endl; b.append("size", sz); b.appendBool("capped", 1); b.appendBool("autoIndexId", false); string err; BSONObj o = b.done(); userCreateNS(ns, o, err, false); logOp( "n", "dummy", BSONObj() ); } void startReplication() { /* this was just to see if anything locks for longer than it should -- we need to be careful not to be locked when trying to connect() or query() the other side. */ //boost::thread tempt(tempThread); if ( !replSettings.slave && !replSettings.master && !replPair ) return; { dblock lk; pairSync->init(); } if ( replSettings.slave || replPair ) { if ( replSettings.slave ) { assert( replSettings.slave == SimpleSlave ); log(1) << "slave=true" << endl; } else replSettings.slave = ReplPairSlave; boost::thread repl_thread(replSlaveThread); } if ( replSettings.master || replPair ) { if ( replSettings.master ) log(1) << "master=true" << endl; replSettings.master = true; createOplog(); } } /* called from main at server startup */ void pairWith(const char *remoteEnd, const char *arb) { replPair = new ReplPair(remoteEnd, arb); } class CmdLogCollection : public Command { public: virtual bool slaveOk() { return false; } virtual LockType locktype(){ return WRITE; } CmdLogCollection() : Command( "logCollection" ) {} virtual void help( stringstream &help ) const { help << "examples: { logCollection: , start: 1 }, " << "{ logCollection: , validateComplete: 1 }"; } virtual bool run(const char *ns, BSONObj& cmdObj, string& errmsg, BSONObjBuilder& result, bool fromRepl) { string logCollection = cmdObj.getStringField( "logCollection" ); if ( logCollection.empty() ) { errmsg = "missing logCollection spec"; return false; } bool start = !cmdObj.getField( "start" ).eoo(); bool validateComplete = !cmdObj.getField( "validateComplete" ).eoo(); if ( start ? validateComplete : !validateComplete ) { errmsg = "Must specify exactly one of start:1 or validateComplete:1"; return false; } int logSizeMb = cmdObj.getIntField( "logSizeMb" ); NamespaceDetailsTransient &t = NamespaceDetailsTransient::get_w( logCollection.c_str() ); if ( start ) { if ( t.cllNS().empty() ) { if ( logSizeMb == INT_MIN ) { t.cllStart(); } else { t.cllStart( logSizeMb ); } } else { errmsg = "Log already started for ns: " + logCollection; return false; } } else { if ( t.cllNS().empty() ) { errmsg = "No log to validateComplete for ns: " + logCollection; return false; } else { if ( !t.cllValidateComplete() ) { errmsg = "Oplog failure, insufficient space allocated"; return false; } } } log() << "started logCollection with cmd obj: " << cmdObj << endl; return true; } } cmdlogcollection; } // namespace mongo