/** * Copyright (C) 2018-present MongoDB, Inc. * * This program is free software: you can redistribute it and/or modify * it under the terms of the Server Side Public License, version 1, * as published by MongoDB, Inc. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * Server Side Public License for more details. * * You should have received a copy of the Server Side Public License * along with this program. If not, see * . * * As a special exception, the copyright holders give permission to link the * code of portions of this program with the OpenSSL library under certain * conditions as described in each individual source file and distribute * linked combinations including the program with the OpenSSL library. You * must comply with the Server Side Public License in all respects for * all of the code used other than as permitted herein. If you modify file(s) * with this exception, you may extend this exception to your version of the * file(s), but you are not obligated to do so. If you do not wish to do so, * delete this exception statement from your version. If you delete this * exception statement from all source files in the program, then also delete * it in the license file. */ #pragma once #include "mongo/base/status_with.h" #include "mongo/db/repl/rollback.h" #include "mongo/db/repl/storage_interface.h" #include "mongo/stdx/functional.h" namespace mongo { class OperationContext; namespace repl { class OplogInterface; class ReplicationCoordinator; class ReplicationProcess; /** * During steady state replication, it is possible to find the local server in a state where it * cannot replicate from a sync source. This can happen if the local server has gone offline and * comes back to find a new primary with an inconsistent set of operations in its oplog from the * local server. For example: * F = node that is on the wrong branch of history * SS = sync source (usually primary) * * F : a b c d e f g * SS : a b c d h * * In the example 'e', 'f', and 'g' are getting rolled back, 'h' is what's getting rolled forward. * * This class performs 'rollback' via the storage engine's 'recover to a stable timestamp' * machinery. This class runs synchronously on the caller's thread. * * Order of actions: * 1. Transition to ROLLBACK * 2. Find the common point between the local and remote oplogs. * a. Keep track of what is rolled back to provide a summary to the user * b. Write rolled back documents to 'Rollback Files' * 3. Increment the Rollback ID (RBID) * 4. Write the common point as the 'OplogTruncateAfterPoint' * 5. Tell the storage engine to recover to the last stable timestamp * 6. Call recovery code * a. Truncate the oplog at the common point * b. Apply all oplog entries to the end of oplog. * 7. Check the shard identity document for roll back * 8. Clear the in-memory transaction table * 9. Transition to SECONDARY * * If the node crashes while in rollback and the storage engine has not recovered to the last * stable timestamp yet, then rollback will simply restart against the new sync source upon restart. * If the node crashes after the storage engine has recovered to the last stable timestamp, * then the normal startup recovery code will run and finish the rollback process. * * If the sync source rolls back while we're searching for a common point, the connection should * get closed and finding the common point should fail. * */ class RollbackImpl : public Rollback { public: /** * A class with functions that get called throughout rollback. These can be overridden to * instrument this class for diagnostics and testing. */ class Listener { public: virtual ~Listener() = default; /** * Function called after we transition to ROLLBACK. */ virtual void onTransitionToRollback() noexcept {} /** * Function called after we find the common point. */ virtual void onCommonPointFound(Timestamp commonPoint) noexcept {} /** * Function called after we recover to the stable timestamp. */ virtual void onRecoverToStableTimestamp() noexcept {} /** * Function called after we recover from the oplog. */ virtual void onRecoverFromOplog() noexcept {} }; /** * Creates a RollbackImpl instance that will run the entire rollback algorithm. This is * called during steady state replication when we determine that we have to roll back after * processing the first batch of oplog entries from the sync source. */ RollbackImpl(OplogInterface* localOplog, OplogInterface* remoteOplog, StorageInterface* storageInterface, ReplicationProcess* replicationProcess, ReplicationCoordinator* replicationCoordinator, Listener* listener); /** * Constructs RollbackImpl with a default noop listener. */ RollbackImpl(OplogInterface* localOplog, OplogInterface* remoteOplog, StorageInterface* storageInterface, ReplicationProcess* replicationProcess, ReplicationCoordinator* replicationCoordinator); virtual ~RollbackImpl(); /** * Runs the rollback algorithm. */ Status runRollback(OperationContext* opCtx); /** * Cancels all outstanding work. */ void shutdown(); private: /** * Returns if shutdown was called on this rollback process. */ bool _isInShutdown() const; /** * Finds the common point between the local and remote oplogs. */ StatusWith _findCommonPoint(); /** * Uses the ReplicationCoordinator to transition the current member state to ROLLBACK. * If the transition to ROLLBACK fails, this could mean that we have been elected PRIMARY. In * this case, we return a NotSecondary error. * * 'opCtx' cannot be null. */ Status _transitionToRollback(OperationContext* opCtx); /** * Recovers to the stable timestamp while holding the global exclusive lock. */ Status _recoverToStableTimestamp(OperationContext* opCtx); /** * Runs the oplog recovery logic. This involves applying oplog operations between the stable * timestamp and the common point. */ Status _oplogRecovery(OperationContext* opCtx); /** * If we detected that we rolled back the shardIdentity document as part of this rollback * then we must shut down the server to clear the in-memory ShardingState associated with the * shardIdentity document. * * 'opCtx' cannot be null. */ void _checkShardIdentityRollback(OperationContext* opCtx); /** * In-memory sessions need to be reset after rollback, so they are forced to refetch from the * transactions collection. * * 'opCtx' cannot be null. */ void _resetSessions(OperationContext* opCtx); /** * Transitions the current member state from ROLLBACK to SECONDARY. * This operation must succeed. Otherwise, we will shut down the server. * * 'opCtx' cannot be null. */ void _transitionFromRollbackToSecondary(OperationContext* opCtx); // All member variables are labeled with one of the following codes indicating the // synchronization rules for accessing them. // // (R) Read-only in concurrent operation; no synchronization required. // (S) Self-synchronizing; access in any way from any context. // (M) Reads and writes guarded by _mutex. // Guards access to member variables. mutable stdx::mutex _mutex; // (S) // Set to true when RollbackImpl should shut down. bool _inShutdown = false; // (M) // This is used to read oplog entries from the local oplog that will be rolled back. OplogInterface* const _localOplog; // (R) // This is used to read oplog entries from the remote oplog to find the common point. OplogInterface* const _remoteOplog; // (R) // The StorageInterface associated with this Rollback instance. Used to execute operations // at the storage layer e.g. recovering to a timestamp. StorageInterface* _storageInterface; // (R) // The ReplicationProcess associated with this Rollback instance. Used to update and persist // various pieces of replication state related to the rollback process. ReplicationProcess* _replicationProcess; // (R) // This is used to read and update global replication settings. This includes: // - update transition member states; ReplicationCoordinator* const _replicationCoordinator; // (R) // A listener that's called at various points throughout rollback. Listener* _listener; // (R) }; } // namespace repl } // namespace mongo