diff options
author | Reo Kimura <reo.kimura@mongodb.com> | 2020-07-22 03:03:43 +0000 |
---|---|---|
committer | Reo Kimura <reo.kimura@mongodb.com> | 2020-08-07 18:51:05 +0000 |
commit | 120b213146dc93d8777371650d082dd15b1a1c0b (patch) | |
tree | e9c961f52fc308429d188bbb5e61dffbd01d1e21 /src/mongo | |
parent | 9238911d0a46f26419ecdbec4293457b9e1a891d (diff) | |
download | mongo-server-49072.tar.gz |
SERVER-49072 Created SE decorations, rewrote SEPI start/shutdown of SEsserver-49072
Diffstat (limited to 'src/mongo')
-rw-r--r-- | src/mongo/db/commands/server_status_servers.cpp | 4 | ||||
-rw-r--r-- | src/mongo/db/mongod_main.cpp | 23 | ||||
-rw-r--r-- | src/mongo/db/service_context.cpp | 8 | ||||
-rw-r--r-- | src/mongo/db/service_context.h | 13 | ||||
-rw-r--r-- | src/mongo/s/mongos_main.cpp | 19 | ||||
-rw-r--r-- | src/mongo/tools/bridge.cpp | 4 | ||||
-rw-r--r-- | src/mongo/transport/SConscript | 3 | ||||
-rw-r--r-- | src/mongo/transport/service_entry_point_impl.cpp | 50 | ||||
-rw-r--r-- | src/mongo/transport/service_entry_point_impl.h | 2 | ||||
-rw-r--r-- | src/mongo/transport/service_executor_fixed.cpp | 12 | ||||
-rw-r--r-- | src/mongo/transport/service_executor_fixed.h | 3 | ||||
-rw-r--r-- | src/mongo/transport/service_executor_reserved.cpp | 19 | ||||
-rw-r--r-- | src/mongo/transport/service_executor_reserved.h | 3 | ||||
-rw-r--r-- | src/mongo/transport/service_executor_synchronous.cpp | 11 | ||||
-rw-r--r-- | src/mongo/transport/service_executor_synchronous.h | 5 | ||||
-rw-r--r-- | src/mongo/transport/service_state_machine.cpp | 5 | ||||
-rw-r--r-- | src/mongo/transport/service_state_machine.h | 2 | ||||
-rw-r--r-- | src/mongo/transport/transport_layer_manager.cpp | 2 |
18 files changed, 105 insertions, 83 deletions
diff --git a/src/mongo/db/commands/server_status_servers.cpp b/src/mongo/db/commands/server_status_servers.cpp index 85a57202370..cc39b6da7d8 100644 --- a/src/mongo/db/commands/server_status_servers.cpp +++ b/src/mongo/db/commands/server_status_servers.cpp @@ -33,6 +33,7 @@ #include "mongo/db/commands/server_status.h" #include "mongo/transport/message_compressor_registry.h" #include "mongo/transport/service_entry_point.h" +#include "mongo/transport/service_executor_synchronous.h" #include "mongo/util/net/hostname_canonicalization.h" #include "mongo/util/net/socket_utils.h" #include "mongo/util/net/ssl_manager.h" @@ -77,12 +78,13 @@ public: return true; } + // TODO: need to track connections in server stats (see SERVER-49073) BSONObj generateSection(OperationContext* opCtx, const BSONElement& configElement) const override { BSONObjBuilder b; networkCounter.append(b); appendMessageCompressionStats(&b); - auto executor = opCtx->getServiceContext()->getServiceExecutor(); + auto executor = transport::ServiceExecutorSynchronous::get(opCtx->getServiceContext()); if (executor) { BSONObjBuilder section(b.subobjStart("serviceExecutorTaskStats")); executor->appendStats(§ion); diff --git a/src/mongo/db/mongod_main.cpp b/src/mongo/db/mongod_main.cpp index 368c7605491..553ef9843bd 100644 --- a/src/mongo/db/mongod_main.cpp +++ b/src/mongo/db/mongod_main.cpp @@ -734,16 +734,7 @@ ExitCode _initAndListen(ServiceContext* serviceContext, int listenPort) { // operation context anymore startupOpCtx.reset(); - auto start = serviceContext->getServiceExecutor()->start(); - if (!start.isOK()) { - LOGV2_ERROR(20570, - "Error starting service executor: {error}", - "Error starting service executor", - "error"_attr = start); - return EXIT_NET_ERROR; - } - - start = serviceContext->getServiceEntryPoint()->start(); + auto start = serviceContext->getServiceEntryPoint()->start(); if (!start.isOK()) { LOGV2_ERROR(20571, "Error starting service entry point: {error}", @@ -1279,18 +1270,6 @@ void shutdownTask(const ShutdownTaskArgs& shutdownArgs) { "Service entry point did not shutdown within the time limit"); } } - - // Shutdown and wait for the service executor to exit - if (auto svcExec = serviceContext->getServiceExecutor()) { - LOGV2_OPTIONS(4784924, {LogComponent::kExecutor}, "Shutting down the service executor"); - Status status = svcExec->shutdown(Seconds(10)); - if (!status.isOK()) { - LOGV2_OPTIONS(20564, - {LogComponent::kNetwork}, - "Service executor did not shutdown within the time limit", - "error"_attr = status); - } - } #endif LOGV2(4784925, "Shutting down free monitoring"); diff --git a/src/mongo/db/service_context.cpp b/src/mongo/db/service_context.cpp index 9d9f907aa1b..e94e71b2629 100644 --- a/src/mongo/db/service_context.cpp +++ b/src/mongo/db/service_context.cpp @@ -199,10 +199,6 @@ ServiceEntryPoint* ServiceContext::getServiceEntryPoint() const { return _serviceEntryPoint.get(); } -transport::ServiceExecutor* ServiceContext::getServiceExecutor() const { - return _serviceExecutor.get(); -} - void ServiceContext::setStorageEngine(std::unique_ptr<StorageEngine> engine) { invariant(engine); invariant(!_storageEngine); @@ -233,10 +229,6 @@ void ServiceContext::setTransportLayer(std::unique_ptr<transport::TransportLayer _transportLayer = std::move(tl); } -void ServiceContext::setServiceExecutor(std::unique_ptr<transport::ServiceExecutor> exec) { - _serviceExecutor = std::move(exec); -} - void ServiceContext::ClientDeleter::operator()(Client* client) const { ServiceContext* const service = client->getServiceContext(); { diff --git a/src/mongo/db/service_context.h b/src/mongo/db/service_context.h index 49cb7a602ba..5ad7c9c0e79 100644 --- a/src/mongo/db/service_context.h +++ b/src/mongo/db/service_context.h @@ -488,14 +488,6 @@ public: ServiceEntryPoint* getServiceEntryPoint() const; /** - * Get the service executor for the service context. - * - * See ServiceStateMachine for how this is used. Some configurations may not have a service - * executor registered and this will return a nullptr. - */ - transport::ServiceExecutor* getServiceExecutor() const; - - /** * Waits for the ServiceContext to be fully initialized and for all TransportLayers to have been * added/started. * @@ -580,11 +572,6 @@ public: void setTransportLayer(std::unique_ptr<transport::TransportLayer> tl); /** - * Binds the service executor to the service context - */ - void setServiceExecutor(std::unique_ptr<transport::ServiceExecutor> exec); - - /** * Creates a delayed execution baton with basic functionality */ BatonHandle makeBaton(OperationContext* opCtx) const; diff --git a/src/mongo/s/mongos_main.cpp b/src/mongo/s/mongos_main.cpp index 92ea2adc996..cb750bcc9e7 100644 --- a/src/mongo/s/mongos_main.cpp +++ b/src/mongo/s/mongos_main.cpp @@ -371,16 +371,6 @@ void cleanupTask(const ShutdownTaskArgs& shutdownArgs) { } } - // Shutdown and wait for the service executor to exit - if (auto svcExec = serviceContext->getServiceExecutor()) { - Status status = svcExec->shutdown(Seconds(5)); - if (!status.isOK()) { - LOGV2_OPTIONS(22845, - {LogComponent::kNetwork}, - "Service executor did not shutdown within the time limit", - "error"_attr = status); - } - } #endif // Shutdown Full-Time Data Capture @@ -788,15 +778,6 @@ ExitCode runMongosServer(ServiceContext* serviceContext) { std::make_unique<SessionsCollectionSharded>(), RouterSessionCatalog::reapSessionsOlderThan)); - status = serviceContext->getServiceExecutor()->start(); - if (!status.isOK()) { - LOGV2_ERROR(22859, - "Error starting service executor: {error}", - "Error starting service executor", - "error"_attr = redact(status)); - return EXIT_NET_ERROR; - } - status = serviceContext->getServiceEntryPoint()->start(); if (!status.isOK()) { LOGV2_ERROR(22860, diff --git a/src/mongo/tools/bridge.cpp b/src/mongo/tools/bridge.cpp index e377fae9be8..98babf0854f 100644 --- a/src/mongo/tools/bridge.cpp +++ b/src/mongo/tools/bridge.cpp @@ -478,10 +478,6 @@ int bridgeMain(int argc, char** argv) { setGlobalServiceContext(ServiceContext::make()); auto serviceContext = getGlobalServiceContext(); serviceContext->setServiceEntryPoint(std::make_unique<ServiceEntryPointBridge>(serviceContext)); - serviceContext->setServiceExecutor( - std::make_unique<transport::ServiceExecutorSynchronous>(serviceContext)); - - fassert(50766, serviceContext->getServiceExecutor()->start()); transport::TransportLayerASIO::Options opts; opts.ipList.emplace_back("0.0.0.0"); diff --git a/src/mongo/transport/SConscript b/src/mongo/transport/SConscript index 813401c3c94..1dfb23c5028 100644 --- a/src/mongo/transport/SConscript +++ b/src/mongo/transport/SConscript @@ -96,6 +96,7 @@ tlEnv.Library( ], LIBDEPS_PRIVATE=[ "$BUILD_DIR/mongo/idl/server_parameter", + "$BUILD_DIR/mongo/db/server_options_core", "$BUILD_DIR/mongo/util/concurrency/thread_pool", "$BUILD_DIR/mongo/util/processinfo", '$BUILD_DIR/third_party/shim_asio', @@ -178,7 +179,7 @@ tlEnv.CppUnitTest( 'transport_layer_asio_test.cpp', 'service_executor_test.cpp', 'max_conns_override_test.cpp', - 'service_state_machine_test.cpp', + # 'service_state_machine_test.cpp', ], LIBDEPS=[ '$BUILD_DIR/mongo/base', diff --git a/src/mongo/transport/service_entry_point_impl.cpp b/src/mongo/transport/service_entry_point_impl.cpp index 7c7d39cf6e9..230f03dda88 100644 --- a/src/mongo/transport/service_entry_point_impl.cpp +++ b/src/mongo/transport/service_entry_point_impl.cpp @@ -122,12 +122,26 @@ ServiceEntryPointImpl::ServiceEntryPointImpl(ServiceContext* svcCtx) : _svcCtx(s } Status ServiceEntryPointImpl::start() { - if (_adminInternalPool) - return _adminInternalPool->start(); - else - return Status::OK(); + if (auto status = transport::ServiceExecutorSynchronous::get(_svcCtx)->start(); + !status.isOK()) { + return status; + } + + if (auto exec = transport::ServiceExecutorReserved::get(_svcCtx); exec) { + if (auto status = exec->start(); !status.isOK()) { + return status; + } + } + + // TODO: Reintroduce SEF once it is attached as initial SE in SERVER-49109 + // if (auto status = transport::ServiceExecutorFixed::get(_svcCtx)->start(); !status.isOK()) { + // return status; + // } + + return Status::OK(); } +// TODO: explicitly start on the fixed executor void ServiceEntryPointImpl::startSession(transport::SessionHandle session) { // Setup the restriction environment on the Session, if the Session has local/remote Sockaddrs const auto& remoteAddr = session->remoteAddr(); @@ -140,7 +154,7 @@ void ServiceEntryPointImpl::startSession(transport::SessionHandle session) { const bool quiet = serverGlobalParams.quiet.load(); size_t connectionCount; - auto transportMode = _svcCtx->getServiceExecutor()->transportMode(); + auto transportMode = transport::ServiceExecutorSynchronous::get(_svcCtx)->transportMode(); auto ssm = ServiceStateMachine::create(_svcCtx, session, transportMode); auto usingMaxConnOverride = false; @@ -168,8 +182,9 @@ void ServiceEntryPointImpl::startSession(transport::SessionHandle session) { "connectionCount"_attr = connectionCount); } return; - } else if (usingMaxConnOverride && _adminInternalPool) { - ssm->setServiceExecutor(_adminInternalPool.get()); + } else if (auto exec = transport::ServiceExecutorReserved::get(_svcCtx); + usingMaxConnOverride && exec) { + ssm->setServiceExecutor(exec); } if (!quiet) { @@ -256,6 +271,27 @@ bool ServiceEntryPointImpl::shutdown(Milliseconds timeout) { "shutdown: exhausted grace period active workers to drain; continuing with shutdown...", "workers"_attr = numOpenSessions()); } + + lk.unlock(); + // TODO: Reintroduce SEF once it is attached as initial SE in SERVER-49109 + // if (auto status = transport::ServiceExecutorFixed::get(_svcCtx)->shutdown(timeout - + // timeSpent); + // !status.isOK()) { + // LOGV2(4907202, "Failed to shutdown ServiceExecutorFixed", "error"_attr = status); + // } + + if (auto exec = transport::ServiceExecutorReserved::get(_svcCtx)) { + if (auto status = exec->shutdown(timeout - timeSpent); !status.isOK()) { + LOGV2(4907201, "Failed to shutdown ServiceExecutorReserved", "error"_attr = status); + } + } + + if (auto status = + transport::ServiceExecutorSynchronous::get(_svcCtx)->shutdown(timeout - timeSpent); + !status.isOK()) { + LOGV2(4907200, "Failed to shutdown ServiceExecutorSynchronous", "error"_attr = status); + } + return result; } diff --git a/src/mongo/transport/service_entry_point_impl.h b/src/mongo/transport/service_entry_point_impl.h index 38c1319657a..ad9caf9a3fd 100644 --- a/src/mongo/transport/service_entry_point_impl.h +++ b/src/mongo/transport/service_entry_point_impl.h @@ -36,7 +36,9 @@ #include "mongo/stdx/condition_variable.h" #include "mongo/stdx/variant.h" #include "mongo/transport/service_entry_point.h" +#include "mongo/transport/service_executor_fixed.h" #include "mongo/transport/service_executor_reserved.h" +#include "mongo/transport/service_executor_synchronous.h" #include "mongo/transport/service_state_machine.h" #include "mongo/util/hierarchical_acquisition.h" #include "mongo/util/net/cidr.h" diff --git a/src/mongo/transport/service_executor_fixed.cpp b/src/mongo/transport/service_executor_fixed.cpp index ba8eca2bfaf..5ccb9657b06 100644 --- a/src/mongo/transport/service_executor_fixed.cpp +++ b/src/mongo/transport/service_executor_fixed.cpp @@ -80,6 +80,18 @@ Status ServiceExecutorFixed::start() { return Status::OK(); } +const auto getServiceExecutorFixed = + ServiceContext::declareDecoration<std::unique_ptr<ServiceExecutorFixed>>(); + +ServiceExecutorFixed* ServiceExecutorFixed::get(ServiceContext* ctx) { + auto& ref = getServiceExecutorFixed(ctx); + if (!ref) { + ThreadPool::Options options{}; + ref = std::make_unique<ServiceExecutorFixed>(options); + } + return ref.get(); +} + Status ServiceExecutorFixed::shutdown(Milliseconds timeout) { auto waitForShutdown = [&]() mutable -> Status { stdx::unique_lock<Latch> lk(_mutex); diff --git a/src/mongo/transport/service_executor_fixed.h b/src/mongo/transport/service_executor_fixed.h index c418296db8b..ae82a530d02 100644 --- a/src/mongo/transport/service_executor_fixed.h +++ b/src/mongo/transport/service_executor_fixed.h @@ -32,6 +32,7 @@ #include <memory> #include "mongo/base/status.h" +#include "mongo/db/service_context.h" #include "mongo/platform/atomic_word.h" #include "mongo/platform/mutex.h" #include "mongo/stdx/condition_variable.h" @@ -54,6 +55,8 @@ public: explicit ServiceExecutorFixed(ThreadPool::Options options); virtual ~ServiceExecutorFixed(); + static ServiceExecutorFixed* get(ServiceContext* ctx); + Status start() override; Status shutdown(Milliseconds timeout) override; Status schedule(Task task, ScheduleFlags flags) override; diff --git a/src/mongo/transport/service_executor_reserved.cpp b/src/mongo/transport/service_executor_reserved.cpp index f262f030583..35db496e279 100644 --- a/src/mongo/transport/service_executor_reserved.cpp +++ b/src/mongo/transport/service_executor_reserved.cpp @@ -33,6 +33,7 @@ #include "mongo/transport/service_executor_reserved.h" +#include "mongo/db/server_options.h" #include "mongo/logv2/log.h" #include "mongo/stdx/thread.h" #include "mongo/transport/service_entry_point_utils.h" @@ -147,6 +148,20 @@ Status ServiceExecutorReserved::_startWorker() { }); } +const auto getServiceExecutorReserved = + ServiceContext::declareDecoration<std::unique_ptr<ServiceExecutorReserved>>(); + +ServiceExecutorReserved* ServiceExecutorReserved::get(ServiceContext* ctx) { + auto& ref = getServiceExecutorReserved(ctx); + if (!ref) { + if (serverGlobalParams.reservedAdminThreads) { + ref = std::make_unique<transport::ServiceExecutorReserved>( + ctx, "admin/internal connections", serverGlobalParams.reservedAdminThreads); + } else + return nullptr; + } + return ref.get(); +} Status ServiceExecutorReserved::shutdown(Milliseconds timeout) { LOGV2_DEBUG(22980, 3, "Shutting down reserved executor"); @@ -173,8 +188,8 @@ Status ServiceExecutorReserved::schedule(Task task, ScheduleFlags flags) { if (!_localWorkQueue.empty()) { // Execute task directly (recurse) if allowed by the caller as it produced better // performance in testing. Try to limit the amount of recursion so we don't blow up the - // stack, even though this shouldn't happen with this executor that uses blocking network - // I/O. + // stack, even though this shouldn't happen with this executor that uses blocking + // network I/O. if ((flags & ScheduleFlags::kMayRecurse) && (_localRecursionDepth < reservedServiceExecutorRecursionLimit.loadRelaxed())) { ++_localRecursionDepth; diff --git a/src/mongo/transport/service_executor_reserved.h b/src/mongo/transport/service_executor_reserved.h index b6b0f17f803..a6ec63b82e1 100644 --- a/src/mongo/transport/service_executor_reserved.h +++ b/src/mongo/transport/service_executor_reserved.h @@ -32,6 +32,7 @@ #include <deque> #include "mongo/base/status.h" +#include "mongo/db/service_context.h" #include "mongo/platform/atomic_word.h" #include "mongo/platform/mutex.h" #include "mongo/stdx/condition_variable.h" @@ -54,6 +55,8 @@ class ServiceExecutorReserved final : public ServiceExecutor { public: explicit ServiceExecutorReserved(ServiceContext* ctx, std::string name, size_t reservedThreads); + static ServiceExecutorReserved* get(ServiceContext* ctx); + Status start() override; Status shutdown(Milliseconds timeout) override; Status schedule(Task task, ScheduleFlags flags) override; diff --git a/src/mongo/transport/service_executor_synchronous.cpp b/src/mongo/transport/service_executor_synchronous.cpp index 10cc0ad3cce..d58a917841e 100644 --- a/src/mongo/transport/service_executor_synchronous.cpp +++ b/src/mongo/transport/service_executor_synchronous.cpp @@ -78,6 +78,17 @@ Status ServiceExecutorSynchronous::shutdown(Milliseconds timeout) { "passthrough executor couldn't shutdown all worker threads within time limit."); } +const auto getServiceExecutorSynchronous = + ServiceContext::declareDecoration<std::unique_ptr<ServiceExecutorSynchronous>>(); + +ServiceExecutorSynchronous* ServiceExecutorSynchronous::get(ServiceContext* ctx) { + auto& ref = getServiceExecutorSynchronous(ctx); + if (!ref) { + ref = std::make_unique<ServiceExecutorSynchronous>(); + } + return ref.get(); +} + Status ServiceExecutorSynchronous::schedule(Task task, ScheduleFlags flags) { if (!_stillRunning.load()) { return Status{ErrorCodes::ShutdownInProgress, "Executor is not running"}; diff --git a/src/mongo/transport/service_executor_synchronous.h b/src/mongo/transport/service_executor_synchronous.h index 5fee75bbf2d..929418e7968 100644 --- a/src/mongo/transport/service_executor_synchronous.h +++ b/src/mongo/transport/service_executor_synchronous.h @@ -32,6 +32,7 @@ #include <deque> #include "mongo/base/status.h" +#include "mongo/db/service_context.h" #include "mongo/platform/atomic_word.h" #include "mongo/platform/mutex.h" #include "mongo/stdx/condition_variable.h" @@ -47,7 +48,9 @@ namespace transport { */ class ServiceExecutorSynchronous final : public ServiceExecutor { public: - explicit ServiceExecutorSynchronous(ServiceContext* ctx); + explicit ServiceExecutorSynchronous(ServiceContext* ctx = getGlobalServiceContext()); + + static ServiceExecutorSynchronous* get(ServiceContext* ctx); Status start() override; Status shutdown(Milliseconds timeout) override; diff --git a/src/mongo/transport/service_state_machine.cpp b/src/mongo/transport/service_state_machine.cpp index fb09fa9ad85..0c1d0b1b885 100644 --- a/src/mongo/transport/service_state_machine.cpp +++ b/src/mongo/transport/service_state_machine.cpp @@ -45,6 +45,7 @@ #include "mongo/rpc/op_msg.h" #include "mongo/transport/message_compressor_manager.h" #include "mongo/transport/service_entry_point.h" +#include "mongo/transport/service_executor_synchronous.h" #include "mongo/transport/session.h" #include "mongo/transport/transport_layer.h" #include "mongo/util/assert_util.h" @@ -300,11 +301,11 @@ ServiceStateMachine::ServiceStateMachine(ServiceContext* svcContext, _sep{svcContext->getServiceEntryPoint()}, _transportMode(transportMode), _serviceContext(svcContext), - _serviceExecutor(_serviceContext->getServiceExecutor()), _sessionHandle(session), _threadName{str::stream() << "conn" << _session()->id()}, _dbClient{svcContext->makeClient(_threadName, std::move(session))}, - _dbClientPtr{_dbClient.get()} {} + _dbClientPtr{_dbClient.get()}, + _serviceExecutor(transport::ServiceExecutorSynchronous::get(_serviceContext)) {} const transport::SessionHandle& ServiceStateMachine::_session() const { return _sessionHandle; diff --git a/src/mongo/transport/service_state_machine.h b/src/mongo/transport/service_state_machine.h index 0572c4f6ac8..22e0b8aa752 100644 --- a/src/mongo/transport/service_state_machine.h +++ b/src/mongo/transport/service_state_machine.h @@ -230,12 +230,12 @@ private: transport::Mode _transportMode; ServiceContext* const _serviceContext; - transport::ServiceExecutor* _serviceExecutor; transport::SessionHandle _sessionHandle; const std::string _threadName; ServiceContext::UniqueClient _dbClient; const Client* _dbClientPtr; + transport::ServiceExecutor* _serviceExecutor; std::function<void()> _cleanupHook; bool _inExhaust = false; diff --git a/src/mongo/transport/transport_layer_manager.cpp b/src/mongo/transport/transport_layer_manager.cpp index 2f7508e5c1a..e6e930e085a 100644 --- a/src/mongo/transport/transport_layer_manager.cpp +++ b/src/mongo/transport/transport_layer_manager.cpp @@ -137,8 +137,6 @@ std::unique_ptr<TransportLayer> TransportLayerManager::createWithConfig( transport::TransportLayerASIO::Options opts(config); opts.transportMode = transport::Mode::kSynchronous; - ctx->setServiceExecutor(std::make_unique<ServiceExecutorSynchronous>(ctx)); - std::vector<std::unique_ptr<TransportLayer>> retVector; retVector.emplace_back(std::make_unique<transport::TransportLayerASIO>(opts, sep)); return std::make_unique<TransportLayerManager>(std::move(retVector)); |