// Copyright 2017-2020 Parity Technologies (UK) Ltd. // This file is part of Polkadot. // Polkadot is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // Polkadot is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with Polkadot. If not, see . //! Utility module for subsystems //! //! Many subsystems have common interests such as canceling a bunch of spawned jobs, //! or determining what their validator ID is. These common interests are factored into //! this module. //! //! This crate also reexports Prometheus metric types which are expected to be implemented by subsystems. #![warn(missing_docs)] use polkadot_node_subsystem::{ errors::RuntimeApiError, messages::{AllMessages, RuntimeApiMessage, RuntimeApiRequest, RuntimeApiSender, BoundToRelayParent}, FromOverseer, SpawnedSubsystem, Subsystem, SubsystemContext, SubsystemError, SubsystemResult, }; use polkadot_node_jaeger as jaeger; use futures::{channel::{mpsc, oneshot}, prelude::*, select, stream::Stream}; use futures_timer::Delay; use parity_scale_codec::Encode; use pin_project::pin_project; use polkadot_primitives::v1::{ CandidateEvent, CommittedCandidateReceipt, CoreState, EncodeAs, PersistedValidationData, GroupRotationInfo, Hash, Id as ParaId, OccupiedCoreAssumption, SessionIndex, Signed, SigningContext, ValidationCode, ValidatorId, ValidatorIndex, SessionInfo, }; use sp_core::{traits::SpawnNamed, Public}; use sp_application_crypto::AppKey; use sp_keystore::{CryptoStore, SyncCryptoStorePtr, Error as KeystoreError}; use std::{ collections::{HashMap, hash_map::Entry}, convert::{TryFrom, TryInto}, marker::Unpin, pin::Pin, task::{Poll, Context}, time::Duration, fmt, sync::Arc, }; use streamunordered::{StreamUnordered, StreamYield}; use thiserror::Error; pub mod validator_discovery; pub use metered_channel as metered; /// These reexports are required so that external crates can use the `delegated_subsystem` macro properly. pub mod reexports { pub use sp_core::traits::SpawnNamed; pub use polkadot_node_subsystem::{ SpawnedSubsystem, Subsystem, SubsystemContext, }; } /// Duration a job will wait after sending a stop signal before hard-aborting. pub const JOB_GRACEFUL_STOP_DURATION: Duration = Duration::from_secs(1); /// Capacity of channels to and from individual jobs pub const JOB_CHANNEL_CAPACITY: usize = 64; /// Utility errors #[derive(Debug, Error)] pub enum Error { /// Attempted to send or receive on a oneshot channel which had been canceled #[error(transparent)] Oneshot(#[from] oneshot::Canceled), /// Attempted to send on a MPSC channel which has been canceled #[error(transparent)] Mpsc(#[from] mpsc::SendError), /// A subsystem error #[error(transparent)] Subsystem(#[from] SubsystemError), /// An error in the Runtime API. #[error(transparent)] RuntimeApi(#[from] RuntimeApiError), /// The type system wants this even though it doesn't make sense #[error(transparent)] Infallible(#[from] std::convert::Infallible), /// Attempted to convert from an AllMessages to a FromJob, and failed. #[error("AllMessage not relevant to Job")] SenderConversion(String), /// The local node is not a validator. #[error("Node is not a validator")] NotAValidator, /// Already forwarding errors to another sender #[error("AlreadyForwarding")] AlreadyForwarding, } /// A type alias for Runtime API receivers. pub type RuntimeApiReceiver = oneshot::Receiver>; /// Request some data from the `RuntimeApi`. pub async fn request_from_runtime( parent: Hash, sender: &mut mpsc::Sender, request_builder: RequestBuilder, ) -> Result, Error> where RequestBuilder: FnOnce(RuntimeApiSender) -> RuntimeApiRequest, FromJob: From, { let (tx, rx) = oneshot::channel(); sender.send(AllMessages::RuntimeApi(RuntimeApiMessage::Request(parent, request_builder(tx))).into()).await?; Ok(rx) } /// Construct specialized request functions for the runtime. /// /// These would otherwise get pretty repetitive. macro_rules! specialize_requests { // expand return type name for documentation purposes (fn $func_name:ident( $( $param_name:ident : $param_ty:ty ),* ) -> $return_ty:ty ; $request_variant:ident;) => { specialize_requests!{ named stringify!($request_variant) ; fn $func_name( $( $param_name : $param_ty ),* ) -> $return_ty ; $request_variant; } }; // create a single specialized request function (named $doc_name:expr ; fn $func_name:ident( $( $param_name:ident : $param_ty:ty ),* ) -> $return_ty:ty ; $request_variant:ident;) => { #[doc = "Request `"] #[doc = $doc_name] #[doc = "` from the runtime"] pub async fn $func_name( parent: Hash, $( $param_name: $param_ty, )* sender: &mut mpsc::Sender, ) -> Result, Error> where FromJob: From, { request_from_runtime(parent, sender, |tx| RuntimeApiRequest::$request_variant( $( $param_name, )* tx )).await } }; // recursive decompose ( fn $func_name:ident( $( $param_name:ident : $param_ty:ty ),* ) -> $return_ty:ty ; $request_variant:ident; $( fn $t_func_name:ident( $( $t_param_name:ident : $t_param_ty:ty ),* ) -> $t_return_ty:ty ; $t_request_variant:ident; )+ ) => { specialize_requests!{ fn $func_name( $( $param_name : $param_ty ),* ) -> $return_ty ; $request_variant ; } specialize_requests!{ $( fn $t_func_name( $( $t_param_name : $t_param_ty ),* ) -> $t_return_ty ; $t_request_variant ; )+ } }; } specialize_requests! { fn request_validators() -> Vec; Validators; fn request_validator_groups() -> (Vec>, GroupRotationInfo); ValidatorGroups; fn request_availability_cores() -> Vec; AvailabilityCores; fn request_persisted_validation_data(para_id: ParaId, assumption: OccupiedCoreAssumption) -> Option; PersistedValidationData; fn request_session_index_for_child() -> SessionIndex; SessionIndexForChild; fn request_validation_code(para_id: ParaId, assumption: OccupiedCoreAssumption) -> Option; ValidationCode; fn request_candidate_pending_availability(para_id: ParaId) -> Option; CandidatePendingAvailability; fn request_candidate_events() -> Vec; CandidateEvents; fn request_session_info(index: SessionIndex) -> Option; SessionInfo; } /// Request some data from the `RuntimeApi` via a SubsystemContext. async fn request_from_runtime_ctx( parent: Hash, ctx: &mut Context, request_builder: RequestBuilder, ) -> Result, Error> where RequestBuilder: FnOnce(RuntimeApiSender) -> RuntimeApiRequest, Context: SubsystemContext, { let (tx, rx) = oneshot::channel(); ctx.send_message( AllMessages::RuntimeApi(RuntimeApiMessage::Request(parent, request_builder(tx))) .try_into() .map_err(|err| Error::SenderConversion(format!("{:?}", err)))?, ).await; Ok(rx) } /// Construct specialized request functions for the runtime. /// /// These would otherwise get pretty repetitive. macro_rules! specialize_requests_ctx { // expand return type name for documentation purposes (fn $func_name:ident( $( $param_name:ident : $param_ty:ty ),* ) -> $return_ty:ty ; $request_variant:ident;) => { specialize_requests_ctx!{ named stringify!($request_variant) ; fn $func_name( $( $param_name : $param_ty ),* ) -> $return_ty ; $request_variant; } }; // create a single specialized request function (named $doc_name:expr ; fn $func_name:ident( $( $param_name:ident : $param_ty:ty ),* ) -> $return_ty:ty ; $request_variant:ident;) => { #[doc = "Request `"] #[doc = $doc_name] #[doc = "` from the runtime via a `SubsystemContext`"] pub async fn $func_name( parent: Hash, $( $param_name: $param_ty, )* ctx: &mut Context, ) -> Result, Error> { request_from_runtime_ctx(parent, ctx, |tx| RuntimeApiRequest::$request_variant( $( $param_name, )* tx )).await } }; // recursive decompose ( fn $func_name:ident( $( $param_name:ident : $param_ty:ty ),* ) -> $return_ty:ty ; $request_variant:ident; $( fn $t_func_name:ident( $( $t_param_name:ident : $t_param_ty:ty ),* ) -> $t_return_ty:ty ; $t_request_variant:ident; )+ ) => { specialize_requests_ctx!{ fn $func_name( $( $param_name : $param_ty ),* ) -> $return_ty ; $request_variant ; } specialize_requests_ctx!{ $( fn $t_func_name( $( $t_param_name : $t_param_ty ),* ) -> $t_return_ty ; $t_request_variant ; )+ } }; } specialize_requests_ctx! { fn request_validators_ctx() -> Vec; Validators; fn request_validator_groups_ctx() -> (Vec>, GroupRotationInfo); ValidatorGroups; fn request_availability_cores_ctx() -> Vec; AvailabilityCores; fn request_persisted_validation_data_ctx(para_id: ParaId, assumption: OccupiedCoreAssumption) -> Option; PersistedValidationData; fn request_session_index_for_child_ctx() -> SessionIndex; SessionIndexForChild; fn request_validation_code_ctx(para_id: ParaId, assumption: OccupiedCoreAssumption) -> Option; ValidationCode; fn request_candidate_pending_availability_ctx(para_id: ParaId) -> Option; CandidatePendingAvailability; fn request_candidate_events_ctx() -> Vec; CandidateEvents; fn request_session_info_ctx(index: SessionIndex) -> Option; SessionInfo; } /// From the given set of validators, find the first key we can sign with, if any. pub async fn signing_key(validators: &[ValidatorId], keystore: SyncCryptoStorePtr) -> Option { for v in validators.iter() { if CryptoStore::has_keys(&*keystore, &[(v.to_raw_vec(), ValidatorId::ID)]).await { return Some(v.clone()); } } None } /// Local validator information /// /// It can be created if the local node is a validator in the context of a particular /// relay chain block. #[derive(Debug)] pub struct Validator { signing_context: SigningContext, key: ValidatorId, index: ValidatorIndex, } impl Validator { /// Get a struct representing this node's validator if this node is in fact a validator in the context of the given block. pub async fn new( parent: Hash, keystore: SyncCryptoStorePtr, mut sender: mpsc::Sender, ) -> Result where FromJob: From, { // Note: request_validators and request_session_index_for_child do not and cannot // run concurrently: they both have a mutable handle to the same sender. // However, each of them returns a oneshot::Receiver, and those are resolved concurrently. let (validators, session_index) = futures::try_join!( request_validators(parent, &mut sender).await?, request_session_index_for_child(parent, &mut sender).await?, )?; let signing_context = SigningContext { session_index: session_index?, parent_hash: parent, }; let validators = validators?; Self::construct(&validators, signing_context, keystore).await } /// Construct a validator instance without performing runtime fetches. /// /// This can be useful if external code also needs the same data. pub async fn construct( validators: &[ValidatorId], signing_context: SigningContext, keystore: SyncCryptoStorePtr, ) -> Result { let key = signing_key(validators, keystore).await.ok_or(Error::NotAValidator)?; let index = validators .iter() .enumerate() .find(|(_, k)| k == &&key) .map(|(idx, _)| ValidatorIndex(idx as u32)) .expect("signing_key would have already returned NotAValidator if the item we're searching for isn't in this list; qed"); Ok(Validator { signing_context, key, index, }) } /// Get this validator's id. pub fn id(&self) -> ValidatorId { self.key.clone() } /// Get this validator's local index. pub fn index(&self) -> ValidatorIndex { self.index } /// Get the current signing context. pub fn signing_context(&self) -> &SigningContext { &self.signing_context } /// Sign a payload with this validator pub async fn sign, RealPayload: Encode>( &self, keystore: SyncCryptoStorePtr, payload: Payload, ) -> Result>, KeystoreError> { Signed::sign(&keystore, payload, &self.signing_context, self.index, &self.key).await } /// Validate the payload with this validator /// /// Validation can only succeed if `signed.validator_index() == self.index()`. /// Normally, this will always be the case for a properly operating program, /// but it's double-checked here anyway. pub fn check_payload, RealPayload: Encode>( &self, signed: Signed, ) -> Result<(), ()> { if signed.validator_index() != self.index { return Err(()); } signed.check_signature(&self.signing_context, &self.id()) } } struct AbortOnDrop(future::AbortHandle); impl Drop for AbortOnDrop { fn drop(&mut self) { self.0.abort(); } } /// A JobHandle manages a particular job for a subsystem. struct JobHandle { _abort_handle: AbortOnDrop, to_job: mpsc::Sender, } impl JobHandle { /// Send a message to the job. async fn send_msg(&mut self, msg: ToJob) -> Result<(), Error> { self.to_job.send(msg).await.map_err(Into::into) } } /// This module reexports Prometheus types and defines the [`Metrics`] trait. pub mod metrics { /// Reexport Substrate Prometheus types. pub use substrate_prometheus_endpoint as prometheus; /// Subsystem- or job-specific Prometheus metrics. /// /// Usually implemented as a wrapper for `Option` /// to ensure `Default` bounds or as a dummy type (). /// Prometheus metrics internally hold an `Arc` reference, so cloning them is fine. pub trait Metrics: Default + Clone { /// Try to register metrics in the Prometheus registry. fn try_register(registry: &prometheus::Registry) -> Result; /// Convenience method to register metrics in the optional Promethius registry. /// /// If no registry is provided, returns `Default::default()`. Otherwise, returns the same /// thing that `try_register` does. fn register(registry: Option<&prometheus::Registry>) -> Result { match registry { None => Ok(Self::default()), Some(registry) => Self::try_register(registry), } } } // dummy impl impl Metrics for () { fn try_register(_registry: &prometheus::Registry) -> Result<(), prometheus::PrometheusError> { Ok(()) } } } /// Commands from a job to the broader subsystem. pub enum FromJobCommand { /// Send a message to another subsystem. SendMessage(AllMessages), /// Spawn a child task on the executor. Spawn(&'static str, Pin + Send>>), /// Spawn a blocking child task on the executor's dedicated thread pool. SpawnBlocking(&'static str, Pin + Send>>), } impl fmt::Debug for FromJobCommand { fn fmt(&self, fmt: &mut fmt::Formatter) -> fmt::Result { match self { Self::SendMessage(msg) => write!(fmt, "FromJobCommand::SendMessage({:?})", msg), Self::Spawn(name, _) => write!(fmt, "FromJobCommand::Spawn({})", name), Self::SpawnBlocking(name, _) => write!(fmt, "FromJobCommand::SpawnBlocking({})", name), } } } impl From for FromJobCommand { fn from(msg: AllMessages) -> Self { Self::SendMessage(msg) } } /// This trait governs jobs. /// /// Jobs are instantiated and killed automatically on appropriate overseer messages. /// Other messages are passed along to and from the job via the overseer to other subsystems. pub trait JobTrait: Unpin { /// Message type used to send messages to the job. type ToJob: 'static + BoundToRelayParent + Send; /// Job runtime error. type Error: 'static + std::error::Error + Send; /// Extra arguments this job needs to run properly. /// /// If no extra information is needed, it is perfectly acceptable to set it to `()`. type RunArgs: 'static + Send; /// Subsystem-specific Prometheus metrics. /// /// Jobs spawned by one subsystem should share the same /// instance of metrics (use `.clone()`). /// The `delegate_subsystem!` macro should take care of this. type Metrics: 'static + metrics::Metrics + Send; /// Name of the job, i.e. `CandidateBackingJob` const NAME: &'static str; /// Run a job for the given relay `parent`. /// /// The job should be ended when `receiver` returns `None`. fn run( parent: Hash, span: Arc, run_args: Self::RunArgs, metrics: Self::Metrics, receiver: mpsc::Receiver, sender: mpsc::Sender, ) -> Pin> + Send>>; } /// Error which can be returned by the jobs manager /// /// Wraps the utility error type and the job-specific error #[derive(Debug, Error)] pub enum JobsError { /// utility error #[error("Utility")] Utility(#[source] Error), /// internal job error #[error("Internal")] Job(#[source] JobError), } /// Jobs manager for a subsystem /// /// - Spawns new jobs for a given relay-parent on demand. /// - Closes old jobs for a given relay-parent on demand. /// - Dispatches messages to the appropriate job for a given relay-parent. /// - When dropped, aborts all remaining jobs. /// - implements `Stream`, collecting all messages from subordinate jobs. #[pin_project] pub struct Jobs { spawner: Spawner, running: HashMap>, outgoing_msgs: StreamUnordered>, #[pin] job: std::marker::PhantomData, errors: Option, JobsError)>>, } impl Jobs { /// Create a new Jobs manager which handles spawning appropriate jobs. pub fn new(spawner: Spawner) -> Self { Self { spawner, running: HashMap::new(), outgoing_msgs: StreamUnordered::new(), job: std::marker::PhantomData, errors: None, } } /// Monitor errors which may occur during handling of a spawned job. /// /// By default, an error in a job is simply logged. Once this is called, /// the error is forwarded onto the provided channel. /// /// Errors if the error channel already exists. pub fn forward_errors( &mut self, tx: mpsc::Sender<(Option, JobsError)>, ) -> Result<(), Error> { if self.errors.is_some() { return Err(Error::AlreadyForwarding); } self.errors = Some(tx); Ok(()) } /// Spawn a new job for this `parent_hash`, with whatever args are appropriate. fn spawn_job( &mut self, parent_hash: Hash, span: Arc, run_args: Job::RunArgs, metrics: Job::Metrics, ) -> Result<(), Error> { let (to_job_tx, to_job_rx) = mpsc::channel(JOB_CHANNEL_CAPACITY); let (from_job_tx, from_job_rx) = mpsc::channel(JOB_CHANNEL_CAPACITY); let err_tx = self.errors.clone(); let (future, abort_handle) = future::abortable(async move { if let Err(e) = Job::run(parent_hash, span, run_args, metrics, to_job_rx, from_job_tx).await { tracing::error!( job = Job::NAME, parent_hash = %parent_hash, err = ?e, "job finished with an error", ); if let Some(mut err_tx) = err_tx { // if we can't send the notification of error on the error channel, then // there's no point trying to propagate this error onto the channel too // all we can do is warn that error propagation has failed if let Err(e) = err_tx.send((Some(parent_hash), JobsError::Job(e))).await { tracing::warn!(err = ?e, "failed to forward error"); } } } }); self.spawner.spawn(Job::NAME, future.map(drop).boxed()); self.outgoing_msgs.push(from_job_rx); let handle = JobHandle { _abort_handle: AbortOnDrop(abort_handle), to_job: to_job_tx, }; self.running.insert(parent_hash, handle); Ok(()) } /// Stop the job associated with this `parent_hash`. pub async fn stop_job(&mut self, parent_hash: Hash) { self.running.remove(&parent_hash); } /// Send a message to the appropriate job for this `parent_hash`. async fn send_msg(&mut self, parent_hash: Hash, msg: Job::ToJob) { if let Entry::Occupied(mut job) = self.running.entry(parent_hash) { if job.get_mut().send_msg(msg).await.is_err() { job.remove(); } } } } impl Stream for Jobs where Spawner: SpawnNamed, Job: JobTrait, { type Item = FromJobCommand; fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context) -> Poll> { loop { match Pin::new(&mut self.outgoing_msgs).poll_next(cx) { Poll::Pending => return Poll::Pending, Poll::Ready(r) => match r.map(|v| v.0) { Some(StreamYield::Item(msg)) => return Poll::Ready(Some(msg)), // If a job is finished, rerun the loop Some(StreamYield::Finished(_)) => continue, // Don't end if there are no jobs running None => return Poll::Pending, } } } } } impl stream::FusedStream for Jobs where Spawner: SpawnNamed, Job: JobTrait, { fn is_terminated(&self) -> bool { false } } /// A basic implementation of a subsystem. /// /// This struct is responsible for handling message traffic between /// this subsystem and the overseer. It spawns and kills jobs on the /// appropriate Overseer messages, and dispatches standard traffic to /// the appropriate job the rest of the time. pub struct JobManager { spawner: Spawner, run_args: Job::RunArgs, metrics: Job::Metrics, context: std::marker::PhantomData, job: std::marker::PhantomData, errors: Option, JobsError)>>, } impl JobManager where Spawner: SpawnNamed + Clone + Send + Unpin, Context: SubsystemContext, Job: 'static + JobTrait, Job::RunArgs: Clone, Job::ToJob: From<::Message> + Sync, { /// Creates a new `Subsystem`. pub fn new(spawner: Spawner, run_args: Job::RunArgs, metrics: Job::Metrics) -> Self { Self { spawner, run_args, metrics, context: std::marker::PhantomData, job: std::marker::PhantomData, errors: None, } } /// Monitor errors which may occur during handling of a spawned job. /// /// By default, an error in a job is simply logged. Once this is called, /// the error is forwarded onto the provided channel. /// /// Errors if the error channel already exists. pub fn forward_errors( &mut self, tx: mpsc::Sender<(Option, JobsError)>, ) -> Result<(), Error> { if self.errors.is_some() { return Err(Error::AlreadyForwarding); } self.errors = Some(tx); Ok(()) } /// Run this subsystem /// /// Conceptually, this is very simple: it just loops forever. /// /// - On incoming overseer messages, it starts or stops jobs as appropriate. /// - On other incoming messages, if they can be converted into Job::ToJob and /// include a hash, then they're forwarded to the appropriate individual job. /// - On outgoing messages from the jobs, it forwards them to the overseer. /// /// If `err_tx` is not `None`, errors are forwarded onto that channel as they occur. /// Otherwise, most are logged and then discarded. pub async fn run( mut ctx: Context, run_args: Job::RunArgs, metrics: Job::Metrics, spawner: Spawner, mut err_tx: Option, JobsError)>>, ) { let mut jobs = Jobs::new(spawner.clone()); if let Some(ref err_tx) = err_tx { jobs.forward_errors(err_tx.clone()) .expect("we never call this twice in this context; qed"); } loop { select! { incoming = ctx.recv().fuse() => if Self::handle_incoming( incoming, &mut jobs, &run_args, &metrics, &mut err_tx, ).await { break }, outgoing = jobs.next() => { if let Err(e) = Self::handle_from_job(outgoing, &mut ctx).await { tracing::warn!(err = ?e, "failed to handle command from job"); } } complete => break, } } } /// Forward a given error to the higher context using the given error channel. async fn fwd_err( hash: Option, err: JobsError, err_tx: &mut Option, JobsError)>>, ) { if let Some(err_tx) = err_tx { // if we can't send on the error transmission channel, we can't do anything useful about it // still, we can at least log the failure if let Err(e) = err_tx.send((hash, err)).await { tracing::warn!(err = ?e, "failed to forward error"); } } } /// Handle an incoming message. /// /// Returns `true` when this job manager should shutdown. async fn handle_incoming( incoming: SubsystemResult>, jobs: &mut Jobs, run_args: &Job::RunArgs, metrics: &Job::Metrics, err_tx: &mut Option, JobsError)>>, ) -> bool { use polkadot_node_subsystem::ActiveLeavesUpdate; use polkadot_node_subsystem::FromOverseer::{Communication, Signal}; use polkadot_node_subsystem::OverseerSignal::{ActiveLeaves, BlockFinalized, Conclude}; match incoming { Ok(Signal(ActiveLeaves(ActiveLeavesUpdate { activated, deactivated, }))) => { for activated in activated { let metrics = metrics.clone(); if let Err(e) = jobs.spawn_job(activated.hash, activated.span, run_args.clone(), metrics) { tracing::error!( job = Job::NAME, err = ?e, "failed to spawn a job", ); Self::fwd_err(Some(activated.hash), JobsError::Utility(e), err_tx).await; return true; } } for hash in deactivated { jobs.stop_job(hash).await; } } Ok(Signal(Conclude)) => { jobs.running.clear(); return true; } Ok(Communication { msg }) => { if let Ok(to_job) = ::try_from(msg) { jobs.send_msg(to_job.relay_parent(), to_job).await; } } Ok(Signal(BlockFinalized(..))) => {} Err(err) => { tracing::error!( job = Job::NAME, err = ?err, "error receiving message from subsystem context for job", ); Self::fwd_err(None, JobsError::Utility(Error::from(err)), err_tx).await; return true; } } false } // handle a command from a job. async fn handle_from_job( outgoing: Option, ctx: &mut Context, ) -> SubsystemResult<()> { match outgoing.expect("the Jobs stream never ends; qed") { FromJobCommand::SendMessage(msg) => ctx.send_message(msg).await, FromJobCommand::Spawn(name, task) => ctx.spawn(name, task).await?, FromJobCommand::SpawnBlocking(name, task) => ctx.spawn_blocking(name, task).await?, } Ok(()) } } impl Subsystem for JobManager where Spawner: SpawnNamed + Send + Clone + Unpin + 'static, Context: SubsystemContext, Job: 'static + JobTrait + Send, Job::RunArgs: Clone + Sync, Job::ToJob: From<::Message> + Sync, Job::Metrics: Sync, { fn start(self, ctx: Context) -> SpawnedSubsystem { let spawner = self.spawner.clone(); let run_args = self.run_args.clone(); let metrics = self.metrics.clone(); let errors = self.errors; let future = Box::pin(async move { Self::run(ctx, run_args, metrics, spawner, errors).await; Ok(()) }); SpawnedSubsystem { name: Job::NAME.strip_suffix("Job").unwrap_or(Job::NAME), future, } } } /// Create a delegated subsystem /// /// It is possible to create a type which implements `Subsystem` by simply doing: /// /// ```ignore /// pub type ExampleSubsystem = JobManager; /// ``` /// /// However, doing this requires that job itself and all types which comprise it (i.e. `ToJob`, `FromJob`, `Error`, `RunArgs`) /// are public, to avoid exposing private types in public interfaces. It's possible to delegate instead, which /// can reduce the total number of public types exposed, i.e. /// /// ```ignore /// type Manager = JobManager; /// pub struct ExampleSubsystem { /// manager: Manager, /// } /// /// impl Subsystem for ExampleSubsystem { ... } /// ``` /// /// This dramatically reduces the number of public types in the crate; the only things which must be public are now /// /// - `struct ExampleSubsystem` (defined by this macro) /// - `type ToJob` (because it appears in a trait bound) /// - `type RunArgs` (because it appears in a function signature) /// /// Implementing this all manually is of course possible, but it's tedious; why bother? This macro exists for /// the purpose of doing it automatically: /// /// ```ignore /// delegated_subsystem!(ExampleJob(ExampleRunArgs) <- ExampleToJob as ExampleSubsystem); /// ``` #[macro_export] macro_rules! delegated_subsystem { ($job:ident($run_args:ty, $metrics:ty) <- $to_job:ty as $subsystem:ident) => { delegated_subsystem!($job($run_args, $metrics) <- $to_job as $subsystem; stringify!($subsystem)); }; ($job:ident($run_args:ty, $metrics:ty) <- $to_job:ty as $subsystem:ident; $subsystem_name:expr) => { #[doc = "Manager type for the "] #[doc = $subsystem_name] type Manager = $crate::JobManager; #[doc = "An implementation of the "] #[doc = $subsystem_name] pub struct $subsystem { manager: Manager, } impl $subsystem where Spawner: Clone + $crate::reexports::SpawnNamed + Send + Unpin, Context: $crate::reexports::SubsystemContext, $to_job: From<::Message>, { #[doc = "Creates a new "] #[doc = $subsystem_name] pub fn new(spawner: Spawner, run_args: $run_args, metrics: $metrics) -> Self { $subsystem { manager: $crate::JobManager::new(spawner, run_args, metrics) } } /// Run this subsystem #[tracing::instrument(skip(ctx, run_args, metrics, spawner), fields(subsystem = $subsystem_name))] pub async fn run(ctx: Context, run_args: $run_args, metrics: $metrics, spawner: Spawner) { >::run(ctx, run_args, metrics, spawner, None).await } } impl $crate::reexports::Subsystem for $subsystem where Spawner: $crate::reexports::SpawnNamed + Send + Clone + Unpin + 'static, Context: $crate::reexports::SubsystemContext, $to_job: From<::Message>, { fn start(self, ctx: Context) -> $crate::reexports::SpawnedSubsystem { self.manager.start(ctx) } } }; } /// A future that wraps another future with a `Delay` allowing for time-limited futures. #[pin_project] pub struct Timeout { #[pin] future: F, #[pin] delay: Delay, } /// Extends `Future` to allow time-limited futures. pub trait TimeoutExt: Future { /// Adds a timeout of `duration` to the given `Future`. /// Returns a new `Future`. fn timeout(self, duration: Duration) -> Timeout where Self: Sized, { Timeout { future: self, delay: Delay::new(duration), } } } impl TimeoutExt for F {} impl Future for Timeout { type Output = Option; fn poll(self: Pin<&mut Self>, ctx: &mut Context) -> Poll { let this = self.project(); if this.delay.poll(ctx).is_ready() { return Poll::Ready(None); } if let Poll::Ready(output) = this.future.poll(ctx) { return Poll::Ready(Some(output)); } Poll::Pending } } #[derive(Copy, Clone)] enum MetronomeState { Snooze, SetAlarm, } /// Create a stream of ticks with a defined cycle duration. pub struct Metronome { delay: Delay, period: Duration, state: MetronomeState, } impl Metronome { /// Create a new metronome source with a defined cycle duration. pub fn new(cycle: Duration) -> Self { let period = cycle.into(); Self { period, delay: Delay::new(period), state: MetronomeState::Snooze, } } } impl futures::Stream for Metronome { type Item = (); fn poll_next( mut self: Pin<&mut Self>, cx: &mut Context<'_> ) -> Poll> { loop { match self.state { MetronomeState::SetAlarm => { let val = self.period.clone(); self.delay.reset(val); self.state = MetronomeState::Snooze; } MetronomeState::Snooze => { if !Pin::new(&mut self.delay).poll(cx).is_ready() { break } self.state = MetronomeState::SetAlarm; return Poll::Ready(Some(())); } } } Poll::Pending } } #[cfg(test)] mod tests { use super::*; use executor::block_on; use thiserror::Error; use polkadot_node_jaeger as jaeger; use polkadot_node_subsystem::{ messages::{AllMessages, CandidateSelectionMessage}, ActiveLeavesUpdate, FromOverseer, OverseerSignal, SpawnedSubsystem, ActivatedLeaf, }; use assert_matches::assert_matches; use futures::{channel::mpsc, executor, StreamExt, future, Future, FutureExt, SinkExt}; use polkadot_primitives::v1::Hash; use polkadot_node_subsystem_test_helpers::{self as test_helpers, make_subsystem_context}; use std::{pin::Pin, sync::{Arc, atomic::{AtomicUsize, Ordering}}, time::Duration}; // basic usage: in a nutshell, when you want to define a subsystem, just focus on what its jobs do; // you can leave the subsystem itself to the job manager. // for purposes of demonstration, we're going to whip up a fake subsystem. // this will 'select' candidates which are pre-loaded in the job // job structs are constructed within JobTrait::run // most will want to retain the sender and receiver, as well as whatever other data they like struct FakeCandidateSelectionJob { receiver: mpsc::Receiver, } // Error will mostly be a wrapper to make the try operator more convenient; // deriving From implementations for most variants is recommended. // It must implement Debug for logging. #[derive(Debug, Error)] enum Error { #[error(transparent)] Sending(#[from]mpsc::SendError), } impl JobTrait for FakeCandidateSelectionJob { type ToJob = CandidateSelectionMessage; type Error = Error; type RunArgs = bool; type Metrics = (); const NAME: &'static str = "FakeCandidateSelectionJob"; /// Run a job for the parent block indicated // // this function is in charge of creating and executing the job's main loop fn run( _: Hash, _: Arc, run_args: Self::RunArgs, _metrics: Self::Metrics, receiver: mpsc::Receiver, mut sender: mpsc::Sender, ) -> Pin> + Send>> { async move { let job = FakeCandidateSelectionJob { receiver }; if run_args { sender.send(FromJobCommand::SendMessage( CandidateSelectionMessage::Invalid( Default::default(), Default::default(), ).into(), )).await?; } // it isn't necessary to break run_loop into its own function, // but it's convenient to separate the concerns in this way job.run_loop().await } .boxed() } } impl FakeCandidateSelectionJob { async fn run_loop(mut self) -> Result<(), Error> { loop { match self.receiver.next().await { Some(_csm) => { unimplemented!("we'd report the collator to the peer set manager here, but that's not implemented yet"); } None => break, } } Ok(()) } } // with the job defined, it's straightforward to get a subsystem implementation. type FakeCandidateSelectionSubsystem = JobManager; // this type lets us pretend to be the overseer type OverseerHandle = test_helpers::TestSubsystemContextHandle; fn test_harness>( run_args: bool, test: impl FnOnce(OverseerHandle, mpsc::Receiver<(Option, JobsError)>) -> T, ) { let _ = env_logger::builder() .is_test(true) .filter( None, log::LevelFilter::Trace, ) .try_init(); let pool = sp_core::testing::TaskExecutor::new(); let (context, overseer_handle) = make_subsystem_context(pool.clone()); let (err_tx, err_rx) = mpsc::channel(16); let subsystem = FakeCandidateSelectionSubsystem::run(context, run_args, (), pool, Some(err_tx)); let test_future = test(overseer_handle, err_rx); futures::pin_mut!(subsystem, test_future); executor::block_on(async move { future::join(subsystem, test_future) .timeout(Duration::from_secs(2)) .await .expect("test timed out instead of completing") }); } #[test] fn starting_and_stopping_job_works() { let relay_parent: Hash = [0; 32].into(); test_harness(true, |mut overseer_handle, err_rx| async move { overseer_handle .send(FromOverseer::Signal(OverseerSignal::ActiveLeaves( ActiveLeavesUpdate::start_work(ActivatedLeaf { hash: relay_parent, number: 1, span: Arc::new(jaeger::Span::Disabled), }), ))) .await; assert_matches!( overseer_handle.recv().await, AllMessages::CandidateSelection(_) ); overseer_handle .send(FromOverseer::Signal(OverseerSignal::ActiveLeaves( ActiveLeavesUpdate::stop_work(relay_parent), ))) .await; overseer_handle .send(FromOverseer::Signal(OverseerSignal::Conclude)) .await; let errs: Vec<_> = err_rx.collect().await; assert_eq!(errs.len(), 0); }); } #[test] fn sending_to_a_non_running_job_do_not_stop_the_subsystem() { let relay_parent = Hash::repeat_byte(0x01); test_harness(true, |mut overseer_handle, err_rx| async move { overseer_handle .send(FromOverseer::Signal(OverseerSignal::ActiveLeaves( ActiveLeavesUpdate::start_work(ActivatedLeaf { hash: relay_parent, number: 1, span: Arc::new(jaeger::Span::Disabled), }), ))) .await; // send to a non running job overseer_handle .send(FromOverseer::Communication { msg: Default::default(), }) .await; // the subsystem is still alive assert_matches!( overseer_handle.recv().await, AllMessages::CandidateSelection(_) ); overseer_handle .send(FromOverseer::Signal(OverseerSignal::Conclude)) .await; let errs: Vec<_> = err_rx.collect().await; assert_eq!(errs.len(), 0); }); } #[test] fn test_subsystem_impl_and_name_derivation() { let pool = sp_core::testing::TaskExecutor::new(); let (context, _) = make_subsystem_context::(pool.clone()); let SpawnedSubsystem { name, .. } = FakeCandidateSelectionSubsystem::new(pool, false, ()).start(context); assert_eq!(name, "FakeCandidateSelection"); } #[test] fn tick_tack_metronome() { let n = Arc::new(AtomicUsize::default()); let (tick, mut block) = mpsc::unbounded(); let metronome = { let n = n.clone(); let stream = Metronome::new(Duration::from_millis(137_u64)); stream.for_each(move |_res| { let _ = n.fetch_add(1, Ordering::Relaxed); let mut tick = tick.clone(); async move { tick.send(()).await.expect("Test helper channel works. qed"); } }).fuse() }; let f2 = async move { block.next().await; assert_eq!(n.load(Ordering::Relaxed), 1_usize); block.next().await; assert_eq!(n.load(Ordering::Relaxed), 2_usize); block.next().await; assert_eq!(n.load(Ordering::Relaxed), 3_usize); block.next().await; assert_eq!(n.load(Ordering::Relaxed), 4_usize); }.fuse(); futures::pin_mut!(f2); futures::pin_mut!(metronome); block_on(async move { // futures::join!(metronome, f2) futures::select!( _ = metronome => unreachable!("Metronome never stops. qed"), _ = f2 => (), ) }); } }