mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-30 08:27:55 +00:00
a2b6470607
* Import changes from archieved repo * Revert erroneous changes * Fix more tests * Resolve discussions * Fix MORE tests * approval-voting: launch_approval better interface (#1355) --------- Co-authored-by: Javier Viola <javier@parity.io> Co-authored-by: ordian <noreply@reusable.software> Co-authored-by: ordian <write@reusable.software>
2039 lines
62 KiB
Rust
2039 lines
62 KiB
Rust
// Copyright (C) Parity Technologies (UK) Ltd.
|
|
// This file is part of Polkadot.
|
|
|
|
// Polkadot is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// Polkadot is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with Polkadot. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
//! Implements the `CandidateBackingSubsystem`.
|
|
//!
|
|
//! This subsystem maintains the entire responsibility of tracking parachain
|
|
//! candidates which can be backed, as well as the issuance of statements
|
|
//! about candidates when run on a validator node.
|
|
//!
|
|
//! There are two types of statements: `Seconded` and `Valid`.
|
|
//! `Seconded` implies `Valid`, and nothing should be stated as
|
|
//! `Valid` unless its already been `Seconded`.
|
|
//!
|
|
//! Validators may only second candidates which fall under their own group
|
|
//! assignment, and they may only second one candidate per depth per active leaf.
|
|
//! Candidates which are stated as either `Second` or `Valid` by a majority of the
|
|
//! assigned group of validators may be backed on-chain and proceed to the availability
|
|
//! stage.
|
|
//!
|
|
//! Depth is a concept relating to asynchronous backing, by which validators
|
|
//! short sub-chains of candidates are backed and extended off-chain, and then placed
|
|
//! asynchronously into blocks of the relay chain as those are authored and as the
|
|
//! relay-chain state becomes ready for them. Asynchronous backing allows parachains to
|
|
//! grow mostly independently from the state of the relay chain, which gives more time for
|
|
//! parachains to be validated and thereby increases performance.
|
|
//!
|
|
//! Most of the work of asynchronous backing is handled by the Prospective Parachains
|
|
//! subsystem. The 'depth' of a parachain block with respect to a relay chain block is
|
|
//! a measure of how many parachain blocks are between the most recent included parachain block
|
|
//! in the post-state of the relay-chain block and the candidate. For instance,
|
|
//! a candidate that descends directly from the most recent parachain block in the relay-chain
|
|
//! state has depth 0. The child of that candidate would have depth 1. And so on.
|
|
//!
|
|
//! The candidate backing subsystem keeps track of a set of 'active leaves' which are the
|
|
//! most recent blocks in the relay-chain (which is in fact a tree) which could be built
|
|
//! upon. Depth is always measured against active leaves, and the valid relay-parent that
|
|
//! each candidate can have is determined by the active leaves. The Prospective Parachains
|
|
//! subsystem enforces that the relay-parent increases monotonically, so that logic
|
|
//! is not handled here. By communicating with the Prospective Parachains subsystem,
|
|
//! this subsystem extrapolates an "implicit view" from the set of currently active leaves,
|
|
//! which determines the set of all recent relay-chain block hashes which could be relay-parents
|
|
//! for candidates backed in children of the active leaves.
|
|
//!
|
|
//! In fact, this subsystem relies on the Statement Distribution subsystem to prevent spam
|
|
//! by enforcing the rule that each validator may second at most one candidate per depth per
|
|
//! active leaf. This bounds the number of candidates that the system needs to consider and
|
|
//! is not handled within this subsystem, except for candidates seconded locally.
|
|
//!
|
|
//! This subsystem also handles relay-chain heads which don't support asynchronous backing.
|
|
//! For such active leaves, the only valid relay-parent is the leaf hash itself and the only
|
|
//! allowed depth is 0.
|
|
|
|
#![deny(unused_crate_dependencies)]
|
|
|
|
use std::{
|
|
collections::{BTreeMap, HashMap, HashSet},
|
|
sync::Arc,
|
|
};
|
|
|
|
use bitvec::vec::BitVec;
|
|
use futures::{
|
|
channel::{mpsc, oneshot},
|
|
future::BoxFuture,
|
|
stream::FuturesOrdered,
|
|
FutureExt, SinkExt, StreamExt, TryFutureExt,
|
|
};
|
|
|
|
use error::{Error, FatalResult};
|
|
use polkadot_node_primitives::{
|
|
AvailableData, InvalidCandidate, PoV, SignedFullStatementWithPVD, StatementWithPVD,
|
|
ValidationResult,
|
|
};
|
|
use polkadot_node_subsystem::{
|
|
messages::{
|
|
AvailabilityDistributionMessage, AvailabilityStoreMessage, CanSecondRequest,
|
|
CandidateBackingMessage, CandidateValidationMessage, CollatorProtocolMessage,
|
|
HypotheticalCandidate, HypotheticalFrontierRequest, IntroduceCandidateRequest,
|
|
ProspectiveParachainsMessage, ProvisionableData, ProvisionerMessage, RuntimeApiMessage,
|
|
RuntimeApiRequest, StatementDistributionMessage, StoreAvailableDataError,
|
|
},
|
|
overseer, ActiveLeavesUpdate, FromOrchestra, OverseerSignal, SpawnedSubsystem, SubsystemError,
|
|
};
|
|
use polkadot_node_subsystem_util::{
|
|
self as util,
|
|
backing_implicit_view::{FetchError as ImplicitViewFetchError, View as ImplicitView},
|
|
executor_params_at_relay_parent, request_from_runtime, request_session_index_for_child,
|
|
request_validator_groups, request_validators,
|
|
runtime::{
|
|
self, prospective_parachains_mode, request_min_backing_votes, ProspectiveParachainsMode,
|
|
},
|
|
Validator,
|
|
};
|
|
use polkadot_primitives::{
|
|
BackedCandidate, CandidateCommitments, CandidateHash, CandidateReceipt,
|
|
CommittedCandidateReceipt, CoreIndex, CoreState, ExecutorParams, Hash, Id as ParaId,
|
|
PersistedValidationData, PvfExecTimeoutKind, SigningContext, ValidationCode, ValidatorId,
|
|
ValidatorIndex, ValidatorSignature, ValidityAttestation,
|
|
};
|
|
use sp_keystore::KeystorePtr;
|
|
use statement_table::{
|
|
generic::AttestedCandidate as TableAttestedCandidate,
|
|
v2::{
|
|
SignedStatement as TableSignedStatement, Statement as TableStatement,
|
|
Summary as TableSummary,
|
|
},
|
|
Config as TableConfig, Context as TableContextTrait, Table,
|
|
};
|
|
|
|
mod error;
|
|
|
|
mod metrics;
|
|
use self::metrics::Metrics;
|
|
|
|
#[cfg(test)]
|
|
mod tests;
|
|
|
|
const LOG_TARGET: &str = "parachain::candidate-backing";
|
|
|
|
/// PoV data to validate.
|
|
enum PoVData {
|
|
/// Already available (from candidate selection).
|
|
Ready(Arc<PoV>),
|
|
/// Needs to be fetched from validator (we are checking a signed statement).
|
|
FetchFromValidator {
|
|
from_validator: ValidatorIndex,
|
|
candidate_hash: CandidateHash,
|
|
pov_hash: Hash,
|
|
},
|
|
}
|
|
|
|
enum ValidatedCandidateCommand {
|
|
// We were instructed to second the candidate that has been already validated.
|
|
Second(BackgroundValidationResult),
|
|
// We were instructed to validate the candidate.
|
|
Attest(BackgroundValidationResult),
|
|
// We were not able to `Attest` because backing validator did not send us the PoV.
|
|
AttestNoPoV(CandidateHash),
|
|
}
|
|
|
|
impl std::fmt::Debug for ValidatedCandidateCommand {
|
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
|
let candidate_hash = self.candidate_hash();
|
|
match *self {
|
|
ValidatedCandidateCommand::Second(_) => write!(f, "Second({})", candidate_hash),
|
|
ValidatedCandidateCommand::Attest(_) => write!(f, "Attest({})", candidate_hash),
|
|
ValidatedCandidateCommand::AttestNoPoV(_) => write!(f, "Attest({})", candidate_hash),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl ValidatedCandidateCommand {
|
|
fn candidate_hash(&self) -> CandidateHash {
|
|
match *self {
|
|
ValidatedCandidateCommand::Second(Ok(ref outputs)) => outputs.candidate.hash(),
|
|
ValidatedCandidateCommand::Second(Err(ref candidate)) => candidate.hash(),
|
|
ValidatedCandidateCommand::Attest(Ok(ref outputs)) => outputs.candidate.hash(),
|
|
ValidatedCandidateCommand::Attest(Err(ref candidate)) => candidate.hash(),
|
|
ValidatedCandidateCommand::AttestNoPoV(candidate_hash) => candidate_hash,
|
|
}
|
|
}
|
|
}
|
|
|
|
/// The candidate backing subsystem.
|
|
pub struct CandidateBackingSubsystem {
|
|
keystore: KeystorePtr,
|
|
metrics: Metrics,
|
|
}
|
|
|
|
impl CandidateBackingSubsystem {
|
|
/// Create a new instance of the `CandidateBackingSubsystem`.
|
|
pub fn new(keystore: KeystorePtr, metrics: Metrics) -> Self {
|
|
Self { keystore, metrics }
|
|
}
|
|
}
|
|
|
|
#[overseer::subsystem(CandidateBacking, error = SubsystemError, prefix = self::overseer)]
|
|
impl<Context> CandidateBackingSubsystem
|
|
where
|
|
Context: Send + Sync,
|
|
{
|
|
fn start(self, ctx: Context) -> SpawnedSubsystem {
|
|
let future = async move {
|
|
run(ctx, self.keystore, self.metrics)
|
|
.await
|
|
.map_err(|e| SubsystemError::with_origin("candidate-backing", e))
|
|
}
|
|
.boxed();
|
|
|
|
SpawnedSubsystem { name: "candidate-backing-subsystem", future }
|
|
}
|
|
}
|
|
|
|
struct PerRelayParentState {
|
|
prospective_parachains_mode: ProspectiveParachainsMode,
|
|
/// The hash of the relay parent on top of which this job is doing it's work.
|
|
parent: Hash,
|
|
/// The `ParaId` assigned to the local validator at this relay parent.
|
|
assignment: Option<ParaId>,
|
|
/// The candidates that are backed by enough validators in their group, by hash.
|
|
backed: HashSet<CandidateHash>,
|
|
/// The table of candidates and statements under this relay-parent.
|
|
table: Table<TableContext>,
|
|
/// The table context, including groups.
|
|
table_context: TableContext,
|
|
/// We issued `Seconded` or `Valid` statements on about these candidates.
|
|
issued_statements: HashSet<CandidateHash>,
|
|
/// These candidates are undergoing validation in the background.
|
|
awaiting_validation: HashSet<CandidateHash>,
|
|
/// Data needed for retrying in case of `ValidatedCandidateCommand::AttestNoPoV`.
|
|
fallbacks: HashMap<CandidateHash, AttestingData>,
|
|
/// The minimum backing votes threshold.
|
|
minimum_backing_votes: u32,
|
|
}
|
|
|
|
struct PerCandidateState {
|
|
persisted_validation_data: PersistedValidationData,
|
|
seconded_locally: bool,
|
|
para_id: ParaId,
|
|
relay_parent: Hash,
|
|
}
|
|
|
|
struct ActiveLeafState {
|
|
prospective_parachains_mode: ProspectiveParachainsMode,
|
|
/// The candidates seconded at various depths under this active
|
|
/// leaf with respect to parachain id. A candidate can only be
|
|
/// seconded when its hypothetical frontier under every active leaf
|
|
/// has an empty entry in this map.
|
|
///
|
|
/// When prospective parachains are disabled, the only depth
|
|
/// which is allowed is 0.
|
|
seconded_at_depth: HashMap<ParaId, BTreeMap<usize, CandidateHash>>,
|
|
}
|
|
|
|
/// The state of the subsystem.
|
|
struct State {
|
|
/// The utility for managing the implicit and explicit views in a consistent way.
|
|
///
|
|
/// We only feed leaves which have prospective parachains enabled to this view.
|
|
implicit_view: ImplicitView,
|
|
/// State tracked for all active leaves, whether or not they have prospective parachains
|
|
/// enabled.
|
|
per_leaf: HashMap<Hash, ActiveLeafState>,
|
|
/// State tracked for all relay-parents backing work is ongoing for. This includes
|
|
/// all active leaves.
|
|
///
|
|
/// relay-parents fall into one of 3 categories.
|
|
/// 1. active leaves which do support prospective parachains
|
|
/// 2. active leaves which do not support prospective parachains
|
|
/// 3. relay-chain blocks which are ancestors of an active leaf and do support prospective
|
|
/// parachains.
|
|
///
|
|
/// Relay-chain blocks which don't support prospective parachains are
|
|
/// never included in the fragment trees of active leaves which do.
|
|
///
|
|
/// While it would be technically possible to support such leaves in
|
|
/// fragment trees, it only benefits the transition period when asynchronous
|
|
/// backing is being enabled and complicates code complexity.
|
|
per_relay_parent: HashMap<Hash, PerRelayParentState>,
|
|
/// State tracked for all candidates relevant to the implicit view.
|
|
///
|
|
/// This is guaranteed to have an entry for each candidate with a relay parent in the implicit
|
|
/// or explicit view for which a `Seconded` statement has been successfully imported.
|
|
per_candidate: HashMap<CandidateHash, PerCandidateState>,
|
|
/// A cloneable sender which is dispatched to background candidate validation tasks to inform
|
|
/// the main task of the result.
|
|
background_validation_tx: mpsc::Sender<(Hash, ValidatedCandidateCommand)>,
|
|
/// The handle to the keystore used for signing.
|
|
keystore: KeystorePtr,
|
|
}
|
|
|
|
impl State {
|
|
fn new(
|
|
background_validation_tx: mpsc::Sender<(Hash, ValidatedCandidateCommand)>,
|
|
keystore: KeystorePtr,
|
|
) -> Self {
|
|
State {
|
|
implicit_view: ImplicitView::default(),
|
|
per_leaf: HashMap::default(),
|
|
per_relay_parent: HashMap::default(),
|
|
per_candidate: HashMap::new(),
|
|
background_validation_tx,
|
|
keystore,
|
|
}
|
|
}
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn run<Context>(
|
|
mut ctx: Context,
|
|
keystore: KeystorePtr,
|
|
metrics: Metrics,
|
|
) -> FatalResult<()> {
|
|
let (background_validation_tx, mut background_validation_rx) = mpsc::channel(16);
|
|
let mut state = State::new(background_validation_tx, keystore);
|
|
|
|
loop {
|
|
let res =
|
|
run_iteration(&mut ctx, &mut state, &metrics, &mut background_validation_rx).await;
|
|
|
|
match res {
|
|
Ok(()) => break,
|
|
Err(e) => crate::error::log_error(Err(e))?,
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn run_iteration<Context>(
|
|
ctx: &mut Context,
|
|
state: &mut State,
|
|
metrics: &Metrics,
|
|
background_validation_rx: &mut mpsc::Receiver<(Hash, ValidatedCandidateCommand)>,
|
|
) -> Result<(), Error> {
|
|
loop {
|
|
futures::select!(
|
|
validated_command = background_validation_rx.next().fuse() => {
|
|
if let Some((relay_parent, command)) = validated_command {
|
|
handle_validated_candidate_command(
|
|
&mut *ctx,
|
|
state,
|
|
relay_parent,
|
|
command,
|
|
metrics,
|
|
).await?;
|
|
} else {
|
|
panic!("background_validation_tx always alive at this point; qed");
|
|
}
|
|
}
|
|
from_overseer = ctx.recv().fuse() => {
|
|
match from_overseer.map_err(Error::OverseerExited)? {
|
|
FromOrchestra::Signal(OverseerSignal::ActiveLeaves(update)) => {
|
|
handle_active_leaves_update(
|
|
&mut *ctx,
|
|
update,
|
|
state,
|
|
).await?;
|
|
}
|
|
FromOrchestra::Signal(OverseerSignal::BlockFinalized(..)) => {}
|
|
FromOrchestra::Signal(OverseerSignal::Conclude) => return Ok(()),
|
|
FromOrchestra::Communication { msg } => {
|
|
handle_communication(&mut *ctx, state, msg, metrics).await?;
|
|
}
|
|
}
|
|
}
|
|
)
|
|
}
|
|
}
|
|
|
|
/// In case a backing validator does not provide a PoV, we need to retry with other backing
|
|
/// validators.
|
|
///
|
|
/// This is the data needed to accomplish this. Basically all the data needed for spawning a
|
|
/// validation job and a list of backing validators, we can try.
|
|
#[derive(Clone)]
|
|
struct AttestingData {
|
|
/// The candidate to attest.
|
|
candidate: CandidateReceipt,
|
|
/// Hash of the PoV we need to fetch.
|
|
pov_hash: Hash,
|
|
/// Validator we are currently trying to get the PoV from.
|
|
from_validator: ValidatorIndex,
|
|
/// Other backing validators we can try in case `from_validator` failed.
|
|
backing: Vec<ValidatorIndex>,
|
|
}
|
|
|
|
#[derive(Default)]
|
|
struct TableContext {
|
|
validator: Option<Validator>,
|
|
groups: HashMap<ParaId, Vec<ValidatorIndex>>,
|
|
validators: Vec<ValidatorId>,
|
|
}
|
|
|
|
impl TableContextTrait for TableContext {
|
|
type AuthorityId = ValidatorIndex;
|
|
type Digest = CandidateHash;
|
|
type GroupId = ParaId;
|
|
type Signature = ValidatorSignature;
|
|
type Candidate = CommittedCandidateReceipt;
|
|
|
|
fn candidate_digest(candidate: &CommittedCandidateReceipt) -> CandidateHash {
|
|
candidate.hash()
|
|
}
|
|
|
|
fn candidate_group(candidate: &CommittedCandidateReceipt) -> ParaId {
|
|
candidate.descriptor().para_id
|
|
}
|
|
|
|
fn is_member_of(&self, authority: &ValidatorIndex, group: &ParaId) -> bool {
|
|
self.groups.get(group).map_or(false, |g| g.iter().any(|a| a == authority))
|
|
}
|
|
|
|
fn get_group_size(&self, group: &ParaId) -> Option<usize> {
|
|
self.groups.get(group).map(|g| g.len())
|
|
}
|
|
}
|
|
|
|
// It looks like it's not possible to do an `impl From` given the current state of
|
|
// the code. So this does the necessary conversion.
|
|
fn primitive_statement_to_table(s: &SignedFullStatementWithPVD) -> TableSignedStatement {
|
|
let statement = match s.payload() {
|
|
StatementWithPVD::Seconded(c, _) => TableStatement::Seconded(c.clone()),
|
|
StatementWithPVD::Valid(h) => TableStatement::Valid(*h),
|
|
};
|
|
|
|
TableSignedStatement {
|
|
statement,
|
|
signature: s.signature().clone(),
|
|
sender: s.validator_index(),
|
|
}
|
|
}
|
|
|
|
fn table_attested_to_backed(
|
|
attested: TableAttestedCandidate<
|
|
ParaId,
|
|
CommittedCandidateReceipt,
|
|
ValidatorIndex,
|
|
ValidatorSignature,
|
|
>,
|
|
table_context: &TableContext,
|
|
) -> Option<BackedCandidate> {
|
|
let TableAttestedCandidate { candidate, validity_votes, group_id: para_id } = attested;
|
|
|
|
let (ids, validity_votes): (Vec<_>, Vec<ValidityAttestation>) =
|
|
validity_votes.into_iter().map(|(id, vote)| (id, vote.into())).unzip();
|
|
|
|
let group = table_context.groups.get(¶_id)?;
|
|
|
|
let mut validator_indices = BitVec::with_capacity(group.len());
|
|
|
|
validator_indices.resize(group.len(), false);
|
|
|
|
// The order of the validity votes in the backed candidate must match
|
|
// the order of bits set in the bitfield, which is not necessarily
|
|
// the order of the `validity_votes` we got from the table.
|
|
let mut vote_positions = Vec::with_capacity(validity_votes.len());
|
|
for (orig_idx, id) in ids.iter().enumerate() {
|
|
if let Some(position) = group.iter().position(|x| x == id) {
|
|
validator_indices.set(position, true);
|
|
vote_positions.push((orig_idx, position));
|
|
} else {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
"Logic error: Validity vote from table does not correspond to group",
|
|
);
|
|
|
|
return None
|
|
}
|
|
}
|
|
vote_positions.sort_by_key(|(_orig, pos_in_group)| *pos_in_group);
|
|
|
|
Some(BackedCandidate {
|
|
candidate,
|
|
validity_votes: vote_positions
|
|
.into_iter()
|
|
.map(|(pos_in_votes, _pos_in_group)| validity_votes[pos_in_votes].clone())
|
|
.collect(),
|
|
validator_indices,
|
|
})
|
|
}
|
|
|
|
async fn store_available_data(
|
|
sender: &mut impl overseer::CandidateBackingSenderTrait,
|
|
n_validators: u32,
|
|
candidate_hash: CandidateHash,
|
|
available_data: AvailableData,
|
|
expected_erasure_root: Hash,
|
|
) -> Result<(), Error> {
|
|
let (tx, rx) = oneshot::channel();
|
|
// Important: the `av-store` subsystem will check if the erasure root of the `available_data`
|
|
// matches `expected_erasure_root` which was provided by the collator in the `CandidateReceipt`.
|
|
// This check is consensus critical and the `backing` subsystem relies on it for ensuring
|
|
// candidate validity.
|
|
sender
|
|
.send_message(AvailabilityStoreMessage::StoreAvailableData {
|
|
candidate_hash,
|
|
n_validators,
|
|
available_data,
|
|
expected_erasure_root,
|
|
tx,
|
|
})
|
|
.await;
|
|
|
|
rx.await
|
|
.map_err(Error::StoreAvailableDataChannel)?
|
|
.map_err(Error::StoreAvailableData)
|
|
}
|
|
|
|
// Make a `PoV` available.
|
|
//
|
|
// This calls the AV store to write the available data to storage. The AV store also checks the
|
|
// erasure root matches the `expected_erasure_root`.
|
|
// This returns `Err()` on erasure root mismatch or due to any AV store subsystem error.
|
|
//
|
|
// Otherwise, it returns `Ok(())`.
|
|
async fn make_pov_available(
|
|
sender: &mut impl overseer::CandidateBackingSenderTrait,
|
|
n_validators: usize,
|
|
pov: Arc<PoV>,
|
|
candidate_hash: CandidateHash,
|
|
validation_data: PersistedValidationData,
|
|
expected_erasure_root: Hash,
|
|
) -> Result<(), Error> {
|
|
store_available_data(
|
|
sender,
|
|
n_validators as u32,
|
|
candidate_hash,
|
|
AvailableData { pov, validation_data },
|
|
expected_erasure_root,
|
|
)
|
|
.await
|
|
}
|
|
|
|
async fn request_pov(
|
|
sender: &mut impl overseer::CandidateBackingSenderTrait,
|
|
relay_parent: Hash,
|
|
from_validator: ValidatorIndex,
|
|
para_id: ParaId,
|
|
candidate_hash: CandidateHash,
|
|
pov_hash: Hash,
|
|
) -> Result<Arc<PoV>, Error> {
|
|
let (tx, rx) = oneshot::channel();
|
|
sender
|
|
.send_message(AvailabilityDistributionMessage::FetchPoV {
|
|
relay_parent,
|
|
from_validator,
|
|
para_id,
|
|
candidate_hash,
|
|
pov_hash,
|
|
tx,
|
|
})
|
|
.await;
|
|
|
|
let pov = rx.await.map_err(|_| Error::FetchPoV)?;
|
|
Ok(Arc::new(pov))
|
|
}
|
|
|
|
async fn request_candidate_validation(
|
|
sender: &mut impl overseer::CandidateBackingSenderTrait,
|
|
pvd: PersistedValidationData,
|
|
code: ValidationCode,
|
|
candidate_receipt: CandidateReceipt,
|
|
pov: Arc<PoV>,
|
|
executor_params: ExecutorParams,
|
|
) -> Result<ValidationResult, Error> {
|
|
let (tx, rx) = oneshot::channel();
|
|
|
|
sender
|
|
.send_message(CandidateValidationMessage::ValidateFromExhaustive(
|
|
pvd,
|
|
code,
|
|
candidate_receipt,
|
|
pov,
|
|
executor_params,
|
|
PvfExecTimeoutKind::Backing,
|
|
tx,
|
|
))
|
|
.await;
|
|
|
|
match rx.await {
|
|
Ok(Ok(validation_result)) => Ok(validation_result),
|
|
Ok(Err(err)) => Err(Error::ValidationFailed(err)),
|
|
Err(err) => Err(Error::ValidateFromExhaustive(err)),
|
|
}
|
|
}
|
|
|
|
struct BackgroundValidationOutputs {
|
|
candidate: CandidateReceipt,
|
|
commitments: CandidateCommitments,
|
|
persisted_validation_data: PersistedValidationData,
|
|
}
|
|
|
|
type BackgroundValidationResult = Result<BackgroundValidationOutputs, CandidateReceipt>;
|
|
|
|
struct BackgroundValidationParams<S: overseer::CandidateBackingSenderTrait, F> {
|
|
sender: S,
|
|
tx_command: mpsc::Sender<(Hash, ValidatedCandidateCommand)>,
|
|
candidate: CandidateReceipt,
|
|
relay_parent: Hash,
|
|
persisted_validation_data: PersistedValidationData,
|
|
pov: PoVData,
|
|
n_validators: usize,
|
|
make_command: F,
|
|
}
|
|
|
|
async fn validate_and_make_available(
|
|
params: BackgroundValidationParams<
|
|
impl overseer::CandidateBackingSenderTrait,
|
|
impl Fn(BackgroundValidationResult) -> ValidatedCandidateCommand + Sync,
|
|
>,
|
|
) -> Result<(), Error> {
|
|
let BackgroundValidationParams {
|
|
mut sender,
|
|
mut tx_command,
|
|
candidate,
|
|
relay_parent,
|
|
persisted_validation_data,
|
|
pov,
|
|
n_validators,
|
|
make_command,
|
|
} = params;
|
|
|
|
let validation_code = {
|
|
let validation_code_hash = candidate.descriptor().validation_code_hash;
|
|
let (tx, rx) = oneshot::channel();
|
|
sender
|
|
.send_message(RuntimeApiMessage::Request(
|
|
relay_parent,
|
|
RuntimeApiRequest::ValidationCodeByHash(validation_code_hash, tx),
|
|
))
|
|
.await;
|
|
|
|
let code = rx.await.map_err(Error::RuntimeApiUnavailable)?;
|
|
match code {
|
|
Err(e) => return Err(Error::FetchValidationCode(validation_code_hash, e)),
|
|
Ok(None) => return Err(Error::NoValidationCode(validation_code_hash)),
|
|
Ok(Some(c)) => c,
|
|
}
|
|
};
|
|
|
|
let executor_params = match executor_params_at_relay_parent(relay_parent, &mut sender).await {
|
|
Ok(ep) => ep,
|
|
Err(e) => return Err(Error::UtilError(e)),
|
|
};
|
|
|
|
let pov = match pov {
|
|
PoVData::Ready(pov) => pov,
|
|
PoVData::FetchFromValidator { from_validator, candidate_hash, pov_hash } =>
|
|
match request_pov(
|
|
&mut sender,
|
|
relay_parent,
|
|
from_validator,
|
|
candidate.descriptor.para_id,
|
|
candidate_hash,
|
|
pov_hash,
|
|
)
|
|
.await
|
|
{
|
|
Err(Error::FetchPoV) => {
|
|
tx_command
|
|
.send((
|
|
relay_parent,
|
|
ValidatedCandidateCommand::AttestNoPoV(candidate.hash()),
|
|
))
|
|
.await
|
|
.map_err(Error::BackgroundValidationMpsc)?;
|
|
return Ok(())
|
|
},
|
|
Err(err) => return Err(err),
|
|
Ok(pov) => pov,
|
|
},
|
|
};
|
|
|
|
let v = {
|
|
request_candidate_validation(
|
|
&mut sender,
|
|
persisted_validation_data,
|
|
validation_code,
|
|
candidate.clone(),
|
|
pov.clone(),
|
|
executor_params,
|
|
)
|
|
.await?
|
|
};
|
|
|
|
let res = match v {
|
|
ValidationResult::Valid(commitments, validation_data) => {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate.hash(),
|
|
"Validation successful",
|
|
);
|
|
|
|
let erasure_valid = make_pov_available(
|
|
&mut sender,
|
|
n_validators,
|
|
pov.clone(),
|
|
candidate.hash(),
|
|
validation_data.clone(),
|
|
candidate.descriptor.erasure_root,
|
|
)
|
|
.await;
|
|
|
|
match erasure_valid {
|
|
Ok(()) => Ok(BackgroundValidationOutputs {
|
|
candidate,
|
|
commitments,
|
|
persisted_validation_data: validation_data,
|
|
}),
|
|
Err(Error::StoreAvailableData(StoreAvailableDataError::InvalidErasureRoot)) => {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate.hash(),
|
|
actual_commitments = ?commitments,
|
|
"Erasure root doesn't match the announced by the candidate receipt",
|
|
);
|
|
Err(candidate)
|
|
},
|
|
// Bubble up any other error.
|
|
Err(e) => return Err(e),
|
|
}
|
|
},
|
|
ValidationResult::Invalid(InvalidCandidate::CommitmentsHashMismatch) => {
|
|
// If validation produces a new set of commitments, we vote the candidate as invalid.
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate.hash(),
|
|
"Validation yielded different commitments",
|
|
);
|
|
Err(candidate)
|
|
},
|
|
ValidationResult::Invalid(reason) => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate.hash(),
|
|
reason = ?reason,
|
|
"Validation yielded an invalid candidate",
|
|
);
|
|
Err(candidate)
|
|
},
|
|
};
|
|
|
|
tx_command.send((relay_parent, make_command(res))).await.map_err(Into::into)
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn handle_communication<Context>(
|
|
ctx: &mut Context,
|
|
state: &mut State,
|
|
message: CandidateBackingMessage,
|
|
metrics: &Metrics,
|
|
) -> Result<(), Error> {
|
|
match message {
|
|
CandidateBackingMessage::Second(_relay_parent, candidate, pvd, pov) => {
|
|
handle_second_message(ctx, state, candidate, pvd, pov, metrics).await?;
|
|
},
|
|
CandidateBackingMessage::Statement(relay_parent, statement) => {
|
|
handle_statement_message(ctx, state, relay_parent, statement, metrics).await?;
|
|
},
|
|
CandidateBackingMessage::GetBackedCandidates(requested_candidates, tx) =>
|
|
handle_get_backed_candidates_message(state, requested_candidates, tx, metrics)?,
|
|
CandidateBackingMessage::CanSecond(request, tx) =>
|
|
handle_can_second_request(ctx, state, request, tx).await,
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn handle_active_leaves_update<Context>(
|
|
ctx: &mut Context,
|
|
update: ActiveLeavesUpdate,
|
|
state: &mut State,
|
|
) -> Result<(), Error> {
|
|
enum LeafHasProspectiveParachains {
|
|
Enabled(Result<ProspectiveParachainsMode, ImplicitViewFetchError>),
|
|
Disabled,
|
|
}
|
|
|
|
// Activate in implicit view before deactivate, per the docs
|
|
// on ImplicitView, this is more efficient.
|
|
let res = if let Some(leaf) = update.activated {
|
|
// Only activate in implicit view if prospective
|
|
// parachains are enabled.
|
|
let mode = prospective_parachains_mode(ctx.sender(), leaf.hash).await?;
|
|
|
|
let leaf_hash = leaf.hash;
|
|
Some((
|
|
leaf,
|
|
match mode {
|
|
ProspectiveParachainsMode::Disabled => LeafHasProspectiveParachains::Disabled,
|
|
ProspectiveParachainsMode::Enabled { .. } => LeafHasProspectiveParachains::Enabled(
|
|
state.implicit_view.activate_leaf(ctx.sender(), leaf_hash).await.map(|_| mode),
|
|
),
|
|
},
|
|
))
|
|
} else {
|
|
None
|
|
};
|
|
|
|
for deactivated in update.deactivated {
|
|
state.per_leaf.remove(&deactivated);
|
|
state.implicit_view.deactivate_leaf(deactivated);
|
|
}
|
|
|
|
// clean up `per_relay_parent` according to ancestry
|
|
// of leaves. we do this so we can clean up candidates right after
|
|
// as a result.
|
|
//
|
|
// when prospective parachains are disabled, the implicit view is empty,
|
|
// which means we'll clean up everything that's not a leaf - the expected behavior
|
|
// for pre-asynchronous backing.
|
|
{
|
|
let remaining: HashSet<_> = state
|
|
.per_leaf
|
|
.keys()
|
|
.chain(state.implicit_view.all_allowed_relay_parents())
|
|
.collect();
|
|
|
|
state.per_relay_parent.retain(|r, _| remaining.contains(&r));
|
|
}
|
|
|
|
// clean up `per_candidate` according to which relay-parents
|
|
// are known.
|
|
//
|
|
// when prospective parachains are disabled, we clean up all candidates
|
|
// because we've cleaned up all relay parents. this is correct.
|
|
state
|
|
.per_candidate
|
|
.retain(|_, pc| state.per_relay_parent.contains_key(&pc.relay_parent));
|
|
|
|
// Get relay parents which might be fresh but might be known already
|
|
// that are explicit or implicit from the new active leaf.
|
|
let (fresh_relay_parents, leaf_mode) = match res {
|
|
None => return Ok(()),
|
|
Some((leaf, LeafHasProspectiveParachains::Disabled)) => {
|
|
// defensive in this case - for enabled, this manifests as an error.
|
|
if state.per_leaf.contains_key(&leaf.hash) {
|
|
return Ok(())
|
|
}
|
|
|
|
state.per_leaf.insert(
|
|
leaf.hash,
|
|
ActiveLeafState {
|
|
prospective_parachains_mode: ProspectiveParachainsMode::Disabled,
|
|
// This is empty because the only allowed relay-parent and depth
|
|
// when prospective parachains are disabled is the leaf hash and 0,
|
|
// respectively. We've just learned about the leaf hash, so we cannot
|
|
// have any candidates seconded with it as a relay-parent yet.
|
|
seconded_at_depth: HashMap::new(),
|
|
},
|
|
);
|
|
|
|
(vec![leaf.hash], ProspectiveParachainsMode::Disabled)
|
|
},
|
|
Some((leaf, LeafHasProspectiveParachains::Enabled(Ok(prospective_parachains_mode)))) => {
|
|
let fresh_relay_parents =
|
|
state.implicit_view.known_allowed_relay_parents_under(&leaf.hash, None);
|
|
|
|
// At this point, all candidates outside of the implicit view
|
|
// have been cleaned up. For all which remain, which we've seconded,
|
|
// we ask the prospective parachains subsystem where they land in the fragment
|
|
// tree for the given active leaf. This comprises our `seconded_at_depth`.
|
|
|
|
let remaining_seconded = state
|
|
.per_candidate
|
|
.iter()
|
|
.filter(|(_, cd)| cd.seconded_locally)
|
|
.map(|(c_hash, cd)| (*c_hash, cd.para_id));
|
|
|
|
// one-to-one correspondence to remaining_seconded
|
|
let mut membership_answers = FuturesOrdered::new();
|
|
|
|
for (candidate_hash, para_id) in remaining_seconded {
|
|
let (tx, rx) = oneshot::channel();
|
|
membership_answers
|
|
.push_back(rx.map_ok(move |membership| (para_id, candidate_hash, membership)));
|
|
|
|
ctx.send_message(ProspectiveParachainsMessage::GetTreeMembership(
|
|
para_id,
|
|
candidate_hash,
|
|
tx,
|
|
))
|
|
.await;
|
|
}
|
|
|
|
let mut seconded_at_depth = HashMap::new();
|
|
if let Some(response) = membership_answers.next().await {
|
|
match response {
|
|
Err(oneshot::Canceled) => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
"Prospective parachains subsystem unreachable for membership request",
|
|
);
|
|
},
|
|
Ok((para_id, candidate_hash, membership)) => {
|
|
// This request gives membership in all fragment trees. We have some
|
|
// wasted data here, and it can be optimized if it proves
|
|
// relevant to performance.
|
|
if let Some((_, depths)) =
|
|
membership.into_iter().find(|(leaf_hash, _)| leaf_hash == &leaf.hash)
|
|
{
|
|
let para_entry: &mut BTreeMap<usize, CandidateHash> =
|
|
seconded_at_depth.entry(para_id).or_default();
|
|
for depth in depths {
|
|
para_entry.insert(depth, candidate_hash);
|
|
}
|
|
}
|
|
},
|
|
}
|
|
}
|
|
|
|
state.per_leaf.insert(
|
|
leaf.hash,
|
|
ActiveLeafState { prospective_parachains_mode, seconded_at_depth },
|
|
);
|
|
|
|
let fresh_relay_parent = match fresh_relay_parents {
|
|
Some(f) => f.to_vec(),
|
|
None => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
leaf_hash = ?leaf.hash,
|
|
"Implicit view gave no relay-parents"
|
|
);
|
|
|
|
vec![leaf.hash]
|
|
},
|
|
};
|
|
(fresh_relay_parent, prospective_parachains_mode)
|
|
},
|
|
Some((leaf, LeafHasProspectiveParachains::Enabled(Err(e)))) => {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
leaf_hash = ?leaf.hash,
|
|
err = ?e,
|
|
"Failed to load implicit view for leaf."
|
|
);
|
|
|
|
return Ok(())
|
|
},
|
|
};
|
|
|
|
// add entries in `per_relay_parent`. for all new relay-parents.
|
|
for maybe_new in fresh_relay_parents {
|
|
if state.per_relay_parent.contains_key(&maybe_new) {
|
|
continue
|
|
}
|
|
|
|
let mode = match state.per_leaf.get(&maybe_new) {
|
|
None => {
|
|
// If the relay-parent isn't a leaf itself,
|
|
// then it is guaranteed by the prospective parachains
|
|
// subsystem that it is an ancestor of a leaf which
|
|
// has prospective parachains enabled and that the
|
|
// block itself did.
|
|
leaf_mode
|
|
},
|
|
Some(l) => l.prospective_parachains_mode,
|
|
};
|
|
|
|
// construct a `PerRelayParent` from the runtime API
|
|
// and insert it.
|
|
let per = construct_per_relay_parent_state(ctx, maybe_new, &state.keystore, mode).await?;
|
|
|
|
if let Some(per) = per {
|
|
state.per_relay_parent.insert(maybe_new, per);
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Load the data necessary to do backing work on top of a relay-parent.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn construct_per_relay_parent_state<Context>(
|
|
ctx: &mut Context,
|
|
relay_parent: Hash,
|
|
keystore: &KeystorePtr,
|
|
mode: ProspectiveParachainsMode,
|
|
) -> Result<Option<PerRelayParentState>, Error> {
|
|
macro_rules! try_runtime_api {
|
|
($x: expr) => {
|
|
match $x {
|
|
Ok(x) => x,
|
|
Err(err) => {
|
|
// Only bubble up fatal errors.
|
|
error::log_error(Err(Into::<runtime::Error>::into(err).into()))?;
|
|
|
|
// We can't do candidate validation work if we don't have the
|
|
// requisite runtime API data. But these errors should not take
|
|
// down the node.
|
|
return Ok(None)
|
|
},
|
|
}
|
|
};
|
|
}
|
|
|
|
let parent = relay_parent;
|
|
|
|
let (session_index, validators, groups, cores) = futures::try_join!(
|
|
request_session_index_for_child(parent, ctx.sender()).await,
|
|
request_validators(parent, ctx.sender()).await,
|
|
request_validator_groups(parent, ctx.sender()).await,
|
|
request_from_runtime(parent, ctx.sender(), |tx| {
|
|
RuntimeApiRequest::AvailabilityCores(tx)
|
|
},)
|
|
.await,
|
|
)
|
|
.map_err(Error::JoinMultiple)?;
|
|
|
|
let session_index = try_runtime_api!(session_index);
|
|
let validators: Vec<_> = try_runtime_api!(validators);
|
|
let (validator_groups, group_rotation_info) = try_runtime_api!(groups);
|
|
let cores = try_runtime_api!(cores);
|
|
let minimum_backing_votes =
|
|
try_runtime_api!(request_min_backing_votes(parent, session_index, ctx.sender()).await);
|
|
|
|
let signing_context = SigningContext { parent_hash: parent, session_index };
|
|
let validator =
|
|
match Validator::construct(&validators, signing_context.clone(), keystore.clone()) {
|
|
Ok(v) => Some(v),
|
|
Err(util::Error::NotAValidator) => None,
|
|
Err(e) => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
err = ?e,
|
|
"Cannot participate in candidate backing",
|
|
);
|
|
|
|
return Ok(None)
|
|
},
|
|
};
|
|
|
|
let mut groups = HashMap::new();
|
|
let n_cores = cores.len();
|
|
let mut assignment = None;
|
|
|
|
for (idx, core) in cores.into_iter().enumerate() {
|
|
let core_para_id = match core {
|
|
CoreState::Scheduled(scheduled) => scheduled.para_id,
|
|
CoreState::Occupied(occupied) =>
|
|
if mode.is_enabled() {
|
|
// Async backing makes it legal to build on top of
|
|
// occupied core.
|
|
occupied.candidate_descriptor.para_id
|
|
} else {
|
|
continue
|
|
},
|
|
CoreState::Free => continue,
|
|
};
|
|
|
|
let core_index = CoreIndex(idx as _);
|
|
let group_index = group_rotation_info.group_for_core(core_index, n_cores);
|
|
if let Some(g) = validator_groups.get(group_index.0 as usize) {
|
|
if validator.as_ref().map_or(false, |v| g.contains(&v.index())) {
|
|
assignment = Some(core_para_id);
|
|
}
|
|
groups.insert(core_para_id, g.clone());
|
|
}
|
|
}
|
|
|
|
let table_context = TableContext { groups, validators, validator };
|
|
let table_config = TableConfig {
|
|
allow_multiple_seconded: match mode {
|
|
ProspectiveParachainsMode::Enabled { .. } => true,
|
|
ProspectiveParachainsMode::Disabled => false,
|
|
},
|
|
};
|
|
|
|
Ok(Some(PerRelayParentState {
|
|
prospective_parachains_mode: mode,
|
|
parent,
|
|
assignment,
|
|
backed: HashSet::new(),
|
|
table: Table::new(table_config),
|
|
table_context,
|
|
issued_statements: HashSet::new(),
|
|
awaiting_validation: HashSet::new(),
|
|
fallbacks: HashMap::new(),
|
|
minimum_backing_votes,
|
|
}))
|
|
}
|
|
|
|
enum SecondingAllowed {
|
|
No,
|
|
Yes(Vec<(Hash, Vec<usize>)>),
|
|
}
|
|
|
|
/// Checks whether a candidate can be seconded based on its hypothetical frontiers in the fragment
|
|
/// tree and what we've already seconded in all active leaves.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn seconding_sanity_check<Context>(
|
|
ctx: &mut Context,
|
|
active_leaves: &HashMap<Hash, ActiveLeafState>,
|
|
implicit_view: &ImplicitView,
|
|
hypothetical_candidate: HypotheticalCandidate,
|
|
backed_in_path_only: bool,
|
|
) -> SecondingAllowed {
|
|
let mut membership = Vec::new();
|
|
let mut responses = FuturesOrdered::<BoxFuture<'_, Result<_, oneshot::Canceled>>>::new();
|
|
|
|
let candidate_para = hypothetical_candidate.candidate_para();
|
|
let candidate_relay_parent = hypothetical_candidate.relay_parent();
|
|
let candidate_hash = hypothetical_candidate.candidate_hash();
|
|
|
|
for (head, leaf_state) in active_leaves {
|
|
if leaf_state.prospective_parachains_mode.is_enabled() {
|
|
// Check that the candidate relay parent is allowed for para, skip the
|
|
// leaf otherwise.
|
|
let allowed_parents_for_para =
|
|
implicit_view.known_allowed_relay_parents_under(head, Some(candidate_para));
|
|
if !allowed_parents_for_para.unwrap_or_default().contains(&candidate_relay_parent) {
|
|
continue
|
|
}
|
|
|
|
let (tx, rx) = oneshot::channel();
|
|
ctx.send_message(ProspectiveParachainsMessage::GetHypotheticalFrontier(
|
|
HypotheticalFrontierRequest {
|
|
candidates: vec![hypothetical_candidate.clone()],
|
|
fragment_tree_relay_parent: Some(*head),
|
|
backed_in_path_only,
|
|
},
|
|
tx,
|
|
))
|
|
.await;
|
|
let response = rx.map_ok(move |frontiers| {
|
|
let depths: Vec<usize> = frontiers
|
|
.into_iter()
|
|
.flat_map(|(candidate, memberships)| {
|
|
debug_assert_eq!(candidate.candidate_hash(), candidate_hash);
|
|
memberships.into_iter().flat_map(|(relay_parent, depths)| {
|
|
debug_assert_eq!(relay_parent, *head);
|
|
depths
|
|
})
|
|
})
|
|
.collect();
|
|
(depths, head, leaf_state)
|
|
});
|
|
responses.push_back(response.boxed());
|
|
} else {
|
|
if *head == candidate_relay_parent {
|
|
if leaf_state
|
|
.seconded_at_depth
|
|
.get(&candidate_para)
|
|
.map_or(false, |occupied| occupied.contains_key(&0))
|
|
{
|
|
// The leaf is already occupied.
|
|
return SecondingAllowed::No
|
|
}
|
|
responses.push_back(futures::future::ok((vec![0], head, leaf_state)).boxed());
|
|
}
|
|
}
|
|
}
|
|
|
|
if responses.is_empty() {
|
|
return SecondingAllowed::No
|
|
}
|
|
|
|
while let Some(response) = responses.next().await {
|
|
match response {
|
|
Err(oneshot::Canceled) => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
"Failed to reach prospective parachains subsystem for hypothetical frontiers",
|
|
);
|
|
|
|
return SecondingAllowed::No
|
|
},
|
|
Ok((depths, head, leaf_state)) => {
|
|
for depth in &depths {
|
|
if leaf_state
|
|
.seconded_at_depth
|
|
.get(&candidate_para)
|
|
.map_or(false, |occupied| occupied.contains_key(&depth))
|
|
{
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
?candidate_hash,
|
|
depth,
|
|
leaf_hash = ?head,
|
|
"Refusing to second candidate at depth - already occupied."
|
|
);
|
|
|
|
return SecondingAllowed::No
|
|
}
|
|
}
|
|
|
|
membership.push((*head, depths));
|
|
},
|
|
}
|
|
}
|
|
|
|
// At this point we've checked the depths of the candidate against all active
|
|
// leaves.
|
|
SecondingAllowed::Yes(membership)
|
|
}
|
|
|
|
/// Performs seconding sanity check for an advertisement.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn handle_can_second_request<Context>(
|
|
ctx: &mut Context,
|
|
state: &State,
|
|
request: CanSecondRequest,
|
|
tx: oneshot::Sender<bool>,
|
|
) {
|
|
let relay_parent = request.candidate_relay_parent;
|
|
let response = if state
|
|
.per_relay_parent
|
|
.get(&relay_parent)
|
|
.map_or(false, |pr_state| pr_state.prospective_parachains_mode.is_enabled())
|
|
{
|
|
let hypothetical_candidate = HypotheticalCandidate::Incomplete {
|
|
candidate_hash: request.candidate_hash,
|
|
candidate_para: request.candidate_para_id,
|
|
parent_head_data_hash: request.parent_head_data_hash,
|
|
candidate_relay_parent: relay_parent,
|
|
};
|
|
|
|
let result = seconding_sanity_check(
|
|
ctx,
|
|
&state.per_leaf,
|
|
&state.implicit_view,
|
|
hypothetical_candidate,
|
|
true,
|
|
)
|
|
.await;
|
|
|
|
match result {
|
|
SecondingAllowed::No => false,
|
|
SecondingAllowed::Yes(membership) => {
|
|
// Candidate should be recognized by at least some fragment tree.
|
|
membership.iter().any(|(_, m)| !m.is_empty())
|
|
},
|
|
}
|
|
} else {
|
|
// Relay parent is unknown or async backing is disabled.
|
|
false
|
|
};
|
|
|
|
let _ = tx.send(response);
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn handle_validated_candidate_command<Context>(
|
|
ctx: &mut Context,
|
|
state: &mut State,
|
|
relay_parent: Hash,
|
|
command: ValidatedCandidateCommand,
|
|
metrics: &Metrics,
|
|
) -> Result<(), Error> {
|
|
match state.per_relay_parent.get_mut(&relay_parent) {
|
|
Some(rp_state) => {
|
|
let candidate_hash = command.candidate_hash();
|
|
rp_state.awaiting_validation.remove(&candidate_hash);
|
|
|
|
match command {
|
|
ValidatedCandidateCommand::Second(res) => match res {
|
|
Ok(outputs) => {
|
|
let BackgroundValidationOutputs {
|
|
candidate,
|
|
commitments,
|
|
persisted_validation_data,
|
|
} = outputs;
|
|
|
|
if rp_state.issued_statements.contains(&candidate_hash) {
|
|
return Ok(())
|
|
}
|
|
|
|
let receipt = CommittedCandidateReceipt {
|
|
descriptor: candidate.descriptor.clone(),
|
|
commitments,
|
|
};
|
|
|
|
let parent_head_data_hash = persisted_validation_data.parent_head.hash();
|
|
// Note that `GetHypotheticalFrontier` doesn't account for recursion,
|
|
// i.e. candidates can appear at multiple depths in the tree and in fact
|
|
// at all depths, and we don't know what depths a candidate will ultimately
|
|
// occupy because that's dependent on other candidates we haven't yet
|
|
// received.
|
|
//
|
|
// The only way to effectively rule this out is to have candidate receipts
|
|
// directly commit to the parachain block number or some other incrementing
|
|
// counter. That requires a major primitives format upgrade, so for now
|
|
// we just rule out trivial cycles.
|
|
if parent_head_data_hash == receipt.commitments.head_data.hash() {
|
|
return Ok(())
|
|
}
|
|
let hypothetical_candidate = HypotheticalCandidate::Complete {
|
|
candidate_hash,
|
|
receipt: Arc::new(receipt.clone()),
|
|
persisted_validation_data: persisted_validation_data.clone(),
|
|
};
|
|
// sanity check that we're allowed to second the candidate
|
|
// and that it doesn't conflict with other candidates we've
|
|
// seconded.
|
|
let fragment_tree_membership = match seconding_sanity_check(
|
|
ctx,
|
|
&state.per_leaf,
|
|
&state.implicit_view,
|
|
hypothetical_candidate,
|
|
false,
|
|
)
|
|
.await
|
|
{
|
|
SecondingAllowed::No => return Ok(()),
|
|
SecondingAllowed::Yes(membership) => membership,
|
|
};
|
|
|
|
let statement =
|
|
StatementWithPVD::Seconded(receipt, persisted_validation_data);
|
|
|
|
// If we get an Error::RejectedByProspectiveParachains,
|
|
// then the statement has not been distributed or imported into
|
|
// the table.
|
|
let res = sign_import_and_distribute_statement(
|
|
ctx,
|
|
rp_state,
|
|
&mut state.per_candidate,
|
|
statement,
|
|
state.keystore.clone(),
|
|
metrics,
|
|
)
|
|
.await;
|
|
|
|
if let Err(Error::RejectedByProspectiveParachains) = res {
|
|
let candidate_hash = candidate.hash();
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
relay_parent = ?candidate.descriptor().relay_parent,
|
|
?candidate_hash,
|
|
"Attempted to second candidate but was rejected by prospective parachains",
|
|
);
|
|
|
|
// Ensure the collator is reported.
|
|
ctx.send_message(CollatorProtocolMessage::Invalid(
|
|
candidate.descriptor().relay_parent,
|
|
candidate,
|
|
))
|
|
.await;
|
|
|
|
return Ok(())
|
|
}
|
|
|
|
if let Some(stmt) = res? {
|
|
match state.per_candidate.get_mut(&candidate_hash) {
|
|
None => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
?candidate_hash,
|
|
"Missing `per_candidate` for seconded candidate.",
|
|
);
|
|
},
|
|
Some(p) => p.seconded_locally = true,
|
|
}
|
|
|
|
// update seconded depths in active leaves.
|
|
for (leaf, depths) in fragment_tree_membership {
|
|
let leaf_data = match state.per_leaf.get_mut(&leaf) {
|
|
None => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
leaf_hash = ?leaf,
|
|
"Missing `per_leaf` for known active leaf."
|
|
);
|
|
|
|
continue
|
|
},
|
|
Some(d) => d,
|
|
};
|
|
|
|
let seconded_at_depth = leaf_data
|
|
.seconded_at_depth
|
|
.entry(candidate.descriptor().para_id)
|
|
.or_default();
|
|
|
|
for depth in depths {
|
|
seconded_at_depth.insert(depth, candidate_hash);
|
|
}
|
|
}
|
|
|
|
rp_state.issued_statements.insert(candidate_hash);
|
|
|
|
metrics.on_candidate_seconded();
|
|
ctx.send_message(CollatorProtocolMessage::Seconded(
|
|
rp_state.parent,
|
|
StatementWithPVD::drop_pvd_from_signed(stmt),
|
|
))
|
|
.await;
|
|
}
|
|
},
|
|
Err(candidate) => {
|
|
ctx.send_message(CollatorProtocolMessage::Invalid(
|
|
rp_state.parent,
|
|
candidate,
|
|
))
|
|
.await;
|
|
},
|
|
},
|
|
ValidatedCandidateCommand::Attest(res) => {
|
|
// We are done - avoid new validation spawns:
|
|
rp_state.fallbacks.remove(&candidate_hash);
|
|
// sanity check.
|
|
if !rp_state.issued_statements.contains(&candidate_hash) {
|
|
if res.is_ok() {
|
|
let statement = StatementWithPVD::Valid(candidate_hash);
|
|
|
|
sign_import_and_distribute_statement(
|
|
ctx,
|
|
rp_state,
|
|
&mut state.per_candidate,
|
|
statement,
|
|
state.keystore.clone(),
|
|
metrics,
|
|
)
|
|
.await?;
|
|
}
|
|
rp_state.issued_statements.insert(candidate_hash);
|
|
}
|
|
},
|
|
ValidatedCandidateCommand::AttestNoPoV(candidate_hash) => {
|
|
if let Some(attesting) = rp_state.fallbacks.get_mut(&candidate_hash) {
|
|
if let Some(index) = attesting.backing.pop() {
|
|
attesting.from_validator = index;
|
|
let attesting = attesting.clone();
|
|
|
|
// The candidate state should be available because we've
|
|
// validated it before, the relay-parent is still around,
|
|
// and candidates are pruned on the basis of relay-parents.
|
|
//
|
|
// If it's not, then no point in validating it anyway.
|
|
if let Some(pvd) = state
|
|
.per_candidate
|
|
.get(&candidate_hash)
|
|
.map(|pc| pc.persisted_validation_data.clone())
|
|
{
|
|
kick_off_validation_work(
|
|
ctx,
|
|
rp_state,
|
|
pvd,
|
|
&state.background_validation_tx,
|
|
attesting,
|
|
)
|
|
.await?;
|
|
}
|
|
}
|
|
} else {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
"AttestNoPoV was triggered without fallback being available."
|
|
);
|
|
debug_assert!(false);
|
|
}
|
|
},
|
|
}
|
|
},
|
|
None => {
|
|
// simple race condition; can be ignored = this relay-parent
|
|
// is no longer relevant.
|
|
},
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
fn sign_statement(
|
|
rp_state: &PerRelayParentState,
|
|
statement: StatementWithPVD,
|
|
keystore: KeystorePtr,
|
|
metrics: &Metrics,
|
|
) -> Option<SignedFullStatementWithPVD> {
|
|
let signed = rp_state
|
|
.table_context
|
|
.validator
|
|
.as_ref()?
|
|
.sign(keystore, statement)
|
|
.ok()
|
|
.flatten()?;
|
|
metrics.on_statement_signed();
|
|
Some(signed)
|
|
}
|
|
|
|
/// Import a statement into the statement table and return the summary of the import.
|
|
///
|
|
/// This will fail with `Error::RejectedByProspectiveParachains` if the message type
|
|
/// is seconded, the candidate is fresh,
|
|
/// and any of the following are true:
|
|
/// 1. There is no `PersistedValidationData` attached.
|
|
/// 2. Prospective parachains are enabled for the relay parent and the prospective parachains
|
|
/// subsystem returned an empty `FragmentTreeMembership` i.e. did not recognize the candidate as
|
|
/// being applicable to any of the active leaves.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn import_statement<Context>(
|
|
ctx: &mut Context,
|
|
rp_state: &mut PerRelayParentState,
|
|
per_candidate: &mut HashMap<CandidateHash, PerCandidateState>,
|
|
statement: &SignedFullStatementWithPVD,
|
|
) -> Result<Option<TableSummary>, Error> {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
statement = ?statement.payload().to_compact(),
|
|
validator_index = statement.validator_index().0,
|
|
"Importing statement",
|
|
);
|
|
|
|
let candidate_hash = statement.payload().candidate_hash();
|
|
|
|
// If this is a new candidate (statement is 'seconded' and candidate is unknown),
|
|
// we need to create an entry in the `PerCandidateState` map.
|
|
//
|
|
// If the relay parent supports prospective parachains, we also need
|
|
// to inform the prospective parachains subsystem of the seconded candidate.
|
|
// If `ProspectiveParachainsMessage::Second` fails, then we return
|
|
// Error::RejectedByProspectiveParachains.
|
|
//
|
|
// Persisted Validation Data should be available - it may already be available
|
|
// if this is a candidate we are seconding.
|
|
//
|
|
// We should also not accept any candidates which have no valid depths under any of
|
|
// our active leaves.
|
|
if let StatementWithPVD::Seconded(candidate, pvd) = statement.payload() {
|
|
if !per_candidate.contains_key(&candidate_hash) {
|
|
if rp_state.prospective_parachains_mode.is_enabled() {
|
|
let (tx, rx) = oneshot::channel();
|
|
ctx.send_message(ProspectiveParachainsMessage::IntroduceCandidate(
|
|
IntroduceCandidateRequest {
|
|
candidate_para: candidate.descriptor().para_id,
|
|
candidate_receipt: candidate.clone(),
|
|
persisted_validation_data: pvd.clone(),
|
|
},
|
|
tx,
|
|
))
|
|
.await;
|
|
|
|
match rx.await {
|
|
Err(oneshot::Canceled) => {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
"Could not reach the Prospective Parachains subsystem."
|
|
);
|
|
|
|
return Err(Error::RejectedByProspectiveParachains)
|
|
},
|
|
Ok(membership) =>
|
|
if membership.is_empty() {
|
|
return Err(Error::RejectedByProspectiveParachains)
|
|
},
|
|
}
|
|
|
|
ctx.send_message(ProspectiveParachainsMessage::CandidateSeconded(
|
|
candidate.descriptor().para_id,
|
|
candidate_hash,
|
|
))
|
|
.await;
|
|
}
|
|
|
|
// Only save the candidate if it was approved by prospective parachains.
|
|
per_candidate.insert(
|
|
candidate_hash,
|
|
PerCandidateState {
|
|
persisted_validation_data: pvd.clone(),
|
|
// This is set after importing when seconding locally.
|
|
seconded_locally: false,
|
|
para_id: candidate.descriptor().para_id,
|
|
relay_parent: candidate.descriptor().relay_parent,
|
|
},
|
|
);
|
|
}
|
|
}
|
|
|
|
let stmt = primitive_statement_to_table(statement);
|
|
|
|
Ok(rp_state.table.import_statement(&rp_state.table_context, stmt))
|
|
}
|
|
|
|
/// Handles a summary received from [`import_statement`] and dispatches `Backed` notifications and
|
|
/// misbehaviors as a result of importing a statement.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn post_import_statement_actions<Context>(
|
|
ctx: &mut Context,
|
|
rp_state: &mut PerRelayParentState,
|
|
summary: Option<&TableSummary>,
|
|
) -> Result<(), Error> {
|
|
if let Some(attested) = summary.as_ref().and_then(|s| {
|
|
rp_state.table.attested_candidate(
|
|
&s.candidate,
|
|
&rp_state.table_context,
|
|
rp_state.minimum_backing_votes,
|
|
)
|
|
}) {
|
|
let candidate_hash = attested.candidate.hash();
|
|
|
|
// `HashSet::insert` returns true if the thing wasn't in there already.
|
|
if rp_state.backed.insert(candidate_hash) {
|
|
if let Some(backed) = table_attested_to_backed(attested, &rp_state.table_context) {
|
|
let para_id = backed.candidate.descriptor.para_id;
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate_hash,
|
|
relay_parent = ?rp_state.parent,
|
|
%para_id,
|
|
"Candidate backed",
|
|
);
|
|
|
|
if rp_state.prospective_parachains_mode.is_enabled() {
|
|
// Inform the prospective parachains subsystem
|
|
// that the candidate is now backed.
|
|
ctx.send_message(ProspectiveParachainsMessage::CandidateBacked(
|
|
para_id,
|
|
candidate_hash,
|
|
))
|
|
.await;
|
|
// Backed candidate potentially unblocks new advertisements,
|
|
// notify collator protocol.
|
|
ctx.send_message(CollatorProtocolMessage::Backed {
|
|
para_id,
|
|
para_head: backed.candidate.descriptor.para_head,
|
|
})
|
|
.await;
|
|
// Notify statement distribution of backed candidate.
|
|
ctx.send_message(StatementDistributionMessage::Backed(candidate_hash)).await;
|
|
} else {
|
|
// The provisioner waits on candidate-backing, which means
|
|
// that we need to send unbounded messages to avoid cycles.
|
|
//
|
|
// Backed candidates are bounded by the number of validators,
|
|
// parachains, and the block production rate of the relay chain.
|
|
let message = ProvisionerMessage::ProvisionableData(
|
|
rp_state.parent,
|
|
ProvisionableData::BackedCandidate(backed.receipt()),
|
|
);
|
|
ctx.send_unbounded_message(message);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
issue_new_misbehaviors(ctx, rp_state.parent, &mut rp_state.table);
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Check if there have happened any new misbehaviors and issue necessary messages.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
fn issue_new_misbehaviors<Context>(
|
|
ctx: &mut Context,
|
|
relay_parent: Hash,
|
|
table: &mut Table<TableContext>,
|
|
) {
|
|
// collect the misbehaviors to avoid double mutable self borrow issues
|
|
let misbehaviors: Vec<_> = table.drain_misbehaviors().collect();
|
|
for (validator_id, report) in misbehaviors {
|
|
// The provisioner waits on candidate-backing, which means
|
|
// that we need to send unbounded messages to avoid cycles.
|
|
//
|
|
// Misbehaviors are bounded by the number of validators and
|
|
// the block production protocol.
|
|
ctx.send_unbounded_message(ProvisionerMessage::ProvisionableData(
|
|
relay_parent,
|
|
ProvisionableData::MisbehaviorReport(relay_parent, validator_id, report),
|
|
));
|
|
}
|
|
}
|
|
|
|
/// Sign, import, and distribute a statement.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn sign_import_and_distribute_statement<Context>(
|
|
ctx: &mut Context,
|
|
rp_state: &mut PerRelayParentState,
|
|
per_candidate: &mut HashMap<CandidateHash, PerCandidateState>,
|
|
statement: StatementWithPVD,
|
|
keystore: KeystorePtr,
|
|
metrics: &Metrics,
|
|
) -> Result<Option<SignedFullStatementWithPVD>, Error> {
|
|
if let Some(signed_statement) = sign_statement(&*rp_state, statement, keystore, metrics) {
|
|
let summary = import_statement(ctx, rp_state, per_candidate, &signed_statement).await?;
|
|
|
|
// `Share` must always be sent before `Backed`. We send the latter in
|
|
// `post_import_statement_action` below.
|
|
let smsg = StatementDistributionMessage::Share(rp_state.parent, signed_statement.clone());
|
|
ctx.send_unbounded_message(smsg);
|
|
|
|
post_import_statement_actions(ctx, rp_state, summary.as_ref()).await?;
|
|
|
|
Ok(Some(signed_statement))
|
|
} else {
|
|
Ok(None)
|
|
}
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn background_validate_and_make_available<Context>(
|
|
ctx: &mut Context,
|
|
rp_state: &mut PerRelayParentState,
|
|
params: BackgroundValidationParams<
|
|
impl overseer::CandidateBackingSenderTrait,
|
|
impl Fn(BackgroundValidationResult) -> ValidatedCandidateCommand + Send + 'static + Sync,
|
|
>,
|
|
) -> Result<(), Error> {
|
|
let candidate_hash = params.candidate.hash();
|
|
if rp_state.awaiting_validation.insert(candidate_hash) {
|
|
// spawn background task.
|
|
let bg = async move {
|
|
if let Err(e) = validate_and_make_available(params).await {
|
|
if let Error::BackgroundValidationMpsc(error) = e {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
?error,
|
|
"Mpsc background validation mpsc died during validation- leaf no longer active?"
|
|
);
|
|
} else {
|
|
gum::error!(
|
|
target: LOG_TARGET,
|
|
"Failed to validate and make available: {:?}",
|
|
e
|
|
);
|
|
}
|
|
}
|
|
};
|
|
|
|
ctx.spawn("backing-validation", bg.boxed())
|
|
.map_err(|_| Error::FailedToSpawnBackgroundTask)?;
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Kick off validation work and distribute the result as a signed statement.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn kick_off_validation_work<Context>(
|
|
ctx: &mut Context,
|
|
rp_state: &mut PerRelayParentState,
|
|
persisted_validation_data: PersistedValidationData,
|
|
background_validation_tx: &mpsc::Sender<(Hash, ValidatedCandidateCommand)>,
|
|
attesting: AttestingData,
|
|
) -> Result<(), Error> {
|
|
let candidate_hash = attesting.candidate.hash();
|
|
if rp_state.issued_statements.contains(&candidate_hash) {
|
|
return Ok(())
|
|
}
|
|
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate_hash,
|
|
candidate_receipt = ?attesting.candidate,
|
|
"Kicking off validation",
|
|
);
|
|
|
|
let bg_sender = ctx.sender().clone();
|
|
let pov = PoVData::FetchFromValidator {
|
|
from_validator: attesting.from_validator,
|
|
candidate_hash,
|
|
pov_hash: attesting.pov_hash,
|
|
};
|
|
|
|
background_validate_and_make_available(
|
|
ctx,
|
|
rp_state,
|
|
BackgroundValidationParams {
|
|
sender: bg_sender,
|
|
tx_command: background_validation_tx.clone(),
|
|
candidate: attesting.candidate,
|
|
relay_parent: rp_state.parent,
|
|
persisted_validation_data,
|
|
pov,
|
|
n_validators: rp_state.table_context.validators.len(),
|
|
make_command: ValidatedCandidateCommand::Attest,
|
|
},
|
|
)
|
|
.await
|
|
}
|
|
|
|
/// Import the statement and kick off validation work if it is a part of our assignment.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn maybe_validate_and_import<Context>(
|
|
ctx: &mut Context,
|
|
state: &mut State,
|
|
relay_parent: Hash,
|
|
statement: SignedFullStatementWithPVD,
|
|
) -> Result<(), Error> {
|
|
let rp_state = match state.per_relay_parent.get_mut(&relay_parent) {
|
|
Some(r) => r,
|
|
None => {
|
|
gum::trace!(
|
|
target: LOG_TARGET,
|
|
?relay_parent,
|
|
"Received statement for unknown relay-parent"
|
|
);
|
|
|
|
return Ok(())
|
|
},
|
|
};
|
|
|
|
let res = import_statement(ctx, rp_state, &mut state.per_candidate, &statement).await;
|
|
|
|
// if we get an Error::RejectedByProspectiveParachains,
|
|
// we will do nothing.
|
|
if let Err(Error::RejectedByProspectiveParachains) = res {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
?relay_parent,
|
|
"Statement rejected by prospective parachains."
|
|
);
|
|
|
|
return Ok(())
|
|
}
|
|
|
|
let summary = res?;
|
|
post_import_statement_actions(ctx, rp_state, summary.as_ref()).await?;
|
|
|
|
if let Some(summary) = summary {
|
|
// import_statement already takes care of communicating with the
|
|
// prospective parachains subsystem. At this point, the candidate
|
|
// has already been accepted into the fragment trees.
|
|
|
|
let candidate_hash = summary.candidate;
|
|
|
|
if Some(summary.group_id) != rp_state.assignment {
|
|
return Ok(())
|
|
}
|
|
let attesting = match statement.payload() {
|
|
StatementWithPVD::Seconded(receipt, _) => {
|
|
let attesting = AttestingData {
|
|
candidate: rp_state
|
|
.table
|
|
.get_candidate(&candidate_hash)
|
|
.ok_or(Error::CandidateNotFound)?
|
|
.to_plain(),
|
|
pov_hash: receipt.descriptor.pov_hash,
|
|
from_validator: statement.validator_index(),
|
|
backing: Vec::new(),
|
|
};
|
|
rp_state.fallbacks.insert(summary.candidate, attesting.clone());
|
|
attesting
|
|
},
|
|
StatementWithPVD::Valid(candidate_hash) => {
|
|
if let Some(attesting) = rp_state.fallbacks.get_mut(candidate_hash) {
|
|
let our_index = rp_state.table_context.validator.as_ref().map(|v| v.index());
|
|
if our_index == Some(statement.validator_index()) {
|
|
return Ok(())
|
|
}
|
|
|
|
if rp_state.awaiting_validation.contains(candidate_hash) {
|
|
// Job already running:
|
|
attesting.backing.push(statement.validator_index());
|
|
return Ok(())
|
|
} else {
|
|
// No job, so start another with current validator:
|
|
attesting.from_validator = statement.validator_index();
|
|
attesting.clone()
|
|
}
|
|
} else {
|
|
return Ok(())
|
|
}
|
|
},
|
|
};
|
|
|
|
// After `import_statement` succeeds, the candidate entry is guaranteed
|
|
// to exist.
|
|
if let Some(pvd) = state
|
|
.per_candidate
|
|
.get(&candidate_hash)
|
|
.map(|pc| pc.persisted_validation_data.clone())
|
|
{
|
|
kick_off_validation_work(
|
|
ctx,
|
|
rp_state,
|
|
pvd,
|
|
&state.background_validation_tx,
|
|
attesting,
|
|
)
|
|
.await?;
|
|
}
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
/// Kick off background validation with intent to second.
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn validate_and_second<Context>(
|
|
ctx: &mut Context,
|
|
rp_state: &mut PerRelayParentState,
|
|
persisted_validation_data: PersistedValidationData,
|
|
candidate: &CandidateReceipt,
|
|
pov: Arc<PoV>,
|
|
background_validation_tx: &mpsc::Sender<(Hash, ValidatedCandidateCommand)>,
|
|
) -> Result<(), Error> {
|
|
let candidate_hash = candidate.hash();
|
|
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
candidate_hash = ?candidate_hash,
|
|
candidate_receipt = ?candidate,
|
|
"Validate and second candidate",
|
|
);
|
|
|
|
let bg_sender = ctx.sender().clone();
|
|
background_validate_and_make_available(
|
|
ctx,
|
|
rp_state,
|
|
BackgroundValidationParams {
|
|
sender: bg_sender,
|
|
tx_command: background_validation_tx.clone(),
|
|
candidate: candidate.clone(),
|
|
relay_parent: rp_state.parent,
|
|
persisted_validation_data,
|
|
pov: PoVData::Ready(pov),
|
|
n_validators: rp_state.table_context.validators.len(),
|
|
make_command: ValidatedCandidateCommand::Second,
|
|
},
|
|
)
|
|
.await?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn handle_second_message<Context>(
|
|
ctx: &mut Context,
|
|
state: &mut State,
|
|
candidate: CandidateReceipt,
|
|
persisted_validation_data: PersistedValidationData,
|
|
pov: PoV,
|
|
metrics: &Metrics,
|
|
) -> Result<(), Error> {
|
|
let _timer = metrics.time_process_second();
|
|
|
|
let candidate_hash = candidate.hash();
|
|
let relay_parent = candidate.descriptor().relay_parent;
|
|
|
|
if candidate.descriptor().persisted_validation_data_hash != persisted_validation_data.hash() {
|
|
gum::warn!(
|
|
target: LOG_TARGET,
|
|
?candidate_hash,
|
|
"Candidate backing was asked to second candidate with wrong PVD",
|
|
);
|
|
|
|
return Ok(())
|
|
}
|
|
|
|
let rp_state = match state.per_relay_parent.get_mut(&relay_parent) {
|
|
None => {
|
|
gum::trace!(
|
|
target: LOG_TARGET,
|
|
?relay_parent,
|
|
?candidate_hash,
|
|
"We were asked to second a candidate outside of our view."
|
|
);
|
|
|
|
return Ok(())
|
|
},
|
|
Some(r) => r,
|
|
};
|
|
|
|
// Sanity check that candidate is from our assignment.
|
|
if Some(candidate.descriptor().para_id) != rp_state.assignment {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
our_assignment = ?rp_state.assignment,
|
|
collation = ?candidate.descriptor().para_id,
|
|
"Subsystem asked to second for para outside of our assignment",
|
|
);
|
|
|
|
return Ok(())
|
|
}
|
|
|
|
// If the message is a `CandidateBackingMessage::Second`, sign and dispatch a
|
|
// Seconded statement only if we have not signed a Valid statement for the requested candidate.
|
|
//
|
|
// The actual logic of issuing the signed statement checks that this isn't
|
|
// conflicting with other seconded candidates. Not doing that check here
|
|
// gives other subsystems the ability to get us to execute arbitrary candidates,
|
|
// but no more.
|
|
if !rp_state.issued_statements.contains(&candidate_hash) {
|
|
let pov = Arc::new(pov);
|
|
|
|
validate_and_second(
|
|
ctx,
|
|
rp_state,
|
|
persisted_validation_data,
|
|
&candidate,
|
|
pov,
|
|
&state.background_validation_tx,
|
|
)
|
|
.await?;
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[overseer::contextbounds(CandidateBacking, prefix = self::overseer)]
|
|
async fn handle_statement_message<Context>(
|
|
ctx: &mut Context,
|
|
state: &mut State,
|
|
relay_parent: Hash,
|
|
statement: SignedFullStatementWithPVD,
|
|
metrics: &Metrics,
|
|
) -> Result<(), Error> {
|
|
let _timer = metrics.time_process_statement();
|
|
|
|
match maybe_validate_and_import(ctx, state, relay_parent, statement).await {
|
|
Err(Error::ValidationFailed(_)) => Ok(()),
|
|
Err(e) => Err(e),
|
|
Ok(()) => Ok(()),
|
|
}
|
|
}
|
|
|
|
fn handle_get_backed_candidates_message(
|
|
state: &State,
|
|
requested_candidates: Vec<(CandidateHash, Hash)>,
|
|
tx: oneshot::Sender<Vec<BackedCandidate>>,
|
|
metrics: &Metrics,
|
|
) -> Result<(), Error> {
|
|
let _timer = metrics.time_get_backed_candidates();
|
|
|
|
let backed = requested_candidates
|
|
.into_iter()
|
|
.filter_map(|(candidate_hash, relay_parent)| {
|
|
let rp_state = match state.per_relay_parent.get(&relay_parent) {
|
|
Some(rp_state) => rp_state,
|
|
None => {
|
|
gum::debug!(
|
|
target: LOG_TARGET,
|
|
?relay_parent,
|
|
?candidate_hash,
|
|
"Requested candidate's relay parent is out of view",
|
|
);
|
|
return None
|
|
},
|
|
};
|
|
rp_state
|
|
.table
|
|
.attested_candidate(
|
|
&candidate_hash,
|
|
&rp_state.table_context,
|
|
rp_state.minimum_backing_votes,
|
|
)
|
|
.and_then(|attested| table_attested_to_backed(attested, &rp_state.table_context))
|
|
})
|
|
.collect();
|
|
|
|
tx.send(backed).map_err(|data| Error::Send(data))?;
|
|
Ok(())
|
|
}
|