mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-29 18:27:56 +00:00
3f5156e866
* refactor View to include finalized_number * guide: update the NetworkBridge on BlockFinalized * av-store: fix the tests * actually fix tests * grumbles * ignore macro doctest * use Hash::repeat_bytes more consistently * broadcast empty leaves updates as well * fix issuing view updates on empty leaves updates
774 lines
24 KiB
Rust
774 lines
24 KiB
Rust
// Copyright 2020 Parity Technologies (UK) Ltd.
|
|
// This file is part of Polkadot.
|
|
|
|
// Polkadot is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// Polkadot is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with Polkadot. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
//! The collation generation subsystem is the interface between polkadot and the collators.
|
|
|
|
#![deny(missing_docs)]
|
|
|
|
use futures::{
|
|
channel::mpsc,
|
|
future::FutureExt,
|
|
join,
|
|
select,
|
|
sink::SinkExt,
|
|
stream::StreamExt,
|
|
};
|
|
use polkadot_node_primitives::CollationGenerationConfig;
|
|
use polkadot_node_subsystem::{
|
|
messages::{AllMessages, CollationGenerationMessage, CollatorProtocolMessage},
|
|
FromOverseer, SpawnedSubsystem, Subsystem, SubsystemContext, SubsystemResult,
|
|
};
|
|
use polkadot_node_subsystem_util::{
|
|
request_availability_cores_ctx, request_full_validation_data_ctx,
|
|
request_validators_ctx,
|
|
metrics::{self, prometheus},
|
|
};
|
|
use polkadot_primitives::v1::{
|
|
collator_signature_payload, AvailableData, CandidateCommitments,
|
|
CandidateDescriptor, CandidateReceipt, CoreState, Hash, OccupiedCoreAssumption,
|
|
PersistedValidationData, PoV,
|
|
};
|
|
use sp_core::crypto::Pair;
|
|
use std::sync::Arc;
|
|
|
|
mod error;
|
|
|
|
const LOG_TARGET: &'static str = "collation_generation";
|
|
|
|
/// Collation Generation Subsystem
|
|
pub struct CollationGenerationSubsystem {
|
|
config: Option<Arc<CollationGenerationConfig>>,
|
|
metrics: Metrics,
|
|
}
|
|
|
|
impl CollationGenerationSubsystem {
|
|
/// Create a new instance of the `CollationGenerationSubsystem`.
|
|
pub fn new(metrics: Metrics) -> Self {
|
|
Self {
|
|
config: None,
|
|
metrics,
|
|
}
|
|
}
|
|
|
|
/// Run this subsystem
|
|
///
|
|
/// Conceptually, this is very simple: it just loops forever.
|
|
///
|
|
/// - On incoming overseer messages, it starts or stops jobs as appropriate.
|
|
/// - On other incoming messages, if they can be converted into Job::ToJob and
|
|
/// include a hash, then they're forwarded to the appropriate individual job.
|
|
/// - On outgoing messages from the jobs, it forwards them to the overseer.
|
|
///
|
|
/// If `err_tx` is not `None`, errors are forwarded onto that channel as they occur.
|
|
/// Otherwise, most are logged and then discarded.
|
|
#[tracing::instrument(skip(self, ctx), fields(subsystem = LOG_TARGET))]
|
|
async fn run<Context>(mut self, mut ctx: Context)
|
|
where
|
|
Context: SubsystemContext<Message = CollationGenerationMessage>,
|
|
{
|
|
// when we activate new leaves, we spawn a bunch of sub-tasks, each of which is
|
|
// expected to generate precisely one message. We don't want to block the main loop
|
|
// at any point waiting for them all, so instead, we create a channel on which they can
|
|
// send those messages. We can then just monitor the channel and forward messages on it
|
|
// to the overseer here, via the context.
|
|
let (sender, receiver) = mpsc::channel(0);
|
|
|
|
let mut receiver = receiver.fuse();
|
|
loop {
|
|
select! {
|
|
incoming = ctx.recv().fuse() => {
|
|
if self.handle_incoming::<Context>(incoming, &mut ctx, &sender).await {
|
|
break;
|
|
}
|
|
},
|
|
msg = receiver.next() => {
|
|
if let Some(msg) = msg {
|
|
ctx.send_message(msg).await;
|
|
}
|
|
},
|
|
}
|
|
}
|
|
}
|
|
|
|
// handle an incoming message. return true if we should break afterwards.
|
|
// note: this doesn't strictly need to be a separate function; it's more an administrative function
|
|
// so that we don't clutter the run loop. It could in principle be inlined directly into there.
|
|
// it should hopefully therefore be ok that it's an async function mutably borrowing self.
|
|
#[tracing::instrument(level = "trace", skip(self, ctx, sender), fields(subsystem = LOG_TARGET))]
|
|
async fn handle_incoming<Context>(
|
|
&mut self,
|
|
incoming: SubsystemResult<FromOverseer<Context::Message>>,
|
|
ctx: &mut Context,
|
|
sender: &mpsc::Sender<AllMessages>,
|
|
) -> bool
|
|
where
|
|
Context: SubsystemContext<Message = CollationGenerationMessage>,
|
|
{
|
|
use polkadot_node_subsystem::ActiveLeavesUpdate;
|
|
use polkadot_node_subsystem::FromOverseer::{Communication, Signal};
|
|
use polkadot_node_subsystem::OverseerSignal::{ActiveLeaves, BlockFinalized, Conclude};
|
|
|
|
match incoming {
|
|
Ok(Signal(ActiveLeaves(ActiveLeavesUpdate { activated, .. }))) => {
|
|
// follow the procedure from the guide
|
|
if let Some(config) = &self.config {
|
|
let metrics = self.metrics.clone();
|
|
if let Err(err) =
|
|
handle_new_activations(config.clone(), &activated, ctx, metrics, sender).await
|
|
{
|
|
tracing::warn!(target: LOG_TARGET, err = ?err, "failed to handle new activations");
|
|
};
|
|
}
|
|
false
|
|
}
|
|
Ok(Signal(Conclude)) => true,
|
|
Ok(Communication {
|
|
msg: CollationGenerationMessage::Initialize(config),
|
|
}) => {
|
|
if self.config.is_some() {
|
|
tracing::error!(target: LOG_TARGET, "double initialization");
|
|
} else {
|
|
self.config = Some(Arc::new(config));
|
|
}
|
|
false
|
|
}
|
|
Ok(Signal(BlockFinalized(..))) => false,
|
|
Err(err) => {
|
|
tracing::error!(
|
|
target: LOG_TARGET,
|
|
err = ?err,
|
|
"error receiving message from subsystem context: {:?}",
|
|
err
|
|
);
|
|
true
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<Context> Subsystem<Context> for CollationGenerationSubsystem
|
|
where
|
|
Context: SubsystemContext<Message = CollationGenerationMessage>,
|
|
{
|
|
fn start(self, ctx: Context) -> SpawnedSubsystem {
|
|
let future = Box::pin(async move {
|
|
self.run(ctx).await;
|
|
Ok(())
|
|
});
|
|
|
|
SpawnedSubsystem {
|
|
name: "collation-generation-subsystem",
|
|
future,
|
|
}
|
|
}
|
|
}
|
|
|
|
#[tracing::instrument(level = "trace", skip(ctx, metrics, sender), fields(subsystem = LOG_TARGET))]
|
|
async fn handle_new_activations<Context: SubsystemContext>(
|
|
config: Arc<CollationGenerationConfig>,
|
|
activated: &[Hash],
|
|
ctx: &mut Context,
|
|
metrics: Metrics,
|
|
sender: &mpsc::Sender<AllMessages>,
|
|
) -> crate::error::Result<()> {
|
|
// follow the procedure from the guide:
|
|
// https://w3f.github.io/parachain-implementers-guide/node/collators/collation-generation.html
|
|
|
|
let _overall_timer = metrics.time_new_activations();
|
|
|
|
for relay_parent in activated.iter().copied() {
|
|
let _relay_parent_timer = metrics.time_new_activations_relay_parent();
|
|
|
|
// double-future magic happens here: the first layer of requests takes a mutable borrow of the context, and
|
|
// returns a receiver. The second layer of requests actually polls those receivers to completion.
|
|
let (availability_cores, validators) = join!(
|
|
request_availability_cores_ctx(relay_parent, ctx).await?,
|
|
request_validators_ctx(relay_parent, ctx).await?,
|
|
);
|
|
|
|
let availability_cores = availability_cores??;
|
|
let n_validators = validators??.len();
|
|
|
|
for (core_idx, core) in availability_cores.into_iter().enumerate() {
|
|
let _availability_core_timer = metrics.time_new_activations_availability_core();
|
|
|
|
let (scheduled_core, assumption) = match core {
|
|
CoreState::Scheduled(scheduled_core) => {
|
|
(scheduled_core, OccupiedCoreAssumption::Free)
|
|
}
|
|
CoreState::Occupied(_occupied_core) => {
|
|
// TODO: https://github.com/paritytech/polkadot/issues/1573
|
|
tracing::trace!(
|
|
target: LOG_TARGET,
|
|
core_idx = %core_idx,
|
|
relay_parent = ?relay_parent,
|
|
"core is occupied. Keep going.",
|
|
);
|
|
continue;
|
|
}
|
|
CoreState::Free => {
|
|
tracing::trace!(
|
|
target: LOG_TARGET,
|
|
core_idx = %core_idx,
|
|
"core is free. Keep going.",
|
|
);
|
|
continue
|
|
}
|
|
};
|
|
|
|
if scheduled_core.para_id != config.para_id {
|
|
tracing::trace!(
|
|
target: LOG_TARGET,
|
|
core_idx = %core_idx,
|
|
relay_parent = ?relay_parent,
|
|
our_para = %config.para_id,
|
|
their_para = %scheduled_core.para_id,
|
|
"core is not assigned to our para. Keep going.",
|
|
);
|
|
continue;
|
|
}
|
|
|
|
// we get validation data synchronously for each core instead of
|
|
// within the subtask loop, because we have only a single mutable handle to the
|
|
// context, so the work can't really be distributed
|
|
let validation_data = match request_full_validation_data_ctx(
|
|
relay_parent,
|
|
scheduled_core.para_id,
|
|
assumption,
|
|
ctx,
|
|
)
|
|
.await?
|
|
.await??
|
|
{
|
|
Some(v) => v,
|
|
None => {
|
|
tracing::trace!(
|
|
target: LOG_TARGET,
|
|
core_idx = %core_idx,
|
|
relay_parent = ?relay_parent,
|
|
our_para = %config.para_id,
|
|
their_para = %scheduled_core.para_id,
|
|
"validation data is not available",
|
|
);
|
|
continue
|
|
}
|
|
};
|
|
|
|
let task_config = config.clone();
|
|
let mut task_sender = sender.clone();
|
|
let metrics = metrics.clone();
|
|
ctx.spawn("collation generation collation builder", Box::pin(async move {
|
|
let persisted_validation_data_hash = validation_data.persisted.hash();
|
|
|
|
let collation = match (task_config.collator)(relay_parent, &validation_data).await {
|
|
Some(collation) => collation,
|
|
None => {
|
|
tracing::debug!(
|
|
target: LOG_TARGET,
|
|
para_id = %scheduled_core.para_id,
|
|
"collator returned no collation on collate",
|
|
);
|
|
return
|
|
}
|
|
};
|
|
|
|
let pov_hash = collation.proof_of_validity.hash();
|
|
|
|
let signature_payload = collator_signature_payload(
|
|
&relay_parent,
|
|
&scheduled_core.para_id,
|
|
&persisted_validation_data_hash,
|
|
&pov_hash,
|
|
);
|
|
|
|
let erasure_root = match erasure_root(
|
|
n_validators,
|
|
validation_data.persisted,
|
|
collation.proof_of_validity.clone(),
|
|
) {
|
|
Ok(erasure_root) => erasure_root,
|
|
Err(err) => {
|
|
tracing::error!(
|
|
target: LOG_TARGET,
|
|
para_id = %scheduled_core.para_id,
|
|
err = ?err,
|
|
"failed to calculate erasure root",
|
|
);
|
|
return
|
|
}
|
|
};
|
|
|
|
let commitments = CandidateCommitments {
|
|
upward_messages: collation.upward_messages,
|
|
horizontal_messages: collation.horizontal_messages,
|
|
new_validation_code: collation.new_validation_code,
|
|
head_data: collation.head_data,
|
|
processed_downward_messages: collation.processed_downward_messages,
|
|
hrmp_watermark: collation.hrmp_watermark,
|
|
};
|
|
|
|
let ccr = CandidateReceipt {
|
|
commitments_hash: commitments.hash(),
|
|
descriptor: CandidateDescriptor {
|
|
signature: task_config.key.sign(&signature_payload),
|
|
para_id: scheduled_core.para_id,
|
|
relay_parent,
|
|
collator: task_config.key.public(),
|
|
persisted_validation_data_hash,
|
|
pov_hash,
|
|
erasure_root,
|
|
},
|
|
};
|
|
|
|
metrics.on_collation_generated();
|
|
|
|
if let Err(err) = task_sender.send(AllMessages::CollatorProtocol(
|
|
CollatorProtocolMessage::DistributeCollation(ccr, collation.proof_of_validity)
|
|
)).await {
|
|
tracing::warn!(
|
|
target: LOG_TARGET,
|
|
para_id = %scheduled_core.para_id,
|
|
err = ?err,
|
|
"failed to send collation result",
|
|
);
|
|
}
|
|
})).await?;
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[tracing::instrument(level = "trace", fields(subsystem = LOG_TARGET))]
|
|
fn erasure_root(
|
|
n_validators: usize,
|
|
persisted_validation: PersistedValidationData,
|
|
pov: PoV,
|
|
) -> crate::error::Result<Hash> {
|
|
let available_data = AvailableData {
|
|
validation_data: persisted_validation,
|
|
pov: Arc::new(pov),
|
|
};
|
|
|
|
let chunks = polkadot_erasure_coding::obtain_chunks_v1(n_validators, &available_data)?;
|
|
Ok(polkadot_erasure_coding::branches(&chunks).root())
|
|
}
|
|
|
|
#[derive(Clone)]
|
|
struct MetricsInner {
|
|
collations_generated_total: prometheus::Counter<prometheus::U64>,
|
|
new_activations_overall: prometheus::Histogram,
|
|
new_activations_per_relay_parent: prometheus::Histogram,
|
|
new_activations_per_availability_core: prometheus::Histogram,
|
|
}
|
|
|
|
/// CollationGenerationSubsystem metrics.
|
|
#[derive(Default, Clone)]
|
|
pub struct Metrics(Option<MetricsInner>);
|
|
|
|
impl Metrics {
|
|
fn on_collation_generated(&self) {
|
|
if let Some(metrics) = &self.0 {
|
|
metrics.collations_generated_total.inc();
|
|
}
|
|
}
|
|
|
|
/// Provide a timer for new activations which updates on drop.
|
|
fn time_new_activations(&self) -> Option<metrics::prometheus::prometheus::HistogramTimer> {
|
|
self.0.as_ref().map(|metrics| metrics.new_activations_overall.start_timer())
|
|
}
|
|
|
|
/// Provide a timer per relay parents which updates on drop.
|
|
fn time_new_activations_relay_parent(&self) -> Option<metrics::prometheus::prometheus::HistogramTimer> {
|
|
self.0.as_ref().map(|metrics| metrics.new_activations_per_relay_parent.start_timer())
|
|
}
|
|
|
|
/// Provide a timer per availability core which updates on drop.
|
|
fn time_new_activations_availability_core(&self) -> Option<metrics::prometheus::prometheus::HistogramTimer> {
|
|
self.0.as_ref().map(|metrics| metrics.new_activations_per_availability_core.start_timer())
|
|
}
|
|
}
|
|
|
|
impl metrics::Metrics for Metrics {
|
|
fn try_register(registry: &prometheus::Registry) -> Result<Self, prometheus::PrometheusError> {
|
|
let metrics = MetricsInner {
|
|
collations_generated_total: prometheus::register(
|
|
prometheus::Counter::new(
|
|
"parachain_collations_generated_total",
|
|
"Number of collations generated."
|
|
)?,
|
|
registry,
|
|
)?,
|
|
new_activations_overall: prometheus::register(
|
|
prometheus::Histogram::with_opts(
|
|
prometheus::HistogramOpts::new(
|
|
"parachain_collation_generation_new_activations",
|
|
"Time spent within fn handle_new_activations",
|
|
)
|
|
)?,
|
|
registry,
|
|
)?,
|
|
new_activations_per_relay_parent: prometheus::register(
|
|
prometheus::Histogram::with_opts(
|
|
prometheus::HistogramOpts::new(
|
|
"parachain_collation_generation_per_relay_parent",
|
|
"Time spent handling a particular relay parent within fn handle_new_activations"
|
|
)
|
|
)?,
|
|
registry,
|
|
)?,
|
|
new_activations_per_availability_core: prometheus::register(
|
|
prometheus::Histogram::with_opts(
|
|
prometheus::HistogramOpts::new(
|
|
"parachain_collation_generation_per_availability_core",
|
|
"Time spent handling a particular availability core for a relay parent in fn handle_new_activations",
|
|
)
|
|
)?,
|
|
registry,
|
|
)?,
|
|
};
|
|
Ok(Metrics(Some(metrics)))
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
mod handle_new_activations {
|
|
use super::super::*;
|
|
use futures::{
|
|
lock::Mutex,
|
|
task::{Context as FuturesContext, Poll},
|
|
Future,
|
|
};
|
|
use polkadot_node_primitives::Collation;
|
|
use polkadot_node_subsystem::messages::{
|
|
AllMessages, RuntimeApiMessage, RuntimeApiRequest,
|
|
};
|
|
use polkadot_node_subsystem_test_helpers::{
|
|
subsystem_test_harness, TestSubsystemContextHandle,
|
|
};
|
|
use polkadot_primitives::v1::{
|
|
BlockData, BlockNumber, CollatorPair, Id as ParaId,
|
|
PersistedValidationData, PoV, ScheduledCore, ValidationData,
|
|
};
|
|
use std::pin::Pin;
|
|
|
|
fn test_collation() -> Collation {
|
|
Collation {
|
|
upward_messages: Default::default(),
|
|
horizontal_messages: Default::default(),
|
|
new_validation_code: Default::default(),
|
|
head_data: Default::default(),
|
|
proof_of_validity: PoV {
|
|
block_data: BlockData(Vec::new()),
|
|
},
|
|
processed_downward_messages: Default::default(),
|
|
hrmp_watermark: Default::default(),
|
|
}
|
|
}
|
|
|
|
// Box<dyn Future<Output = Collation> + Unpin + Send
|
|
struct TestCollator;
|
|
|
|
impl Future for TestCollator {
|
|
type Output = Option<Collation>;
|
|
|
|
fn poll(self: Pin<&mut Self>, _cx: &mut FuturesContext) -> Poll<Self::Output> {
|
|
Poll::Ready(Some(test_collation()))
|
|
}
|
|
}
|
|
|
|
impl Unpin for TestCollator {}
|
|
|
|
fn test_config<Id: Into<ParaId>>(para_id: Id) -> Arc<CollationGenerationConfig> {
|
|
Arc::new(CollationGenerationConfig {
|
|
key: CollatorPair::generate().0,
|
|
collator: Box::new(|_: Hash, _vd: &ValidationData| {
|
|
TestCollator.boxed()
|
|
}),
|
|
para_id: para_id.into(),
|
|
})
|
|
}
|
|
|
|
fn scheduled_core_for<Id: Into<ParaId>>(para_id: Id) -> ScheduledCore {
|
|
ScheduledCore {
|
|
para_id: para_id.into(),
|
|
collator: None,
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn requests_availability_per_relay_parent() {
|
|
let activated_hashes: Vec<Hash> = vec![
|
|
[1; 32].into(),
|
|
[4; 32].into(),
|
|
[9; 32].into(),
|
|
[16; 32].into(),
|
|
];
|
|
|
|
let requested_availability_cores = Arc::new(Mutex::new(Vec::new()));
|
|
|
|
let overseer_requested_availability_cores = requested_availability_cores.clone();
|
|
let overseer = |mut handle: TestSubsystemContextHandle<CollationGenerationMessage>| async move {
|
|
loop {
|
|
match handle.try_recv().await {
|
|
None => break,
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(hash, RuntimeApiRequest::AvailabilityCores(tx)))) => {
|
|
overseer_requested_availability_cores.lock().await.push(hash);
|
|
tx.send(Ok(vec![])).unwrap();
|
|
}
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(_hash, RuntimeApiRequest::Validators(tx)))) => {
|
|
tx.send(Ok(vec![Default::default(); 3])).unwrap();
|
|
}
|
|
Some(msg) => panic!("didn't expect any other overseer requests given no availability cores; got {:?}", msg),
|
|
}
|
|
}
|
|
};
|
|
|
|
let (tx, _rx) = mpsc::channel(0);
|
|
|
|
let subsystem_activated_hashes = activated_hashes.clone();
|
|
subsystem_test_harness(overseer, |mut ctx| async move {
|
|
handle_new_activations(
|
|
test_config(123u32),
|
|
&subsystem_activated_hashes,
|
|
&mut ctx,
|
|
Metrics(None),
|
|
&tx,
|
|
)
|
|
.await
|
|
.unwrap();
|
|
});
|
|
|
|
let mut requested_availability_cores = Arc::try_unwrap(requested_availability_cores)
|
|
.expect("overseer should have shut down by now")
|
|
.into_inner();
|
|
requested_availability_cores.sort();
|
|
|
|
assert_eq!(requested_availability_cores, activated_hashes);
|
|
}
|
|
|
|
#[test]
|
|
fn requests_validation_data_for_scheduled_matches() {
|
|
let activated_hashes: Vec<Hash> = vec![
|
|
Hash::repeat_byte(1),
|
|
Hash::repeat_byte(4),
|
|
Hash::repeat_byte(9),
|
|
Hash::repeat_byte(16),
|
|
];
|
|
|
|
let requested_full_validation_data = Arc::new(Mutex::new(Vec::new()));
|
|
|
|
let overseer_requested_full_validation_data = requested_full_validation_data.clone();
|
|
let overseer = |mut handle: TestSubsystemContextHandle<CollationGenerationMessage>| async move {
|
|
loop {
|
|
match handle.try_recv().await {
|
|
None => break,
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(
|
|
hash,
|
|
RuntimeApiRequest::AvailabilityCores(tx),
|
|
))) => {
|
|
tx.send(Ok(vec![
|
|
CoreState::Free,
|
|
// this is weird, see explanation below
|
|
CoreState::Scheduled(scheduled_core_for(
|
|
(hash.as_fixed_bytes()[0] * 4) as u32,
|
|
)),
|
|
CoreState::Scheduled(scheduled_core_for(
|
|
(hash.as_fixed_bytes()[0] * 5) as u32,
|
|
)),
|
|
]))
|
|
.unwrap();
|
|
}
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(
|
|
hash,
|
|
RuntimeApiRequest::FullValidationData(
|
|
_para_id,
|
|
_occupied_core_assumption,
|
|
tx,
|
|
),
|
|
))) => {
|
|
overseer_requested_full_validation_data
|
|
.lock()
|
|
.await
|
|
.push(hash);
|
|
tx.send(Ok(Default::default())).unwrap();
|
|
}
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(
|
|
_hash,
|
|
RuntimeApiRequest::Validators(tx),
|
|
))) => {
|
|
tx.send(Ok(vec![Default::default(); 3])).unwrap();
|
|
}
|
|
Some(msg) => {
|
|
panic!("didn't expect any other overseer requests; got {:?}", msg)
|
|
}
|
|
}
|
|
}
|
|
};
|
|
|
|
let (tx, _rx) = mpsc::channel(0);
|
|
|
|
subsystem_test_harness(overseer, |mut ctx| async move {
|
|
handle_new_activations(test_config(16), &activated_hashes, &mut ctx, Metrics(None), &tx)
|
|
.await
|
|
.unwrap();
|
|
});
|
|
|
|
let requested_full_validation_data = Arc::try_unwrap(requested_full_validation_data)
|
|
.expect("overseer should have shut down by now")
|
|
.into_inner();
|
|
|
|
// the only activated hash should be from the 4 hash:
|
|
// each activated hash generates two scheduled cores: one with its value * 4, one with its value * 5
|
|
// given that the test configuration has a para_id of 16, there's only one way to get that value: with the 4
|
|
// hash.
|
|
assert_eq!(requested_full_validation_data, vec![[4; 32].into()]);
|
|
}
|
|
|
|
#[test]
|
|
fn sends_distribute_collation_message() {
|
|
let activated_hashes: Vec<Hash> = vec![
|
|
Hash::repeat_byte(1),
|
|
Hash::repeat_byte(4),
|
|
Hash::repeat_byte(9),
|
|
Hash::repeat_byte(16),
|
|
];
|
|
|
|
let overseer = |mut handle: TestSubsystemContextHandle<CollationGenerationMessage>| async move {
|
|
loop {
|
|
match handle.try_recv().await {
|
|
None => break,
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(
|
|
hash,
|
|
RuntimeApiRequest::AvailabilityCores(tx),
|
|
))) => {
|
|
tx.send(Ok(vec![
|
|
CoreState::Free,
|
|
// this is weird, see explanation below
|
|
CoreState::Scheduled(scheduled_core_for(
|
|
(hash.as_fixed_bytes()[0] * 4) as u32,
|
|
)),
|
|
CoreState::Scheduled(scheduled_core_for(
|
|
(hash.as_fixed_bytes()[0] * 5) as u32,
|
|
)),
|
|
]))
|
|
.unwrap();
|
|
}
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(
|
|
_hash,
|
|
RuntimeApiRequest::FullValidationData(
|
|
_para_id,
|
|
_occupied_core_assumption,
|
|
tx,
|
|
),
|
|
))) => {
|
|
tx.send(Ok(Some(Default::default()))).unwrap();
|
|
}
|
|
Some(AllMessages::RuntimeApi(RuntimeApiMessage::Request(
|
|
_hash,
|
|
RuntimeApiRequest::Validators(tx),
|
|
))) => {
|
|
tx.send(Ok(vec![Default::default(); 3])).unwrap();
|
|
}
|
|
Some(msg) => {
|
|
panic!("didn't expect any other overseer requests; got {:?}", msg)
|
|
}
|
|
}
|
|
}
|
|
};
|
|
|
|
let config = test_config(16);
|
|
let subsystem_config = config.clone();
|
|
|
|
let (tx, rx) = mpsc::channel(0);
|
|
|
|
// empty vec doesn't allocate on the heap, so it's ok we throw it away
|
|
let sent_messages = Arc::new(Mutex::new(Vec::new()));
|
|
let subsystem_sent_messages = sent_messages.clone();
|
|
subsystem_test_harness(overseer, |mut ctx| async move {
|
|
handle_new_activations(subsystem_config, &activated_hashes, &mut ctx, Metrics(None), &tx)
|
|
.await
|
|
.unwrap();
|
|
|
|
std::mem::drop(tx);
|
|
|
|
// collect all sent messages
|
|
*subsystem_sent_messages.lock().await = rx.collect().await;
|
|
});
|
|
|
|
let sent_messages = Arc::try_unwrap(sent_messages)
|
|
.expect("subsystem should have shut down by now")
|
|
.into_inner();
|
|
|
|
// we expect a single message to be sent, containing a candidate receipt.
|
|
// we don't care too much about the commitments_hash right now, but let's ensure that we've calculated the
|
|
// correct descriptor
|
|
let expect_pov_hash = test_collation().proof_of_validity.hash();
|
|
let expect_validation_data_hash
|
|
= PersistedValidationData::<BlockNumber>::default().hash();
|
|
let expect_relay_parent = Hash::repeat_byte(4);
|
|
let expect_payload = collator_signature_payload(
|
|
&expect_relay_parent,
|
|
&config.para_id,
|
|
&expect_validation_data_hash,
|
|
&expect_pov_hash,
|
|
);
|
|
let expect_descriptor = CandidateDescriptor {
|
|
signature: config.key.sign(&expect_payload),
|
|
para_id: config.para_id,
|
|
relay_parent: expect_relay_parent,
|
|
collator: config.key.public(),
|
|
persisted_validation_data_hash: expect_validation_data_hash,
|
|
pov_hash: expect_pov_hash,
|
|
erasure_root: Default::default(), // this isn't something we're checking right now
|
|
};
|
|
|
|
assert_eq!(sent_messages.len(), 1);
|
|
match &sent_messages[0] {
|
|
AllMessages::CollatorProtocol(CollatorProtocolMessage::DistributeCollation(
|
|
CandidateReceipt { descriptor, .. },
|
|
_pov,
|
|
)) => {
|
|
// signature generation is non-deterministic, so we can't just assert that the
|
|
// expected descriptor is correct. What we can do is validate that the produced
|
|
// descriptor has a valid signature, then just copy in the generated signature
|
|
// and check the rest of the fields for equality.
|
|
assert!(CollatorPair::verify(
|
|
&descriptor.signature,
|
|
&collator_signature_payload(
|
|
&descriptor.relay_parent,
|
|
&descriptor.para_id,
|
|
&descriptor.persisted_validation_data_hash,
|
|
&descriptor.pov_hash,
|
|
)
|
|
.as_ref(),
|
|
&descriptor.collator,
|
|
));
|
|
let expect_descriptor = {
|
|
let mut expect_descriptor = expect_descriptor;
|
|
expect_descriptor.signature = descriptor.signature.clone();
|
|
expect_descriptor.erasure_root = descriptor.erasure_root.clone();
|
|
expect_descriptor
|
|
};
|
|
assert_eq!(descriptor, &expect_descriptor);
|
|
}
|
|
_ => panic!("received wrong message type"),
|
|
}
|
|
}
|
|
}
|
|
}
|