mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-30 01:27:56 +00:00
b832e35c5e
* WIP * Making progress * Almost ready * Get service tests compiling * Fix node screenshot * Line widths * Fix node cli tests * Fix node cli warning * ChainComponents -> ServiceComponents, fix tests * make spawn_handle public * Remove spawnnamed impl for taskmanager * Move the keep alive stuff to the task manager * Move the telemetry, base path, rpc keep_alive to the service builder * Make the task manager keep alive an internal detail * Rewrite the browser start_client future * Remove run_node etc * Revert my personal changes to browser-demo/build.sh * use |config| * Add a runtime_version function to SubstrateCli * Reexport role and runtime version from sc cli * Update Cargo.lock * runtime_version -> native_runtime_version * Pass chain spec to native_runtime_version for polkadot * Fix line widths * Traitify ServiceComponents Client
647 lines
20 KiB
Rust
647 lines
20 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) 2018-2020 Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
//! Service integration test utils.
|
|
|
|
use std::iter;
|
|
use std::sync::Arc;
|
|
use std::net::Ipv4Addr;
|
|
use std::pin::Pin;
|
|
use std::time::Duration;
|
|
use log::{info, debug};
|
|
use futures01::{Future, Stream, Poll};
|
|
use futures::{FutureExt as _, TryFutureExt as _};
|
|
use tempfile::TempDir;
|
|
use tokio::{runtime::Runtime, prelude::FutureExt};
|
|
use tokio::timer::Interval;
|
|
use sc_service::{
|
|
TaskManager,
|
|
GenericChainSpec,
|
|
ChainSpecExtension,
|
|
Configuration,
|
|
config::{BasePath, DatabaseConfig, KeystoreConfig},
|
|
RuntimeGenesis,
|
|
Role,
|
|
Error,
|
|
TaskExecutor,
|
|
client::Client,
|
|
};
|
|
use sp_blockchain::HeaderBackend;
|
|
use sc_network::{multiaddr, Multiaddr};
|
|
use sc_network::config::{NetworkConfiguration, TransportConfig};
|
|
use sp_runtime::{generic::BlockId, traits::Block as BlockT};
|
|
use sp_transaction_pool::TransactionPool;
|
|
use sc_client_api::{Backend, CallExecutor};
|
|
use parking_lot::Mutex;
|
|
|
|
#[cfg(test)]
|
|
mod client;
|
|
|
|
/// Maximum duration of single wait call.
|
|
const MAX_WAIT_TIME: Duration = Duration::from_secs(60 * 3);
|
|
|
|
struct TestNet<G, E, F, L, U> {
|
|
runtime: Runtime,
|
|
authority_nodes: Vec<(usize, F, U, Multiaddr)>,
|
|
full_nodes: Vec<(usize, F, U, Multiaddr)>,
|
|
light_nodes: Vec<(usize, L, Multiaddr)>,
|
|
chain_spec: GenericChainSpec<G, E>,
|
|
base_port: u16,
|
|
nodes: usize,
|
|
}
|
|
|
|
pub trait TestNetNode: Clone + Future<Item = (), Error = sc_service::Error> + Send + 'static {
|
|
type Block: BlockT;
|
|
type Backend: Backend<Self::Block>;
|
|
type Executor: CallExecutor<Self::Block> + Send + Sync;
|
|
type RuntimeApi: Send + Sync;
|
|
type TransactionPool: TransactionPool<Block = Self::Block>;
|
|
|
|
fn client(&self) -> Arc<Client<Self::Backend, Self::Executor, Self::Block, Self::RuntimeApi>>;
|
|
fn transaction_pool(&self) -> Arc<Self::TransactionPool>;
|
|
fn network(&self) -> Arc<sc_network::NetworkService<Self::Block, <Self::Block as BlockT>::Hash>>;
|
|
}
|
|
|
|
pub struct TestNetComponents<TBl: BlockT, TBackend, TExec, TRtApi, TExPool> {
|
|
task_manager: Arc<Mutex<TaskManager>>,
|
|
client: Arc<Client<TBackend, TExec, TBl, TRtApi>>,
|
|
transaction_pool: Arc<TExPool>,
|
|
network: Arc<sc_network::NetworkService<TBl, <TBl as BlockT>::Hash>>,
|
|
}
|
|
|
|
impl<TBl: BlockT, TBackend, TExec, TRtApi, TExPool>
|
|
TestNetComponents<TBl, TBackend, TExec, TRtApi, TExPool> {
|
|
pub fn new(
|
|
task_manager: TaskManager,
|
|
client: Arc<Client<TBackend, TExec, TBl, TRtApi>>,
|
|
network: Arc<sc_network::NetworkService<TBl, <TBl as BlockT>::Hash>>,
|
|
transaction_pool: Arc<TExPool>,
|
|
) -> Self {
|
|
Self {
|
|
client, transaction_pool, network,
|
|
task_manager: Arc::new(Mutex::new(task_manager)),
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
impl<TBl: BlockT, TBackend, TExec, TRtApi, TExPool> Clone for
|
|
TestNetComponents<TBl, TBackend, TExec, TRtApi, TExPool> {
|
|
fn clone(&self) -> Self {
|
|
Self {
|
|
task_manager: self.task_manager.clone(),
|
|
client: self.client.clone(),
|
|
transaction_pool: self.transaction_pool.clone(),
|
|
network: self.network.clone(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<TBl: BlockT, TBackend, TExec, TRtApi, TExPool> Future for
|
|
TestNetComponents<TBl, TBackend, TExec, TRtApi, TExPool>
|
|
{
|
|
type Item = ();
|
|
type Error = sc_service::Error;
|
|
|
|
fn poll(&mut self) -> Poll<Self::Item, Self::Error> {
|
|
futures::compat::Compat::new(&mut self.task_manager.lock().future()).poll()
|
|
}
|
|
}
|
|
|
|
impl<TBl, TBackend, TExec, TRtApi, TExPool> TestNetNode for
|
|
TestNetComponents<TBl, TBackend, TExec, TRtApi, TExPool>
|
|
where
|
|
TBl: BlockT,
|
|
TBackend: sc_client_api::Backend<TBl> + Send + Sync + 'static,
|
|
TExec: CallExecutor<TBl> + Send + Sync + 'static,
|
|
TRtApi: Send + Sync + 'static,
|
|
TExPool: TransactionPool<Block = TBl> + Send + Sync + 'static,
|
|
{
|
|
type Block = TBl;
|
|
type Backend = TBackend;
|
|
type Executor = TExec;
|
|
type RuntimeApi = TRtApi;
|
|
type TransactionPool = TExPool;
|
|
|
|
fn client(&self) -> Arc<Client<Self::Backend, Self::Executor, Self::Block, Self::RuntimeApi>> {
|
|
self.client.clone()
|
|
}
|
|
fn transaction_pool(&self) -> Arc<Self::TransactionPool> {
|
|
self.transaction_pool.clone()
|
|
}
|
|
fn network(&self) -> Arc<sc_network::NetworkService<Self::Block, <Self::Block as BlockT>::Hash>> {
|
|
self.network.clone()
|
|
}
|
|
}
|
|
|
|
impl<G, E, F, L, U> TestNet<G, E, F, L, U>
|
|
where F: Clone + Send + 'static, L: Clone + Send +'static, U: Clone + Send + 'static
|
|
{
|
|
pub fn run_until_all_full<FP, LP>(
|
|
&mut self,
|
|
full_predicate: FP,
|
|
light_predicate: LP,
|
|
)
|
|
where
|
|
FP: Send + Fn(usize, &F) -> bool + 'static,
|
|
LP: Send + Fn(usize, &L) -> bool + 'static,
|
|
{
|
|
let full_nodes = self.full_nodes.clone();
|
|
let light_nodes = self.light_nodes.clone();
|
|
let interval = Interval::new_interval(Duration::from_millis(100))
|
|
.map_err(|_| ())
|
|
.for_each(move |_| {
|
|
let full_ready = full_nodes.iter().all(|&(ref id, ref service, _, _)|
|
|
full_predicate(*id, service)
|
|
);
|
|
|
|
if !full_ready {
|
|
return Ok(());
|
|
}
|
|
|
|
let light_ready = light_nodes.iter().all(|&(ref id, ref service, _)|
|
|
light_predicate(*id, service)
|
|
);
|
|
|
|
if !light_ready {
|
|
Ok(())
|
|
} else {
|
|
Err(())
|
|
}
|
|
})
|
|
.timeout(MAX_WAIT_TIME);
|
|
|
|
match self.runtime.block_on(interval) {
|
|
Ok(()) => unreachable!("interval always fails; qed"),
|
|
Err(ref err) if err.is_inner() => (),
|
|
Err(_) => panic!("Waited for too long"),
|
|
}
|
|
}
|
|
}
|
|
|
|
fn node_config<G: RuntimeGenesis + 'static, E: ChainSpecExtension + Clone + 'static + Send> (
|
|
index: usize,
|
|
spec: &GenericChainSpec<G, E>,
|
|
role: Role,
|
|
task_executor: TaskExecutor,
|
|
key_seed: Option<String>,
|
|
base_port: u16,
|
|
root: &TempDir,
|
|
) -> Configuration
|
|
{
|
|
let root = root.path().join(format!("node-{}", index));
|
|
|
|
let mut network_config = NetworkConfiguration::new(
|
|
format!("Node {}", index),
|
|
"network/test/0.1",
|
|
Default::default(),
|
|
None,
|
|
);
|
|
|
|
network_config.allow_non_globals_in_dht = true;
|
|
|
|
network_config.listen_addresses.push(
|
|
iter::once(multiaddr::Protocol::Ip4(Ipv4Addr::new(127, 0, 0, 1)))
|
|
.chain(iter::once(multiaddr::Protocol::Tcp(base_port + index as u16)))
|
|
.collect()
|
|
);
|
|
|
|
network_config.transport = TransportConfig::Normal {
|
|
enable_mdns: false,
|
|
allow_private_ipv4: true,
|
|
wasm_external_transport: None,
|
|
use_yamux_flow_control: true,
|
|
};
|
|
|
|
Configuration {
|
|
impl_name: "network-test-impl",
|
|
impl_version: "0.1",
|
|
role,
|
|
task_executor,
|
|
transaction_pool: Default::default(),
|
|
network: network_config,
|
|
keystore: KeystoreConfig::Path {
|
|
path: root.join("key"),
|
|
password: None
|
|
},
|
|
database: DatabaseConfig::RocksDb {
|
|
path: root.join("db"),
|
|
cache_size: 128,
|
|
},
|
|
state_cache_size: 16777216,
|
|
state_cache_child_ratio: None,
|
|
pruning: Default::default(),
|
|
chain_spec: Box::new((*spec).clone()),
|
|
wasm_method: sc_service::config::WasmExecutionMethod::Interpreted,
|
|
execution_strategies: Default::default(),
|
|
rpc_http: None,
|
|
rpc_ipc: None,
|
|
rpc_ws: None,
|
|
rpc_ws_max_connections: None,
|
|
rpc_cors: None,
|
|
rpc_methods: Default::default(),
|
|
prometheus_config: None,
|
|
telemetry_endpoints: None,
|
|
telemetry_external_transport: None,
|
|
default_heap_pages: None,
|
|
offchain_worker: Default::default(),
|
|
force_authoring: false,
|
|
disable_grandpa: false,
|
|
dev_key_seed: key_seed,
|
|
tracing_targets: None,
|
|
tracing_receiver: Default::default(),
|
|
max_runtime_instances: 8,
|
|
announce_block: true,
|
|
base_path: Some(BasePath::new(root)),
|
|
informant_output_format: Default::default(),
|
|
}
|
|
}
|
|
|
|
impl<G, E, F, L, U> TestNet<G, E, F, L, U> where
|
|
F: TestNetNode,
|
|
L: TestNetNode,
|
|
E: ChainSpecExtension + Clone + 'static + Send,
|
|
G: RuntimeGenesis + 'static,
|
|
{
|
|
fn new(
|
|
temp: &TempDir,
|
|
spec: GenericChainSpec<G, E>,
|
|
full: impl Iterator<Item = impl FnOnce(Configuration) -> Result<(F, U), Error>>,
|
|
light: impl Iterator<Item = impl FnOnce(Configuration) -> Result<L, Error>>,
|
|
authorities: impl Iterator<Item = (
|
|
String,
|
|
impl FnOnce(Configuration) -> Result<(F, U), Error>
|
|
)>,
|
|
base_port: u16
|
|
) -> TestNet<G, E, F, L, U> {
|
|
let _ = env_logger::try_init();
|
|
fdlimit::raise_fd_limit();
|
|
let runtime = Runtime::new().expect("Error creating tokio runtime");
|
|
let mut net = TestNet {
|
|
runtime,
|
|
authority_nodes: Default::default(),
|
|
full_nodes: Default::default(),
|
|
light_nodes: Default::default(),
|
|
chain_spec: spec,
|
|
base_port,
|
|
nodes: 0,
|
|
};
|
|
net.insert_nodes(temp, full, light, authorities);
|
|
net
|
|
}
|
|
|
|
fn insert_nodes(
|
|
&mut self,
|
|
temp: &TempDir,
|
|
full: impl Iterator<Item = impl FnOnce(Configuration) -> Result<(F, U), Error>>,
|
|
light: impl Iterator<Item = impl FnOnce(Configuration) -> Result<L, Error>>,
|
|
authorities: impl Iterator<Item = (String, impl FnOnce(Configuration) -> Result<(F, U), Error>)>
|
|
) {
|
|
let executor = self.runtime.executor();
|
|
let task_executor: TaskExecutor = {
|
|
let executor = executor.clone();
|
|
(move |fut: Pin<Box<dyn futures::Future<Output = ()> + Send>>, _| {
|
|
executor.spawn(fut.unit_error().compat());
|
|
}).into()
|
|
};
|
|
|
|
for (key, authority) in authorities {
|
|
let node_config = node_config(
|
|
self.nodes,
|
|
&self.chain_spec,
|
|
Role::Authority { sentry_nodes: Vec::new() },
|
|
task_executor.clone(),
|
|
Some(key),
|
|
self.base_port,
|
|
&temp,
|
|
);
|
|
let addr = node_config.network.listen_addresses.iter().next().unwrap().clone();
|
|
let (service, user_data) = authority(node_config).expect("Error creating test node service");
|
|
|
|
executor.spawn(service.clone().map_err(|_| ()));
|
|
let addr = addr.with(multiaddr::Protocol::P2p(service.network().local_peer_id().clone().into()));
|
|
self.authority_nodes.push((self.nodes, service, user_data, addr));
|
|
self.nodes += 1;
|
|
}
|
|
|
|
for full in full {
|
|
let node_config = node_config(
|
|
self.nodes,
|
|
&self.chain_spec,
|
|
Role::Full,
|
|
task_executor.clone(),
|
|
None,
|
|
self.base_port,
|
|
&temp,
|
|
);
|
|
let addr = node_config.network.listen_addresses.iter().next().unwrap().clone();
|
|
let (service, user_data) = full(node_config).expect("Error creating test node service");
|
|
|
|
executor.spawn(service.clone().map_err(|_| ()));
|
|
let addr = addr.with(multiaddr::Protocol::P2p(service.network().local_peer_id().clone().into()));
|
|
self.full_nodes.push((self.nodes, service, user_data, addr));
|
|
self.nodes += 1;
|
|
}
|
|
|
|
for light in light {
|
|
let node_config = node_config(
|
|
self.nodes,
|
|
&self.chain_spec,
|
|
Role::Light,
|
|
task_executor.clone(),
|
|
None,
|
|
self.base_port,
|
|
&temp,
|
|
);
|
|
let addr = node_config.network.listen_addresses.iter().next().unwrap().clone();
|
|
let service = light(node_config).expect("Error creating test node service");
|
|
|
|
executor.spawn(service.clone().map_err(|_| ()));
|
|
let addr = addr.with(multiaddr::Protocol::P2p(service.network().local_peer_id().clone().into()));
|
|
self.light_nodes.push((self.nodes, service, addr));
|
|
self.nodes += 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
fn tempdir_with_prefix(prefix: &str) -> TempDir {
|
|
tempfile::Builder::new().prefix(prefix).tempdir().expect("Error creating test dir")
|
|
}
|
|
|
|
pub fn connectivity<G, E, Fb, F, Lb, L>(
|
|
spec: GenericChainSpec<G, E>,
|
|
full_builder: Fb,
|
|
light_builder: Lb,
|
|
) where
|
|
E: ChainSpecExtension + Clone + 'static + Send,
|
|
G: RuntimeGenesis + 'static,
|
|
Fb: Fn(Configuration) -> Result<F, Error>,
|
|
F: TestNetNode,
|
|
Lb: Fn(Configuration) -> Result<L, Error>,
|
|
L: TestNetNode,
|
|
{
|
|
const NUM_FULL_NODES: usize = 5;
|
|
const NUM_LIGHT_NODES: usize = 5;
|
|
|
|
let expected_full_connections = NUM_FULL_NODES - 1 + NUM_LIGHT_NODES;
|
|
let expected_light_connections = NUM_FULL_NODES;
|
|
|
|
{
|
|
let temp = tempdir_with_prefix("substrate-connectivity-test");
|
|
let runtime = {
|
|
let mut network = TestNet::new(
|
|
&temp,
|
|
spec.clone(),
|
|
(0..NUM_FULL_NODES).map(|_| { |cfg| full_builder(cfg).map(|s| (s, ())) }),
|
|
(0..NUM_LIGHT_NODES).map(|_| { |cfg| light_builder(cfg) }),
|
|
// Note: this iterator is empty but we can't just use `iter::empty()`, otherwise
|
|
// the type of the closure cannot be inferred.
|
|
(0..0).map(|_| (String::new(), { |cfg| full_builder(cfg).map(|s| (s, ())) })),
|
|
30400,
|
|
);
|
|
info!("Checking star topology");
|
|
let first_address = network.full_nodes[0].3.clone();
|
|
for (_, service, _, _) in network.full_nodes.iter().skip(1) {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
for (_, service, _) in network.light_nodes.iter() {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
|
|
network.run_until_all_full(
|
|
move |_index, service| {
|
|
let connected = service.network().num_connected();
|
|
debug!("Got {}/{} full connections...", connected, expected_full_connections);
|
|
connected == expected_full_connections
|
|
},
|
|
move |_index, service| {
|
|
let connected = service.network().num_connected();
|
|
debug!("Got {}/{} light connections...", connected, expected_light_connections);
|
|
connected == expected_light_connections
|
|
},
|
|
);
|
|
|
|
network.runtime
|
|
};
|
|
|
|
runtime.shutdown_now().wait().expect("Error shutting down runtime");
|
|
|
|
temp.close().expect("Error removing temp dir");
|
|
}
|
|
{
|
|
let temp = tempdir_with_prefix("substrate-connectivity-test");
|
|
{
|
|
let mut network = TestNet::new(
|
|
&temp,
|
|
spec,
|
|
(0..NUM_FULL_NODES).map(|_| { |cfg| full_builder(cfg).map(|s| (s, ())) }),
|
|
(0..NUM_LIGHT_NODES).map(|_| { |cfg| light_builder(cfg) }),
|
|
// Note: this iterator is empty but we can't just use `iter::empty()`, otherwise
|
|
// the type of the closure cannot be inferred.
|
|
(0..0).map(|_| (String::new(), { |cfg| full_builder(cfg).map(|s| (s, ())) })),
|
|
30400,
|
|
);
|
|
info!("Checking linked topology");
|
|
let mut address = network.full_nodes[0].3.clone();
|
|
let max_nodes = std::cmp::max(NUM_FULL_NODES, NUM_LIGHT_NODES);
|
|
for i in 0..max_nodes {
|
|
if i != 0 {
|
|
if let Some((_, service, _, node_id)) = network.full_nodes.get(i) {
|
|
service.network().add_reserved_peer(address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
address = node_id.clone();
|
|
}
|
|
}
|
|
|
|
if let Some((_, service, node_id)) = network.light_nodes.get(i) {
|
|
service.network().add_reserved_peer(address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
address = node_id.clone();
|
|
}
|
|
}
|
|
|
|
network.run_until_all_full(
|
|
move |_index, service| {
|
|
let connected = service.network().num_connected();
|
|
debug!("Got {}/{} full connections...", connected, expected_full_connections);
|
|
connected == expected_full_connections
|
|
},
|
|
move |_index, service| {
|
|
let connected = service.network().num_connected();
|
|
debug!("Got {}/{} light connections...", connected, expected_light_connections);
|
|
connected == expected_light_connections
|
|
},
|
|
);
|
|
}
|
|
temp.close().expect("Error removing temp dir");
|
|
}
|
|
}
|
|
|
|
pub fn sync<G, E, Fb, F, Lb, L, B, ExF, U>(
|
|
spec: GenericChainSpec<G, E>,
|
|
full_builder: Fb,
|
|
light_builder: Lb,
|
|
mut make_block_and_import: B,
|
|
mut extrinsic_factory: ExF
|
|
) where
|
|
Fb: Fn(Configuration) -> Result<(F, U), Error>,
|
|
F: TestNetNode,
|
|
Lb: Fn(Configuration) -> Result<L, Error>,
|
|
L: TestNetNode,
|
|
B: FnMut(&F, &mut U),
|
|
ExF: FnMut(&F, &U) -> <F::Block as BlockT>::Extrinsic,
|
|
U: Clone + Send + 'static,
|
|
E: ChainSpecExtension + Clone + 'static + Send,
|
|
G: RuntimeGenesis + 'static,
|
|
{
|
|
const NUM_FULL_NODES: usize = 10;
|
|
// FIXME: BABE light client support is currently not working.
|
|
const NUM_LIGHT_NODES: usize = 10;
|
|
const NUM_BLOCKS: usize = 512;
|
|
let temp = tempdir_with_prefix("substrate-sync-test");
|
|
let mut network = TestNet::new(
|
|
&temp,
|
|
spec.clone(),
|
|
(0..NUM_FULL_NODES).map(|_| { |cfg| full_builder(cfg) }),
|
|
(0..NUM_LIGHT_NODES).map(|_| { |cfg| light_builder(cfg) }),
|
|
// Note: this iterator is empty but we can't just use `iter::empty()`, otherwise
|
|
// the type of the closure cannot be inferred.
|
|
(0..0).map(|_| (String::new(), { |cfg| full_builder(cfg) })),
|
|
30500,
|
|
);
|
|
info!("Checking block sync");
|
|
let first_address = {
|
|
let &mut (_, ref first_service, ref mut first_user_data, _) = &mut network.full_nodes[0];
|
|
for i in 0 .. NUM_BLOCKS {
|
|
if i % 128 == 0 {
|
|
info!("Generating #{}", i + 1);
|
|
}
|
|
|
|
make_block_and_import(&first_service, first_user_data);
|
|
}
|
|
network.full_nodes[0].1.network().update_chain();
|
|
network.full_nodes[0].3.clone()
|
|
};
|
|
|
|
info!("Running sync");
|
|
for (_, service, _, _) in network.full_nodes.iter().skip(1) {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
for (_, service, _) in network.light_nodes.iter() {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
network.run_until_all_full(
|
|
|_index, service|
|
|
service.client().info().best_number == (NUM_BLOCKS as u32).into(),
|
|
|_index, service|
|
|
service.client().info().best_number == (NUM_BLOCKS as u32).into(),
|
|
);
|
|
|
|
info!("Checking extrinsic propagation");
|
|
let first_service = network.full_nodes[0].1.clone();
|
|
let first_user_data = &network.full_nodes[0].2;
|
|
let best_block = BlockId::number(first_service.client().info().best_number);
|
|
let extrinsic = extrinsic_factory(&first_service, first_user_data);
|
|
let source = sp_transaction_pool::TransactionSource::External;
|
|
|
|
futures::executor::block_on(
|
|
first_service.transaction_pool().submit_one(&best_block, source, extrinsic)
|
|
).expect("failed to submit extrinsic");
|
|
|
|
network.run_until_all_full(
|
|
|_index, service| service.transaction_pool().ready().count() == 1,
|
|
|_index, _service| true,
|
|
);
|
|
}
|
|
|
|
pub fn consensus<G, E, Fb, F, Lb, L>(
|
|
spec: GenericChainSpec<G, E>,
|
|
full_builder: Fb,
|
|
light_builder: Lb,
|
|
authorities: impl IntoIterator<Item = String>
|
|
) where
|
|
Fb: Fn(Configuration) -> Result<F, Error>,
|
|
F: TestNetNode,
|
|
Lb: Fn(Configuration) -> Result<L, Error>,
|
|
L: TestNetNode,
|
|
E: ChainSpecExtension + Clone + 'static + Send,
|
|
G: RuntimeGenesis + 'static,
|
|
{
|
|
const NUM_FULL_NODES: usize = 10;
|
|
const NUM_LIGHT_NODES: usize = 10;
|
|
const NUM_BLOCKS: usize = 10; // 10 * 2 sec block production time = ~20 seconds
|
|
let temp = tempdir_with_prefix("substrate-consensus-test");
|
|
let mut network = TestNet::new(
|
|
&temp,
|
|
spec.clone(),
|
|
(0..NUM_FULL_NODES / 2).map(|_| { |cfg| full_builder(cfg).map(|s| (s, ())) }),
|
|
(0..NUM_LIGHT_NODES / 2).map(|_| { |cfg| light_builder(cfg) }),
|
|
authorities.into_iter().map(|key| (key, { |cfg| full_builder(cfg).map(|s| (s, ())) })),
|
|
30600,
|
|
);
|
|
|
|
info!("Checking consensus");
|
|
let first_address = network.authority_nodes[0].3.clone();
|
|
for (_, service, _, _) in network.full_nodes.iter() {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
for (_, service, _) in network.light_nodes.iter() {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
for (_, service, _, _) in network.authority_nodes.iter().skip(1) {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
network.run_until_all_full(
|
|
|_index, service|
|
|
service.client().info().finalized_number >= (NUM_BLOCKS as u32 / 2).into(),
|
|
|_index, service|
|
|
service.client().info().best_number >= (NUM_BLOCKS as u32 / 2).into(),
|
|
);
|
|
|
|
info!("Adding more peers");
|
|
network.insert_nodes(
|
|
&temp,
|
|
(0..NUM_FULL_NODES / 2).map(|_| { |cfg| full_builder(cfg).map(|s| (s, ())) }),
|
|
(0..NUM_LIGHT_NODES / 2).map(|_| { |cfg| light_builder(cfg) }),
|
|
// Note: this iterator is empty but we can't just use `iter::empty()`, otherwise
|
|
// the type of the closure cannot be inferred.
|
|
(0..0).map(|_| (String::new(), { |cfg| full_builder(cfg).map(|s| (s, ())) })),
|
|
);
|
|
for (_, service, _, _) in network.full_nodes.iter() {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
for (_, service, _) in network.light_nodes.iter() {
|
|
service.network().add_reserved_peer(first_address.to_string())
|
|
.expect("Error adding reserved peer");
|
|
}
|
|
network.run_until_all_full(
|
|
|_index, service|
|
|
service.client().info().finalized_number >= (NUM_BLOCKS as u32).into(),
|
|
|_index, service|
|
|
service.client().info().best_number >= (NUM_BLOCKS as u32).into(),
|
|
);
|
|
}
|