mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-05-06 20:48:02 +00:00
ecf8035da6
* Yep * Try to get it working everywhere * Make `from_raw_storage` start with an empty db * More fixes! * Make everything compile * Fix `child_storage_root` * Fix after merge * Cleanups * Update primitives/state-machine/src/overlayed_changes/mod.rs Co-authored-by: Davide Galassi <davxy@datawok.net> * Review comments * Fix issues * Silence warning * FMT * Clippy --------- Co-authored-by: Davide Galassi <davxy@datawok.net>
822 lines
24 KiB
Rust
822 lines
24 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
use futures::prelude::*;
|
|
use libp2p::{Multiaddr, PeerId};
|
|
|
|
use sc_consensus::{ImportQueue, Link};
|
|
use sc_network::{
|
|
config::{self, FullNetworkConfiguration, MultiaddrWithPeerId, ProtocolId, TransportConfig},
|
|
event::Event,
|
|
peer_store::PeerStore,
|
|
NetworkEventStream, NetworkNotification, NetworkPeers, NetworkService, NetworkStateInfo,
|
|
NetworkWorker,
|
|
};
|
|
use sc_network_common::role::Roles;
|
|
use sc_network_light::light_client_requests::handler::LightClientRequestHandler;
|
|
use sc_network_sync::{
|
|
block_request_handler::BlockRequestHandler,
|
|
engine::SyncingEngine,
|
|
service::network::{NetworkServiceHandle, NetworkServiceProvider},
|
|
state_request_handler::StateRequestHandler,
|
|
};
|
|
use sp_blockchain::HeaderBackend;
|
|
use sp_runtime::traits::{Block as BlockT, Zero};
|
|
use substrate_test_runtime_client::{
|
|
runtime::{Block as TestBlock, Hash as TestHash},
|
|
TestClientBuilder, TestClientBuilderExt as _,
|
|
};
|
|
|
|
use std::{sync::Arc, time::Duration};
|
|
|
|
type TestNetworkWorker = NetworkWorker<TestBlock, TestHash>;
|
|
type TestNetworkService = NetworkService<TestBlock, TestHash>;
|
|
|
|
const PROTOCOL_NAME: &str = "/foo";
|
|
|
|
struct TestNetwork {
|
|
network: TestNetworkWorker,
|
|
}
|
|
|
|
impl TestNetwork {
|
|
pub fn new(network: TestNetworkWorker) -> Self {
|
|
Self { network }
|
|
}
|
|
|
|
pub fn start_network(
|
|
self,
|
|
) -> (Arc<TestNetworkService>, (impl Stream<Item = Event> + std::marker::Unpin)) {
|
|
let worker = self.network;
|
|
let service = worker.service().clone();
|
|
let event_stream = service.event_stream("test");
|
|
|
|
tokio::spawn(worker.run());
|
|
|
|
(service, event_stream)
|
|
}
|
|
}
|
|
|
|
struct TestNetworkBuilder {
|
|
import_queue: Option<Box<dyn ImportQueue<TestBlock>>>,
|
|
link: Option<Box<dyn Link<TestBlock>>>,
|
|
client: Option<Arc<substrate_test_runtime_client::TestClient>>,
|
|
listen_addresses: Vec<Multiaddr>,
|
|
set_config: Option<config::SetConfig>,
|
|
chain_sync_network: Option<(NetworkServiceProvider, NetworkServiceHandle)>,
|
|
notification_protocols: Vec<config::NonDefaultSetConfig>,
|
|
config: Option<config::NetworkConfiguration>,
|
|
}
|
|
|
|
impl TestNetworkBuilder {
|
|
pub fn new() -> Self {
|
|
Self {
|
|
import_queue: None,
|
|
link: None,
|
|
client: None,
|
|
listen_addresses: Vec::new(),
|
|
set_config: None,
|
|
chain_sync_network: None,
|
|
notification_protocols: Vec::new(),
|
|
config: None,
|
|
}
|
|
}
|
|
|
|
pub fn with_config(mut self, config: config::NetworkConfiguration) -> Self {
|
|
self.config = Some(config);
|
|
self
|
|
}
|
|
|
|
pub fn with_notification_protocol(mut self, config: config::NonDefaultSetConfig) -> Self {
|
|
self.notification_protocols.push(config);
|
|
self
|
|
}
|
|
|
|
pub fn with_listen_addresses(mut self, addresses: Vec<Multiaddr>) -> Self {
|
|
self.listen_addresses = addresses;
|
|
self
|
|
}
|
|
|
|
pub fn with_set_config(mut self, set_config: config::SetConfig) -> Self {
|
|
self.set_config = Some(set_config);
|
|
self
|
|
}
|
|
|
|
pub fn build(mut self) -> TestNetwork {
|
|
let client = self.client.as_mut().map_or(
|
|
Arc::new(TestClientBuilder::with_default_backend().build_with_longest_chain().0),
|
|
|v| v.clone(),
|
|
);
|
|
|
|
let network_config = self.config.unwrap_or(config::NetworkConfiguration {
|
|
listen_addresses: self.listen_addresses,
|
|
transport: TransportConfig::MemoryOnly,
|
|
..config::NetworkConfiguration::new_local()
|
|
});
|
|
|
|
#[derive(Clone)]
|
|
struct PassThroughVerifier(bool);
|
|
|
|
#[async_trait::async_trait]
|
|
impl<B: BlockT> sc_consensus::Verifier<B> for PassThroughVerifier {
|
|
async fn verify(
|
|
&mut self,
|
|
mut block: sc_consensus::BlockImportParams<B>,
|
|
) -> Result<sc_consensus::BlockImportParams<B>, String> {
|
|
block.finalized = self.0;
|
|
block.fork_choice = Some(sc_consensus::ForkChoiceStrategy::LongestChain);
|
|
Ok(block)
|
|
}
|
|
}
|
|
|
|
let mut import_queue =
|
|
self.import_queue.unwrap_or(Box::new(sc_consensus::BasicQueue::new(
|
|
PassThroughVerifier(false),
|
|
Box::new(client.clone()),
|
|
None,
|
|
&sp_core::testing::TaskExecutor::new(),
|
|
None,
|
|
)));
|
|
|
|
let protocol_id = ProtocolId::from("test-protocol-name");
|
|
let fork_id = Some(String::from("test-fork-id"));
|
|
let mut full_net_config = FullNetworkConfiguration::new(&network_config);
|
|
|
|
let block_request_protocol_config = {
|
|
let (handler, protocol_config) =
|
|
BlockRequestHandler::new(&protocol_id, None, client.clone(), 50);
|
|
tokio::spawn(handler.run().boxed());
|
|
protocol_config
|
|
};
|
|
|
|
let state_request_protocol_config = {
|
|
let (handler, protocol_config) =
|
|
StateRequestHandler::new(&protocol_id, None, client.clone(), 50);
|
|
tokio::spawn(handler.run().boxed());
|
|
protocol_config
|
|
};
|
|
|
|
let light_client_request_protocol_config = {
|
|
let (handler, protocol_config) =
|
|
LightClientRequestHandler::new(&protocol_id, None, client.clone());
|
|
tokio::spawn(handler.run().boxed());
|
|
protocol_config
|
|
};
|
|
|
|
let (chain_sync_network_provider, chain_sync_network_handle) =
|
|
self.chain_sync_network.unwrap_or(NetworkServiceProvider::new());
|
|
let (tx, rx) = sc_utils::mpsc::tracing_unbounded("mpsc_syncing_engine_protocol", 100_000);
|
|
let (engine, chain_sync_service, block_announce_config) = SyncingEngine::new(
|
|
Roles::from(&config::Role::Full),
|
|
client.clone(),
|
|
None,
|
|
&full_net_config,
|
|
protocol_id.clone(),
|
|
&None,
|
|
Box::new(sp_consensus::block_validation::DefaultBlockAnnounceValidator),
|
|
None,
|
|
chain_sync_network_handle,
|
|
import_queue.service(),
|
|
block_request_protocol_config.name.clone(),
|
|
state_request_protocol_config.name.clone(),
|
|
None,
|
|
rx,
|
|
)
|
|
.unwrap();
|
|
let mut link = self.link.unwrap_or(Box::new(chain_sync_service.clone()));
|
|
|
|
if !self.notification_protocols.is_empty() {
|
|
for config in self.notification_protocols {
|
|
full_net_config.add_notification_protocol(config);
|
|
}
|
|
} else {
|
|
full_net_config.add_notification_protocol(config::NonDefaultSetConfig {
|
|
notifications_protocol: PROTOCOL_NAME.into(),
|
|
fallback_names: Vec::new(),
|
|
max_notification_size: 1024 * 1024,
|
|
handshake: None,
|
|
set_config: self.set_config.unwrap_or_default(),
|
|
});
|
|
}
|
|
|
|
for config in [
|
|
block_request_protocol_config,
|
|
state_request_protocol_config,
|
|
light_client_request_protocol_config,
|
|
] {
|
|
full_net_config.add_request_response_protocol(config);
|
|
}
|
|
|
|
let peer_store = PeerStore::new(
|
|
network_config.boot_nodes.iter().map(|bootnode| bootnode.peer_id).collect(),
|
|
);
|
|
let peer_store_handle = peer_store.handle();
|
|
tokio::spawn(peer_store.run().boxed());
|
|
|
|
let genesis_hash =
|
|
client.hash(Zero::zero()).ok().flatten().expect("Genesis block exists; qed");
|
|
let worker = NetworkWorker::<
|
|
substrate_test_runtime_client::runtime::Block,
|
|
substrate_test_runtime_client::runtime::Hash,
|
|
>::new(config::Params::<substrate_test_runtime_client::runtime::Block> {
|
|
block_announce_config,
|
|
role: config::Role::Full,
|
|
executor: Box::new(|f| {
|
|
tokio::spawn(f);
|
|
}),
|
|
genesis_hash,
|
|
network_config: full_net_config,
|
|
peer_store: peer_store_handle,
|
|
protocol_id,
|
|
fork_id,
|
|
metrics_registry: None,
|
|
tx,
|
|
})
|
|
.unwrap();
|
|
|
|
let service = worker.service().clone();
|
|
tokio::spawn(async move {
|
|
let _ = chain_sync_network_provider.run(service).await;
|
|
});
|
|
tokio::spawn(async move {
|
|
loop {
|
|
futures::future::poll_fn(|cx| {
|
|
import_queue.poll_actions(cx, &mut *link);
|
|
std::task::Poll::Ready(())
|
|
})
|
|
.await;
|
|
tokio::time::sleep(std::time::Duration::from_millis(250)).await;
|
|
}
|
|
});
|
|
tokio::spawn(engine.run());
|
|
|
|
TestNetwork::new(worker)
|
|
}
|
|
}
|
|
|
|
/// Builds two nodes and their associated events stream.
|
|
/// The nodes are connected together and have the `PROTOCOL_NAME` protocol registered.
|
|
fn build_nodes_one_proto() -> (
|
|
Arc<TestNetworkService>,
|
|
impl Stream<Item = Event>,
|
|
Arc<TestNetworkService>,
|
|
impl Stream<Item = Event>,
|
|
) {
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
|
|
let (node1, events_stream1) = TestNetworkBuilder::new()
|
|
.with_listen_addresses(vec![listen_addr.clone()])
|
|
.build()
|
|
.start_network();
|
|
|
|
let (node2, events_stream2) = TestNetworkBuilder::new()
|
|
.with_set_config(config::SetConfig {
|
|
reserved_nodes: vec![MultiaddrWithPeerId {
|
|
multiaddr: listen_addr,
|
|
peer_id: node1.local_peer_id(),
|
|
}],
|
|
..Default::default()
|
|
})
|
|
.build()
|
|
.start_network();
|
|
|
|
(node1, events_stream1, node2, events_stream2)
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn notifications_state_consistent() {
|
|
// Runs two nodes and ensures that events are propagated out of the API in a consistent
|
|
// correct order, which means no notification received on a closed substream.
|
|
|
|
let (node1, mut events_stream1, node2, mut events_stream2) = build_nodes_one_proto();
|
|
|
|
// Write some initial notifications that shouldn't get through.
|
|
for _ in 0..(rand::random::<u8>() % 5) {
|
|
node1.write_notification(
|
|
node2.local_peer_id(),
|
|
PROTOCOL_NAME.into(),
|
|
b"hello world".to_vec(),
|
|
);
|
|
}
|
|
for _ in 0..(rand::random::<u8>() % 5) {
|
|
node2.write_notification(
|
|
node1.local_peer_id(),
|
|
PROTOCOL_NAME.into(),
|
|
b"hello world".to_vec(),
|
|
);
|
|
}
|
|
|
|
// True if we have an active substream from node1 to node2.
|
|
let mut node1_to_node2_open = false;
|
|
// True if we have an active substream from node2 to node1.
|
|
let mut node2_to_node1_open = false;
|
|
// We stop the test after a certain number of iterations.
|
|
let mut iterations = 0;
|
|
// Safe guard because we don't want the test to pass if no substream has been open.
|
|
let mut something_happened = false;
|
|
|
|
loop {
|
|
iterations += 1;
|
|
if iterations >= 1_000 {
|
|
assert!(something_happened);
|
|
break
|
|
}
|
|
|
|
// Start by sending a notification from node1 to node2 and vice-versa. Part of the
|
|
// test consists in ensuring that notifications get ignored if the stream isn't open.
|
|
if rand::random::<u8>() % 5 >= 3 {
|
|
node1.write_notification(
|
|
node2.local_peer_id(),
|
|
PROTOCOL_NAME.into(),
|
|
b"hello world".to_vec(),
|
|
);
|
|
}
|
|
if rand::random::<u8>() % 5 >= 3 {
|
|
node2.write_notification(
|
|
node1.local_peer_id(),
|
|
PROTOCOL_NAME.into(),
|
|
b"hello world".to_vec(),
|
|
);
|
|
}
|
|
|
|
// Also randomly disconnect the two nodes from time to time.
|
|
if rand::random::<u8>() % 20 == 0 {
|
|
node1.disconnect_peer(node2.local_peer_id(), PROTOCOL_NAME.into());
|
|
}
|
|
if rand::random::<u8>() % 20 == 0 {
|
|
node2.disconnect_peer(node1.local_peer_id(), PROTOCOL_NAME.into());
|
|
}
|
|
|
|
// Grab next event from either `events_stream1` or `events_stream2`.
|
|
let next_event = {
|
|
let next1 = events_stream1.next();
|
|
let next2 = events_stream2.next();
|
|
// We also await on a small timer, otherwise it is possible for the test to wait
|
|
// forever while nothing at all happens on the network.
|
|
let continue_test = futures_timer::Delay::new(Duration::from_millis(20));
|
|
match future::select(future::select(next1, next2), continue_test).await {
|
|
future::Either::Left((future::Either::Left((Some(ev), _)), _)) =>
|
|
future::Either::Left(ev),
|
|
future::Either::Left((future::Either::Right((Some(ev), _)), _)) =>
|
|
future::Either::Right(ev),
|
|
future::Either::Right(_) => continue,
|
|
_ => break,
|
|
}
|
|
};
|
|
|
|
match next_event {
|
|
future::Either::Left(Event::NotificationStreamOpened { remote, protocol, .. }) =>
|
|
if protocol == PROTOCOL_NAME.into() {
|
|
something_happened = true;
|
|
assert!(!node1_to_node2_open);
|
|
node1_to_node2_open = true;
|
|
assert_eq!(remote, node2.local_peer_id());
|
|
},
|
|
future::Either::Right(Event::NotificationStreamOpened { remote, protocol, .. }) =>
|
|
if protocol == PROTOCOL_NAME.into() {
|
|
something_happened = true;
|
|
assert!(!node2_to_node1_open);
|
|
node2_to_node1_open = true;
|
|
assert_eq!(remote, node1.local_peer_id());
|
|
},
|
|
future::Either::Left(Event::NotificationStreamClosed { remote, protocol, .. }) =>
|
|
if protocol == PROTOCOL_NAME.into() {
|
|
assert!(node1_to_node2_open);
|
|
node1_to_node2_open = false;
|
|
assert_eq!(remote, node2.local_peer_id());
|
|
},
|
|
future::Either::Right(Event::NotificationStreamClosed { remote, protocol, .. }) =>
|
|
if protocol == PROTOCOL_NAME.into() {
|
|
assert!(node2_to_node1_open);
|
|
node2_to_node1_open = false;
|
|
assert_eq!(remote, node1.local_peer_id());
|
|
},
|
|
future::Either::Left(Event::NotificationsReceived { remote, .. }) => {
|
|
assert!(node1_to_node2_open);
|
|
assert_eq!(remote, node2.local_peer_id());
|
|
if rand::random::<u8>() % 5 >= 4 {
|
|
node1.write_notification(
|
|
node2.local_peer_id(),
|
|
PROTOCOL_NAME.into(),
|
|
b"hello world".to_vec(),
|
|
);
|
|
}
|
|
},
|
|
future::Either::Right(Event::NotificationsReceived { remote, .. }) => {
|
|
assert!(node2_to_node1_open);
|
|
assert_eq!(remote, node1.local_peer_id());
|
|
if rand::random::<u8>() % 5 >= 4 {
|
|
node2.write_notification(
|
|
node1.local_peer_id(),
|
|
PROTOCOL_NAME.into(),
|
|
b"hello world".to_vec(),
|
|
);
|
|
}
|
|
},
|
|
|
|
// Add new events here.
|
|
future::Either::Left(Event::Dht(_)) => {},
|
|
future::Either::Right(Event::Dht(_)) => {},
|
|
};
|
|
}
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn lots_of_incoming_peers_works() {
|
|
sp_tracing::try_init_simple();
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
|
|
let (main_node, _) = TestNetworkBuilder::new()
|
|
.with_listen_addresses(vec![listen_addr.clone()])
|
|
.with_set_config(config::SetConfig { in_peers: u32::MAX, ..Default::default() })
|
|
.build()
|
|
.start_network();
|
|
|
|
let main_node_peer_id = main_node.local_peer_id();
|
|
|
|
// We spawn background tasks and push them in this `Vec`. They will all be waited upon before
|
|
// this test ends.
|
|
let mut background_tasks_to_wait = Vec::new();
|
|
|
|
for _ in 0..32 {
|
|
let (_dialing_node, event_stream) = TestNetworkBuilder::new()
|
|
.with_set_config(config::SetConfig {
|
|
reserved_nodes: vec![MultiaddrWithPeerId {
|
|
multiaddr: listen_addr.clone(),
|
|
peer_id: main_node_peer_id,
|
|
}],
|
|
..Default::default()
|
|
})
|
|
.build()
|
|
.start_network();
|
|
|
|
background_tasks_to_wait.push(tokio::spawn(async move {
|
|
// Create a dummy timer that will "never" fire, and that will be overwritten when we
|
|
// actually need the timer. Using an Option would be technically cleaner, but it would
|
|
// make the code below way more complicated.
|
|
let mut timer = futures_timer::Delay::new(Duration::from_secs(3600 * 24 * 7)).fuse();
|
|
|
|
let mut event_stream = event_stream.fuse();
|
|
let mut sync_protocol_name = None;
|
|
loop {
|
|
futures::select! {
|
|
_ = timer => {
|
|
// Test succeeds when timer fires.
|
|
return;
|
|
}
|
|
ev = event_stream.next() => {
|
|
match ev.unwrap() {
|
|
Event::NotificationStreamOpened { protocol, remote, .. } => {
|
|
if let None = sync_protocol_name {
|
|
sync_protocol_name = Some(protocol.clone());
|
|
}
|
|
|
|
assert_eq!(remote, main_node_peer_id);
|
|
// Test succeeds after 5 seconds. This timer is here in order to
|
|
// detect a potential problem after opening.
|
|
timer = futures_timer::Delay::new(Duration::from_secs(5)).fuse();
|
|
}
|
|
Event::NotificationStreamClosed { protocol, .. } => {
|
|
if Some(protocol) != sync_protocol_name {
|
|
// Test failed.
|
|
panic!();
|
|
}
|
|
}
|
|
_ => {}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}));
|
|
}
|
|
|
|
future::join_all(background_tasks_to_wait).await;
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn notifications_back_pressure() {
|
|
// Node 1 floods node 2 with notifications. Random sleeps are done on node 2 to simulate the
|
|
// node being busy. We make sure that all notifications are received.
|
|
|
|
const TOTAL_NOTIFS: usize = 10_000;
|
|
|
|
let (node1, mut events_stream1, node2, mut events_stream2) = build_nodes_one_proto();
|
|
let node2_id = node2.local_peer_id();
|
|
|
|
let receiver = tokio::spawn(async move {
|
|
let mut received_notifications = 0;
|
|
let mut sync_protocol_name = None;
|
|
|
|
while received_notifications < TOTAL_NOTIFS {
|
|
match events_stream2.next().await.unwrap() {
|
|
Event::NotificationStreamOpened { protocol, .. } => {
|
|
if let None = sync_protocol_name {
|
|
sync_protocol_name = Some(protocol);
|
|
}
|
|
},
|
|
Event::NotificationStreamClosed { protocol, .. } => {
|
|
if Some(&protocol) != sync_protocol_name.as_ref() {
|
|
panic!()
|
|
}
|
|
},
|
|
Event::NotificationsReceived { messages, .. } =>
|
|
for message in messages {
|
|
assert_eq!(message.0, PROTOCOL_NAME.into());
|
|
assert_eq!(message.1, format!("hello #{}", received_notifications));
|
|
received_notifications += 1;
|
|
},
|
|
_ => {},
|
|
};
|
|
|
|
if rand::random::<u8>() < 2 {
|
|
tokio::time::sleep(Duration::from_millis(rand::random::<u64>() % 750)).await;
|
|
}
|
|
}
|
|
});
|
|
|
|
// Wait for the `NotificationStreamOpened`.
|
|
loop {
|
|
match events_stream1.next().await.unwrap() {
|
|
Event::NotificationStreamOpened { .. } => break,
|
|
_ => {},
|
|
};
|
|
}
|
|
|
|
// Sending!
|
|
for num in 0..TOTAL_NOTIFS {
|
|
let notif = node1.notification_sender(node2_id, PROTOCOL_NAME.into()).unwrap();
|
|
notif
|
|
.ready()
|
|
.await
|
|
.unwrap()
|
|
.send(format!("hello #{}", num).into_bytes())
|
|
.unwrap();
|
|
}
|
|
|
|
receiver.await.unwrap();
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn fallback_name_working() {
|
|
// Node 1 supports the protocols "new" and "old". Node 2 only supports "old". Checks whether
|
|
// they can connect.
|
|
const NEW_PROTOCOL_NAME: &str = "/new-shiny-protocol-that-isnt-PROTOCOL_NAME";
|
|
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
let (node1, mut events_stream1) = TestNetworkBuilder::new()
|
|
.with_notification_protocol(config::NonDefaultSetConfig {
|
|
notifications_protocol: NEW_PROTOCOL_NAME.into(),
|
|
fallback_names: vec![PROTOCOL_NAME.into()],
|
|
max_notification_size: 1024 * 1024,
|
|
handshake: None,
|
|
set_config: Default::default(),
|
|
})
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
transport: TransportConfig::MemoryOnly,
|
|
..config::NetworkConfiguration::new_local()
|
|
})
|
|
.build()
|
|
.start_network();
|
|
|
|
let (_, mut events_stream2) = TestNetworkBuilder::new()
|
|
.with_set_config(config::SetConfig {
|
|
reserved_nodes: vec![MultiaddrWithPeerId {
|
|
multiaddr: listen_addr,
|
|
peer_id: node1.local_peer_id(),
|
|
}],
|
|
..Default::default()
|
|
})
|
|
.build()
|
|
.start_network();
|
|
|
|
let receiver = tokio::spawn(async move {
|
|
// Wait for the `NotificationStreamOpened`.
|
|
loop {
|
|
match events_stream2.next().await.unwrap() {
|
|
Event::NotificationStreamOpened { protocol, negotiated_fallback, .. } => {
|
|
assert_eq!(protocol, PROTOCOL_NAME.into());
|
|
assert_eq!(negotiated_fallback, None);
|
|
break
|
|
},
|
|
_ => {},
|
|
};
|
|
}
|
|
});
|
|
|
|
// Wait for the `NotificationStreamOpened`.
|
|
loop {
|
|
match events_stream1.next().await.unwrap() {
|
|
Event::NotificationStreamOpened { protocol, negotiated_fallback, .. }
|
|
if protocol == NEW_PROTOCOL_NAME.into() =>
|
|
{
|
|
assert_eq!(negotiated_fallback, Some(PROTOCOL_NAME.into()));
|
|
break
|
|
},
|
|
_ => {},
|
|
};
|
|
}
|
|
|
|
receiver.await.unwrap();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_listen_addresses_consistent_with_transport_memory() {
|
|
let listen_addr = config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)];
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
transport: TransportConfig::MemoryOnly,
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_listen_addresses_consistent_with_transport_not_memory() {
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_boot_node_addresses_consistent_with_transport_memory() {
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
let boot_node = MultiaddrWithPeerId {
|
|
multiaddr: config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)],
|
|
peer_id: PeerId::random(),
|
|
};
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
transport: TransportConfig::MemoryOnly,
|
|
boot_nodes: vec![boot_node],
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_boot_node_addresses_consistent_with_transport_not_memory() {
|
|
let listen_addr = config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)];
|
|
let boot_node = MultiaddrWithPeerId {
|
|
multiaddr: config::build_multiaddr![Memory(rand::random::<u64>())],
|
|
peer_id: PeerId::random(),
|
|
};
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
boot_nodes: vec![boot_node],
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_reserved_node_addresses_consistent_with_transport_memory() {
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
let reserved_node = MultiaddrWithPeerId {
|
|
multiaddr: config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)],
|
|
peer_id: PeerId::random(),
|
|
};
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
transport: TransportConfig::MemoryOnly,
|
|
default_peers_set: config::SetConfig {
|
|
reserved_nodes: vec![reserved_node],
|
|
..Default::default()
|
|
},
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_reserved_node_addresses_consistent_with_transport_not_memory() {
|
|
let listen_addr = config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)];
|
|
let reserved_node = MultiaddrWithPeerId {
|
|
multiaddr: config::build_multiaddr![Memory(rand::random::<u64>())],
|
|
peer_id: PeerId::random(),
|
|
};
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
default_peers_set: config::SetConfig {
|
|
reserved_nodes: vec![reserved_node],
|
|
..Default::default()
|
|
},
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_public_addresses_consistent_with_transport_memory() {
|
|
let listen_addr = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
let public_address = config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)];
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
transport: TransportConfig::MemoryOnly,
|
|
public_addresses: vec![public_address],
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[should_panic(expected = "don't match the transport")]
|
|
async fn ensure_public_addresses_consistent_with_transport_not_memory() {
|
|
let listen_addr = config::build_multiaddr![Ip4([127, 0, 0, 1]), Tcp(0_u16)];
|
|
let public_address = config::build_multiaddr![Memory(rand::random::<u64>())];
|
|
|
|
let _ = TestNetworkBuilder::new()
|
|
.with_config(config::NetworkConfiguration {
|
|
listen_addresses: vec![listen_addr.clone()],
|
|
public_addresses: vec![public_address],
|
|
..config::NetworkConfiguration::new(
|
|
"test-node",
|
|
"test-client",
|
|
Default::default(),
|
|
None,
|
|
)
|
|
})
|
|
.build()
|
|
.start_network();
|
|
}
|