mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-29 20:47:56 +00:00
e71c484d5b
This commit introduces a new concept called `NotificationService` which allows Polkadot protocols to communicate with the underlying notification protocol implementation directly, without routing events through `NetworkWorker`. This implies that each protocol has its own service which it uses to communicate with remote peers and that each `NotificationService` is unique with respect to the underlying notification protocol, meaning `NotificationService` for the transaction protocol can only be used to send and receive transaction-related notifications. The `NotificationService` concept introduces two additional benefits: * allow protocols to start using custom handshakes * allow protocols to accept/reject inbound peers Previously the validation of inbound connections was solely the responsibility of `ProtocolController`. This caused issues with light peers and `SyncingEngine` as `ProtocolController` would accept more peers than `SyncingEngine` could accept which caused peers to have differing views of their own states. `SyncingEngine` would reject excess peers but these rejections were not properly communicated to those peers causing them to assume that they were accepted. With `NotificationService`, the local handshake is not sent to remote peer if peer is rejected which allows it to detect that it was rejected. This commit also deprecates the use of `NetworkEventStream` for all notification-related events and going forward only DHT events are provided through `NetworkEventStream`. If protocols wish to follow each other's events, they must introduce additional abtractions, as is done for GRANDPA and transactions protocols by following the syncing protocol through `SyncEventStream`. Fixes https://github.com/paritytech/polkadot-sdk/issues/512 Fixes https://github.com/paritytech/polkadot-sdk/issues/514 Fixes https://github.com/paritytech/polkadot-sdk/issues/515 Fixes https://github.com/paritytech/polkadot-sdk/issues/554 Fixes https://github.com/paritytech/polkadot-sdk/issues/556 --- These changes are transferred from https://github.com/paritytech/substrate/pull/14197 but there are no functional changes compared to that PR --------- Co-authored-by: Dmitry Markin <dmitry@markin.tech> Co-authored-by: Alexandru Vasile <60601340+lexnv@users.noreply.github.com>
474 lines
13 KiB
Rust
474 lines
13 KiB
Rust
// Copyright (C) Parity Technologies (UK) Ltd.
|
|
// This file is part of Polkadot.
|
|
|
|
// Polkadot is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// Polkadot is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with Polkadot. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
use super::*;
|
|
use futures::executor;
|
|
use polkadot_node_subsystem_util::TimeoutExt;
|
|
|
|
use async_trait::async_trait;
|
|
use parking_lot::Mutex;
|
|
use std::collections::HashSet;
|
|
|
|
use sc_network::{
|
|
IfDisconnected, ObservedRole as SubstrateObservedRole, ProtocolName, ReputationChange, Roles,
|
|
};
|
|
|
|
use parity_scale_codec::DecodeAll;
|
|
use polkadot_node_network_protocol::{
|
|
peer_set::{PeerSetProtocolNames, ValidationVersion},
|
|
request_response::{outgoing::Requests, ReqProtocolNames},
|
|
v1 as protocol_v1, v2 as protocol_v2, ObservedRole, Versioned,
|
|
};
|
|
use polkadot_node_subsystem::{FromOrchestra, OverseerSignal};
|
|
use polkadot_node_subsystem_test_helpers::TestSubsystemContextHandle;
|
|
use polkadot_node_subsystem_util::metered;
|
|
use polkadot_primitives::{AuthorityDiscoveryId, Hash};
|
|
use polkadot_primitives_test_helpers::dummy_collator_signature;
|
|
use sc_network::Multiaddr;
|
|
use sp_keyring::Sr25519Keyring;
|
|
|
|
const TIMEOUT: std::time::Duration = polkadot_node_subsystem_test_helpers::TestSubsystemContextHandle::<NetworkBridgeTxMessage>::TIMEOUT;
|
|
|
|
use crate::{network::Network, validator_discovery::AuthorityDiscovery};
|
|
|
|
#[derive(Debug, PartialEq)]
|
|
pub enum NetworkAction {
|
|
/// Note a change in reputation for a peer.
|
|
ReputationChange(PeerId, ReputationChange),
|
|
/// Disconnect a peer from the given peer-set.
|
|
DisconnectPeer(PeerId, PeerSet),
|
|
/// Write a notification to a given peer on the given peer-set.
|
|
WriteNotification(PeerId, PeerSet, Vec<u8>),
|
|
}
|
|
|
|
// The subsystem's view of the network.
|
|
#[derive(Clone)]
|
|
struct TestNetwork {
|
|
action_tx: Arc<Mutex<metered::UnboundedMeteredSender<NetworkAction>>>,
|
|
peerset_protocol_names: Arc<PeerSetProtocolNames>,
|
|
}
|
|
|
|
#[derive(Clone, Debug)]
|
|
struct TestAuthorityDiscovery;
|
|
|
|
// The test's view of the network. This receives updates from the subsystem in the form
|
|
// of `NetworkAction`s.
|
|
struct TestNetworkHandle {
|
|
action_rx: metered::UnboundedMeteredReceiver<NetworkAction>,
|
|
_peerset_protocol_names: PeerSetProtocolNames,
|
|
notification_sinks: Arc<Mutex<HashMap<(PeerSet, PeerId), Box<dyn MessageSink>>>>,
|
|
action_tx: Arc<Mutex<metered::UnboundedMeteredSender<NetworkAction>>>,
|
|
}
|
|
|
|
struct TestMessageSink {
|
|
peer: PeerId,
|
|
peer_set: PeerSet,
|
|
action_tx: Arc<Mutex<metered::UnboundedMeteredSender<NetworkAction>>>,
|
|
}
|
|
|
|
impl TestMessageSink {
|
|
fn new(
|
|
peer: PeerId,
|
|
peer_set: PeerSet,
|
|
action_tx: Arc<Mutex<metered::UnboundedMeteredSender<NetworkAction>>>,
|
|
) -> TestMessageSink {
|
|
Self { peer, peer_set, action_tx }
|
|
}
|
|
}
|
|
|
|
#[async_trait::async_trait]
|
|
impl MessageSink for TestMessageSink {
|
|
fn send_sync_notification(&self, notification: Vec<u8>) {
|
|
self.action_tx
|
|
.lock()
|
|
.unbounded_send(NetworkAction::WriteNotification(
|
|
self.peer,
|
|
self.peer_set,
|
|
notification,
|
|
))
|
|
.unwrap();
|
|
}
|
|
|
|
async fn send_async_notification(
|
|
&self,
|
|
_notification: Vec<u8>,
|
|
) -> Result<(), sc_network::error::Error> {
|
|
unimplemented!();
|
|
}
|
|
}
|
|
|
|
fn new_test_network(
|
|
peerset_protocol_names: PeerSetProtocolNames,
|
|
) -> (
|
|
TestNetwork,
|
|
TestNetworkHandle,
|
|
TestAuthorityDiscovery,
|
|
Arc<Mutex<HashMap<(PeerSet, PeerId), Box<dyn MessageSink>>>>,
|
|
) {
|
|
let (action_tx, action_rx) = metered::unbounded();
|
|
let notification_sinks = Arc::new(Mutex::new(HashMap::new()));
|
|
let action_tx = Arc::new(Mutex::new(action_tx));
|
|
|
|
(
|
|
TestNetwork {
|
|
action_tx: action_tx.clone(),
|
|
peerset_protocol_names: Arc::new(peerset_protocol_names.clone()),
|
|
},
|
|
TestNetworkHandle {
|
|
action_rx,
|
|
_peerset_protocol_names: peerset_protocol_names,
|
|
action_tx,
|
|
notification_sinks: notification_sinks.clone(),
|
|
},
|
|
TestAuthorityDiscovery,
|
|
notification_sinks,
|
|
)
|
|
}
|
|
|
|
#[async_trait]
|
|
impl Network for TestNetwork {
|
|
async fn set_reserved_peers(
|
|
&mut self,
|
|
_protocol: ProtocolName,
|
|
_: HashSet<Multiaddr>,
|
|
) -> Result<(), String> {
|
|
Ok(())
|
|
}
|
|
|
|
async fn remove_from_peers_set(
|
|
&mut self,
|
|
_protocol: ProtocolName,
|
|
_: Vec<PeerId>,
|
|
) -> Result<(), String> {
|
|
Ok(())
|
|
}
|
|
|
|
async fn start_request<AD: AuthorityDiscovery>(
|
|
&self,
|
|
_: &mut AD,
|
|
_: Requests,
|
|
_: &ReqProtocolNames,
|
|
_: IfDisconnected,
|
|
) {
|
|
}
|
|
|
|
fn report_peer(&self, who: PeerId, rep: ReputationChange) {
|
|
self.action_tx
|
|
.lock()
|
|
.unbounded_send(NetworkAction::ReputationChange(who, rep))
|
|
.unwrap();
|
|
}
|
|
|
|
fn disconnect_peer(&self, who: PeerId, protocol: ProtocolName) {
|
|
let (peer_set, version) = self.peerset_protocol_names.try_get_protocol(&protocol).unwrap();
|
|
assert_eq!(version, peer_set.get_main_version());
|
|
|
|
self.action_tx
|
|
.lock()
|
|
.unbounded_send(NetworkAction::DisconnectPeer(who, peer_set))
|
|
.unwrap();
|
|
}
|
|
|
|
fn peer_role(&self, _peer_id: PeerId, handshake: Vec<u8>) -> Option<SubstrateObservedRole> {
|
|
Roles::decode_all(&mut &handshake[..])
|
|
.ok()
|
|
.and_then(|role| Some(SubstrateObservedRole::from(role)))
|
|
}
|
|
}
|
|
|
|
#[async_trait]
|
|
impl validator_discovery::AuthorityDiscovery for TestAuthorityDiscovery {
|
|
async fn get_addresses_by_authority_id(
|
|
&mut self,
|
|
_authority: AuthorityDiscoveryId,
|
|
) -> Option<HashSet<Multiaddr>> {
|
|
None
|
|
}
|
|
|
|
async fn get_authority_ids_by_peer_id(
|
|
&mut self,
|
|
_peer_id: PeerId,
|
|
) -> Option<HashSet<AuthorityDiscoveryId>> {
|
|
None
|
|
}
|
|
}
|
|
|
|
impl TestNetworkHandle {
|
|
// Get the next network action.
|
|
async fn next_network_action(&mut self) -> NetworkAction {
|
|
self.action_rx.next().await.expect("subsystem concluded early")
|
|
}
|
|
|
|
async fn connect_peer(
|
|
&mut self,
|
|
peer: PeerId,
|
|
_protocol_version: ValidationVersion,
|
|
peer_set: PeerSet,
|
|
_role: ObservedRole,
|
|
) {
|
|
self.notification_sinks.lock().insert(
|
|
(peer_set, peer),
|
|
Box::new(TestMessageSink::new(peer, peer_set, self.action_tx.clone())),
|
|
);
|
|
}
|
|
}
|
|
|
|
type VirtualOverseer = TestSubsystemContextHandle<NetworkBridgeTxMessage>;
|
|
|
|
struct TestHarness {
|
|
network_handle: TestNetworkHandle,
|
|
virtual_overseer: VirtualOverseer,
|
|
}
|
|
|
|
fn test_harness<T: Future<Output = VirtualOverseer>>(test: impl FnOnce(TestHarness) -> T) {
|
|
let genesis_hash = Hash::repeat_byte(0xff);
|
|
let fork_id = None;
|
|
let req_protocol_names = ReqProtocolNames::new(genesis_hash, fork_id);
|
|
let peerset_protocol_names = PeerSetProtocolNames::new(genesis_hash, fork_id);
|
|
|
|
let pool = sp_core::testing::TaskExecutor::new();
|
|
let (network, network_handle, discovery, network_notification_sinks) =
|
|
new_test_network(peerset_protocol_names.clone());
|
|
|
|
let (context, virtual_overseer) =
|
|
polkadot_node_subsystem_test_helpers::make_subsystem_context(pool);
|
|
|
|
let bridge_out = NetworkBridgeTx::new(
|
|
network,
|
|
discovery,
|
|
Metrics(None),
|
|
req_protocol_names,
|
|
peerset_protocol_names,
|
|
network_notification_sinks,
|
|
);
|
|
|
|
let network_bridge_out_fut = run_network_out(bridge_out, context)
|
|
.map_err(|e| panic!("bridge-out subsystem execution failed {:?}", e))
|
|
.map(|_| ());
|
|
|
|
let test_fut = test(TestHarness { network_handle, virtual_overseer });
|
|
|
|
futures::pin_mut!(test_fut);
|
|
futures::pin_mut!(network_bridge_out_fut);
|
|
|
|
let _ = executor::block_on(future::join(
|
|
async move {
|
|
let mut virtual_overseer = test_fut.await;
|
|
virtual_overseer.send(FromOrchestra::Signal(OverseerSignal::Conclude)).await;
|
|
},
|
|
network_bridge_out_fut,
|
|
));
|
|
}
|
|
|
|
#[test]
|
|
fn send_messages_to_peers() {
|
|
test_harness(|test_harness| async move {
|
|
let TestHarness { mut network_handle, mut virtual_overseer } = test_harness;
|
|
|
|
let peer = PeerId::random();
|
|
|
|
network_handle
|
|
.connect_peer(peer, ValidationVersion::V1, PeerSet::Validation, ObservedRole::Full)
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur");
|
|
|
|
// the outgoing side does not consume network messages
|
|
// so the single item sink has to be free explicitly
|
|
|
|
network_handle
|
|
.connect_peer(peer, ValidationVersion::V1, PeerSet::Collation, ObservedRole::Full)
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur");
|
|
|
|
// send a validation protocol message.
|
|
|
|
{
|
|
let approval_distribution_message =
|
|
protocol_v1::ApprovalDistributionMessage::Approvals(Vec::new());
|
|
|
|
let message_v1 = protocol_v1::ValidationProtocol::ApprovalDistribution(
|
|
approval_distribution_message.clone(),
|
|
);
|
|
|
|
virtual_overseer
|
|
.send(FromOrchestra::Communication {
|
|
msg: NetworkBridgeTxMessage::SendValidationMessage(
|
|
vec![peer],
|
|
Versioned::V1(message_v1.clone()),
|
|
),
|
|
})
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur");
|
|
|
|
assert_eq!(
|
|
network_handle
|
|
.next_network_action()
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur"),
|
|
NetworkAction::WriteNotification(
|
|
peer,
|
|
PeerSet::Validation,
|
|
WireMessage::ProtocolMessage(message_v1).encode(),
|
|
)
|
|
);
|
|
}
|
|
|
|
// send a collation protocol message.
|
|
|
|
{
|
|
let collator_protocol_message = protocol_v1::CollatorProtocolMessage::Declare(
|
|
Sr25519Keyring::Alice.public().into(),
|
|
0_u32.into(),
|
|
dummy_collator_signature(),
|
|
);
|
|
|
|
let message_v1 =
|
|
protocol_v1::CollationProtocol::CollatorProtocol(collator_protocol_message.clone());
|
|
|
|
virtual_overseer
|
|
.send(FromOrchestra::Communication {
|
|
msg: NetworkBridgeTxMessage::SendCollationMessage(
|
|
vec![peer],
|
|
Versioned::V1(message_v1.clone()),
|
|
),
|
|
})
|
|
.await;
|
|
|
|
assert_eq!(
|
|
network_handle
|
|
.next_network_action()
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur"),
|
|
NetworkAction::WriteNotification(
|
|
peer,
|
|
PeerSet::Collation,
|
|
WireMessage::ProtocolMessage(message_v1).encode(),
|
|
)
|
|
);
|
|
}
|
|
virtual_overseer
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn network_protocol_versioning_send() {
|
|
test_harness(|test_harness| async move {
|
|
let TestHarness { mut network_handle, mut virtual_overseer } = test_harness;
|
|
|
|
let peer_ids: Vec<_> = (0..4).map(|_| PeerId::random()).collect();
|
|
let peers = [
|
|
(peer_ids[0], PeerSet::Validation, ValidationVersion::V2),
|
|
(peer_ids[1], PeerSet::Collation, ValidationVersion::V1),
|
|
(peer_ids[2], PeerSet::Validation, ValidationVersion::V1),
|
|
(peer_ids[3], PeerSet::Collation, ValidationVersion::V2),
|
|
];
|
|
|
|
for &(peer_id, peer_set, version) in &peers {
|
|
network_handle
|
|
.connect_peer(peer_id, version, peer_set, ObservedRole::Full)
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur");
|
|
}
|
|
|
|
// send a validation protocol message.
|
|
{
|
|
let approval_distribution_message =
|
|
protocol_v2::ApprovalDistributionMessage::Approvals(Vec::new());
|
|
|
|
let msg = protocol_v2::ValidationProtocol::ApprovalDistribution(
|
|
approval_distribution_message.clone(),
|
|
);
|
|
|
|
// only `peer_ids[0]` opened the validation protocol v2
|
|
// so only they will be sent a notification
|
|
let receivers = vec![peer_ids[0]];
|
|
virtual_overseer
|
|
.send(FromOrchestra::Communication {
|
|
msg: NetworkBridgeTxMessage::SendValidationMessage(
|
|
receivers.clone(),
|
|
Versioned::V2(msg.clone()),
|
|
),
|
|
})
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur");
|
|
|
|
for peer in &receivers {
|
|
assert_eq!(
|
|
network_handle
|
|
.next_network_action()
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur"),
|
|
NetworkAction::WriteNotification(
|
|
*peer,
|
|
PeerSet::Validation,
|
|
WireMessage::ProtocolMessage(msg.clone()).encode(),
|
|
)
|
|
);
|
|
}
|
|
}
|
|
|
|
// send a collation protocol message.
|
|
|
|
{
|
|
let collator_protocol_message = protocol_v2::CollatorProtocolMessage::Declare(
|
|
Sr25519Keyring::Alice.public().into(),
|
|
0_u32.into(),
|
|
dummy_collator_signature(),
|
|
);
|
|
|
|
let msg =
|
|
protocol_v2::CollationProtocol::CollatorProtocol(collator_protocol_message.clone());
|
|
|
|
// only `peer_ids[0]` opened the collation protocol v2
|
|
// so only they will be sent a notification
|
|
let receivers = vec![peer_ids[1]];
|
|
|
|
virtual_overseer
|
|
.send(FromOrchestra::Communication {
|
|
msg: NetworkBridgeTxMessage::SendCollationMessages(vec![(
|
|
receivers.clone(),
|
|
Versioned::V2(msg.clone()),
|
|
)]),
|
|
})
|
|
.await;
|
|
|
|
for peer in &receivers {
|
|
assert_eq!(
|
|
network_handle
|
|
.next_network_action()
|
|
.timeout(TIMEOUT)
|
|
.await
|
|
.expect("Timeout does not occur"),
|
|
NetworkAction::WriteNotification(
|
|
*peer,
|
|
PeerSet::Collation,
|
|
WireMessage::ProtocolMessage(msg.clone()).encode(),
|
|
)
|
|
);
|
|
}
|
|
}
|
|
virtual_overseer
|
|
});
|
|
}
|