// This file is part of Substrate. // Copyright (C) 2017-2022 Parity Technologies (UK) Ltd. // SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0 // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see . use crate::config; use bytes::Bytes; use codec::{Decode, DecodeAll, Encode}; use futures::{channel::oneshot, prelude::*}; use libp2p::{ core::{connection::ConnectionId, transport::ListenerId, ConnectedPoint}, request_response::OutboundFailure, swarm::{ ConnectionHandler, IntoConnectionHandler, NetworkBehaviour, NetworkBehaviourAction, PollParameters, }, Multiaddr, PeerId, }; use log::{debug, error, info, log, trace, warn, Level}; use message::{generic::Message as GenericMessage, Message}; use notifications::{Notifications, NotificationsOut}; use prometheus_endpoint::{register, Gauge, GaugeVec, Opts, PrometheusError, Registry, U64}; use sc_client_api::HeaderBackend; use sc_consensus::import_queue::{ BlockImportError, BlockImportStatus, IncomingBlock, RuntimeOrigin, }; use sc_network_common::{ config::NonReservedPeerMode, error, protocol::{role::Roles, ProtocolName}, request_responses::RequestFailure, sync::{ message::{ BlockAnnounce, BlockAnnouncesHandshake, BlockAttributes, BlockData, BlockRequest, BlockResponse, BlockState, }, warp::{EncodedProof, WarpProofRequest}, BadPeer, ChainSync, OnBlockData, OnBlockJustification, OnStateData, OpaqueBlockRequest, OpaqueBlockResponse, OpaqueStateRequest, OpaqueStateResponse, PollBlockAnnounceValidation, SyncStatus, }, utils::{interval, LruHashSet}, }; use sp_arithmetic::traits::SaturatedConversion; use sp_consensus::BlockOrigin; use sp_runtime::{ generic::BlockId, traits::{Block as BlockT, CheckedSub, Header as HeaderT, NumberFor, Zero}, Justifications, }; use std::{ collections::{HashMap, HashSet, VecDeque}, io, iter, num::NonZeroUsize, pin::Pin, sync::Arc, task::Poll, time, }; mod notifications; pub mod message; pub use notifications::{NotificationsSink, NotifsHandlerError, Ready}; /// Interval at which we perform time based maintenance const TICK_TIMEOUT: time::Duration = time::Duration::from_millis(1100); /// Maximum number of known block hashes to keep for a peer. const MAX_KNOWN_BLOCKS: usize = 1024; // ~32kb per peer + LruHashSet overhead /// Maximum size used for notifications in the block announce and transaction protocols. // Must be equal to `max(MAX_BLOCK_ANNOUNCE_SIZE, MAX_TRANSACTIONS_SIZE)`. pub(crate) const BLOCK_ANNOUNCES_TRANSACTIONS_SUBSTREAM_SIZE: u64 = 16 * 1024 * 1024; /// Identifier of the peerset for the block announces protocol. const HARDCODED_PEERSETS_SYNC: sc_peerset::SetId = sc_peerset::SetId::from(0); /// Number of hardcoded peersets (the constants right above). Any set whose identifier is equal or /// superior to this value corresponds to a user-defined protocol. const NUM_HARDCODED_PEERSETS: usize = 1; /// When light node connects to the full node and the full node is behind light node /// for at least `LIGHT_MAXIMAL_BLOCKS_DIFFERENCE` blocks, we consider it not useful /// and disconnect to free connection slot. const LIGHT_MAXIMAL_BLOCKS_DIFFERENCE: u64 = 8192; mod rep { use sc_peerset::ReputationChange as Rep; /// Reputation change when a peer doesn't respond in time to our messages. pub const TIMEOUT: Rep = Rep::new(-(1 << 10), "Request timeout"); /// Reputation change when a peer refuses a request. pub const REFUSED: Rep = Rep::new(-(1 << 10), "Request refused"); /// Reputation change when we are a light client and a peer is behind us. pub const PEER_BEHIND_US_LIGHT: Rep = Rep::new(-(1 << 8), "Useless for a light peer"); /// We received a message that failed to decode. pub const BAD_MESSAGE: Rep = Rep::new(-(1 << 12), "Bad message"); /// Peer has different genesis. pub const GENESIS_MISMATCH: Rep = Rep::new_fatal("Genesis mismatch"); /// Peer is on unsupported protocol version. pub const BAD_PROTOCOL: Rep = Rep::new_fatal("Unsupported protocol"); /// Peer role does not match (e.g. light peer connecting to another light peer). pub const BAD_ROLE: Rep = Rep::new_fatal("Unsupported role"); /// Peer send us a block announcement that failed at validation. pub const BAD_BLOCK_ANNOUNCEMENT: Rep = Rep::new(-(1 << 12), "Bad block announcement"); } struct Metrics { peers: Gauge, queued_blocks: Gauge, fork_targets: Gauge, justifications: GaugeVec, } impl Metrics { fn register(r: &Registry) -> Result { Ok(Self { peers: { let g = Gauge::new("substrate_sync_peers", "Number of peers we sync with")?; register(g, r)? }, queued_blocks: { let g = Gauge::new("substrate_sync_queued_blocks", "Number of blocks in import queue")?; register(g, r)? }, fork_targets: { let g = Gauge::new("substrate_sync_fork_targets", "Number of fork sync targets")?; register(g, r)? }, justifications: { let g = GaugeVec::new( Opts::new( "substrate_sync_extra_justifications", "Number of extra justifications requests", ), &["status"], )?; register(g, r)? }, }) } } // Lock must always be taken in order declared here. pub struct Protocol { /// Interval at which we call `tick`. tick_timeout: Pin + Send>>, /// Pending list of messages to return from `poll` as a priority. pending_messages: VecDeque>, /// Assigned roles. roles: Roles, genesis_hash: B::Hash, /// State machine that handles the list of in-progress requests. Only full node peers are /// registered. chain_sync: Box>, // All connected peers. Contains both full and light node peers. peers: HashMap>, chain: Arc, /// List of nodes for which we perform additional logging because they are important for the /// user. important_peers: HashSet, /// List of nodes that should never occupy peer slots. default_peers_set_no_slot_peers: HashSet, /// Actual list of connected no-slot nodes. default_peers_set_no_slot_connected_peers: HashSet, /// Value that was passed as part of the configuration. Used to cap the number of full nodes. default_peers_set_num_full: usize, /// Number of slots to allocate to light nodes. default_peers_set_num_light: usize, /// Used to report reputation changes. peerset_handle: sc_peerset::PeersetHandle, /// Handles opening the unique substream and sending and receiving raw messages. behaviour: Notifications, /// List of notifications protocols that have been registered. notification_protocols: Vec, /// If we receive a new "substream open" event that contains an invalid handshake, we ask the /// inner layer to force-close the substream. Force-closing the substream will generate a /// "substream closed" event. This is a problem: since we can't propagate the "substream open" /// event to the outer layers, we also shouldn't propagate this "substream closed" event. To /// solve this, an entry is added to this map whenever an invalid handshake is received. /// Entries are removed when the corresponding "substream closed" is later received. bad_handshake_substreams: HashSet<(PeerId, sc_peerset::SetId)>, /// Prometheus metrics. metrics: Option, /// The `PeerId`'s of all boot nodes. boot_node_ids: HashSet, /// A cache for the data that was associated to a block announcement. block_announce_data_cache: lru::LruCache>, } #[derive(Debug)] enum PeerRequest { Block(BlockRequest), State, WarpProof, } /// Peer information #[derive(Debug)] struct Peer { info: PeerInfo, /// Current request, if any. Started by emitting [`CustomMessageOutcome::BlockRequest`]. request: Option<(PeerRequest, oneshot::Receiver, RequestFailure>>)>, /// Holds a set of blocks known to this peer. known_blocks: LruHashSet, } /// Info about a peer's known state. #[derive(Clone, Debug)] pub struct PeerInfo { /// Roles pub roles: Roles, /// Peer best block hash pub best_hash: B::Hash, /// Peer best block number pub best_number: ::Number, } impl Protocol where B: BlockT, Client: HeaderBackend + 'static, { /// Create a new instance. pub fn new( roles: Roles, chain: Arc, network_config: &config::NetworkConfiguration, metrics_registry: Option<&Registry>, chain_sync: Box>, block_announces_protocol: sc_network_common::config::NonDefaultSetConfig, ) -> error::Result<(Self, sc_peerset::PeersetHandle, Vec<(PeerId, Multiaddr)>)> { let info = chain.info(); let boot_node_ids = { let mut list = HashSet::new(); for node in &network_config.boot_nodes { list.insert(node.peer_id); } list.shrink_to_fit(); list }; let important_peers = { let mut imp_p = HashSet::new(); for reserved in &network_config.default_peers_set.reserved_nodes { imp_p.insert(reserved.peer_id); } for reserved in network_config .extra_sets .iter() .flat_map(|s| s.set_config.reserved_nodes.iter()) { imp_p.insert(reserved.peer_id); } imp_p.shrink_to_fit(); imp_p }; let default_peers_set_no_slot_peers = { let mut no_slot_p: HashSet = network_config .default_peers_set .reserved_nodes .iter() .map(|reserved| reserved.peer_id) .collect(); no_slot_p.shrink_to_fit(); no_slot_p }; let mut known_addresses = Vec::new(); let (peerset, peerset_handle) = { let mut sets = Vec::with_capacity(NUM_HARDCODED_PEERSETS + network_config.extra_sets.len()); let mut default_sets_reserved = HashSet::new(); for reserved in network_config.default_peers_set.reserved_nodes.iter() { default_sets_reserved.insert(reserved.peer_id); if !reserved.multiaddr.is_empty() { known_addresses.push((reserved.peer_id, reserved.multiaddr.clone())); } } let mut bootnodes = Vec::with_capacity(network_config.boot_nodes.len()); for bootnode in network_config.boot_nodes.iter() { bootnodes.push(bootnode.peer_id); } // Set number 0 is used for block announces. sets.push(sc_peerset::SetConfig { in_peers: network_config.default_peers_set.in_peers, out_peers: network_config.default_peers_set.out_peers, bootnodes, reserved_nodes: default_sets_reserved.clone(), reserved_only: network_config.default_peers_set.non_reserved_mode == NonReservedPeerMode::Deny, }); for set_cfg in &network_config.extra_sets { let mut reserved_nodes = HashSet::new(); for reserved in set_cfg.set_config.reserved_nodes.iter() { reserved_nodes.insert(reserved.peer_id); known_addresses.push((reserved.peer_id, reserved.multiaddr.clone())); } let reserved_only = set_cfg.set_config.non_reserved_mode == NonReservedPeerMode::Deny; sets.push(sc_peerset::SetConfig { in_peers: set_cfg.set_config.in_peers, out_peers: set_cfg.set_config.out_peers, bootnodes: Vec::new(), reserved_nodes, reserved_only, }); } sc_peerset::Peerset::from_config(sc_peerset::PeersetConfig { sets }) }; let behaviour = { Notifications::new( peerset, // NOTE: Block announcement protocol is still very much hardcoded into `Protocol`. // This protocol must be the first notification protocol given to // `Notifications` iter::once(notifications::ProtocolConfig { name: block_announces_protocol.notifications_protocol.clone(), fallback_names: block_announces_protocol.fallback_names.clone(), handshake: block_announces_protocol.handshake.as_ref().unwrap().to_vec(), max_notification_size: block_announces_protocol.max_notification_size, }) .chain(network_config.extra_sets.iter().map(|s| notifications::ProtocolConfig { name: s.notifications_protocol.clone(), fallback_names: s.fallback_names.clone(), handshake: s.handshake.as_ref().map_or(roles.encode(), |h| (*h).to_vec()), max_notification_size: s.max_notification_size, })), ) }; let block_announce_data_cache = lru::LruCache::new( network_config.default_peers_set.in_peers as usize + network_config.default_peers_set.out_peers as usize, ); let protocol = Self { tick_timeout: Box::pin(interval(TICK_TIMEOUT)), pending_messages: VecDeque::new(), roles, peers: HashMap::new(), chain, genesis_hash: info.genesis_hash, chain_sync, important_peers, default_peers_set_no_slot_peers, default_peers_set_no_slot_connected_peers: HashSet::new(), default_peers_set_num_full: network_config.default_peers_set_num_full as usize, default_peers_set_num_light: { let total = network_config.default_peers_set.out_peers + network_config.default_peers_set.in_peers; total.saturating_sub(network_config.default_peers_set_num_full) as usize }, peerset_handle: peerset_handle.clone(), behaviour, notification_protocols: iter::once(block_announces_protocol.notifications_protocol) .chain(network_config.extra_sets.iter().map(|s| s.notifications_protocol.clone())) .collect(), bad_handshake_substreams: Default::default(), metrics: if let Some(r) = metrics_registry { Some(Metrics::register(r)?) } else { None }, boot_node_ids, block_announce_data_cache, }; Ok((protocol, peerset_handle, known_addresses)) } /// Returns the list of all the peers we have an open channel to. pub fn open_peers(&self) -> impl Iterator { self.behaviour.open_peers() } /// Returns the number of discovered nodes that we keep in memory. pub fn num_discovered_peers(&self) -> usize { self.behaviour.num_discovered_peers() } /// Disconnects the given peer if we are connected to it. pub fn disconnect_peer(&mut self, peer_id: &PeerId, protocol_name: ProtocolName) { if let Some(position) = self.notification_protocols.iter().position(|p| *p == protocol_name) { self.behaviour.disconnect_peer(peer_id, sc_peerset::SetId::from(position)); } else { warn!(target: "sub-libp2p", "disconnect_peer() with invalid protocol name") } } /// Returns the state of the peerset manager, for debugging purposes. pub fn peerset_debug_info(&mut self) -> serde_json::Value { self.behaviour.peerset_debug_info() } /// Returns the number of peers we're connected to. pub fn num_connected_peers(&self) -> usize { self.peers.len() } /// Returns the number of peers we're connected to and that are being queried. pub fn num_active_peers(&self) -> usize { self.peers.values().filter(|p| p.request.is_some()).count() } /// Current global sync state. pub fn sync_state(&self) -> SyncStatus { self.chain_sync.status() } /// Target sync block number. pub fn best_seen_block(&self) -> Option> { self.chain_sync.status().best_seen_block } /// Number of peers participating in syncing. pub fn num_sync_peers(&self) -> u32 { self.chain_sync.status().num_peers } /// Number of blocks in the import queue. pub fn num_queued_blocks(&self) -> u32 { self.chain_sync.status().queued_blocks } /// Number of downloaded blocks. pub fn num_downloaded_blocks(&self) -> usize { self.chain_sync.num_downloaded_blocks() } /// Number of active sync requests. pub fn num_sync_requests(&self) -> usize { self.chain_sync.num_sync_requests() } /// Inform sync about new best imported block. pub fn new_best_block_imported(&mut self, hash: B::Hash, number: NumberFor) { debug!(target: "sync", "New best block imported {:?}/#{}", hash, number); self.chain_sync.update_chain_info(&hash, number); self.behaviour.set_notif_protocol_handshake( HARDCODED_PEERSETS_SYNC, BlockAnnouncesHandshake::::build(self.roles, number, hash, self.genesis_hash) .encode(), ); } fn update_peer_info(&mut self, who: &PeerId) { if let Some(info) = self.chain_sync.peer_info(who) { if let Some(ref mut peer) = self.peers.get_mut(who) { peer.info.best_hash = info.best_hash; peer.info.best_number = info.best_number; } } } /// Returns information about all the peers we are connected to after the handshake message. pub fn peers_info(&self) -> impl Iterator)> { self.peers.iter().map(|(id, peer)| (id, &peer.info)) } /// Called by peer when it is disconnecting. /// /// Returns a result if the handshake of this peer was indeed accepted. pub fn on_sync_peer_disconnected(&mut self, peer: PeerId) -> Result<(), ()> { if self.important_peers.contains(&peer) { warn!(target: "sync", "Reserved peer {} disconnected", peer); } else { debug!(target: "sync", "{} disconnected", peer); } if let Some(_peer_data) = self.peers.remove(&peer) { if let Some(OnBlockData::Import(origin, blocks)) = self.chain_sync.peer_disconnected(&peer) { self.pending_messages .push_back(CustomMessageOutcome::BlockImport(origin, blocks)); } self.default_peers_set_no_slot_connected_peers.remove(&peer); Ok(()) } else { Err(()) } } /// Adjusts the reputation of a node. pub fn report_peer(&self, who: PeerId, reputation: sc_peerset::ReputationChange) { self.peerset_handle.report_peer(who, reputation) } /// Must be called in response to a [`CustomMessageOutcome::BlockRequest`] being emitted. /// Must contain the same `PeerId` and request that have been emitted. pub fn on_block_response( &mut self, peer_id: PeerId, request: BlockRequest, response: OpaqueBlockResponse, ) -> CustomMessageOutcome { let blocks = match self.chain_sync.block_response_into_blocks(&request, response) { Ok(blocks) => blocks, Err(err) => { debug!(target: "sync", "Failed to decode block response from {}: {}", peer_id, err); self.peerset_handle.report_peer(peer_id, rep::BAD_MESSAGE); return CustomMessageOutcome::None }, }; let block_response = BlockResponse:: { id: request.id, blocks }; let blocks_range = || match ( block_response .blocks .first() .and_then(|b| b.header.as_ref().map(|h| h.number())), block_response.blocks.last().and_then(|b| b.header.as_ref().map(|h| h.number())), ) { (Some(first), Some(last)) if first != last => format!(" ({}..{})", first, last), (Some(first), Some(_)) => format!(" ({})", first), _ => Default::default(), }; trace!(target: "sync", "BlockResponse {} from {} with {} blocks {}", block_response.id, peer_id, block_response.blocks.len(), blocks_range(), ); if request.fields == BlockAttributes::JUSTIFICATION { match self.chain_sync.on_block_justification(peer_id, block_response) { Ok(OnBlockJustification::Nothing) => CustomMessageOutcome::None, Ok(OnBlockJustification::Import { peer, hash, number, justifications }) => CustomMessageOutcome::JustificationImport(peer, hash, number, justifications), Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu); CustomMessageOutcome::None }, } } else { match self.chain_sync.on_block_data(&peer_id, Some(request), block_response) { Ok(OnBlockData::Import(origin, blocks)) => CustomMessageOutcome::BlockImport(origin, blocks), Ok(OnBlockData::Request(peer, req)) => prepare_block_request(self.chain_sync.as_ref(), &mut self.peers, peer, req), Ok(OnBlockData::Continue) => CustomMessageOutcome::None, Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu); CustomMessageOutcome::None }, } } } /// Must be called in response to a [`CustomMessageOutcome::StateRequest`] being emitted. /// Must contain the same `PeerId` and request that have been emitted. pub fn on_state_response( &mut self, peer_id: PeerId, response: OpaqueStateResponse, ) -> CustomMessageOutcome { match self.chain_sync.on_state_data(&peer_id, response) { Ok(OnStateData::Import(origin, block)) => CustomMessageOutcome::BlockImport(origin, vec![block]), Ok(OnStateData::Continue) => CustomMessageOutcome::None, Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu); CustomMessageOutcome::None }, } } /// Must be called in response to a [`CustomMessageOutcome::WarpSyncRequest`] being emitted. /// Must contain the same `PeerId` and request that have been emitted. pub fn on_warp_sync_response( &mut self, peer_id: PeerId, response: EncodedProof, ) -> CustomMessageOutcome { match self.chain_sync.on_warp_sync_data(&peer_id, response) { Ok(()) => CustomMessageOutcome::None, Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu); CustomMessageOutcome::None }, } } /// Perform time based maintenance. /// /// > **Note**: This method normally doesn't have to be called except for testing purposes. pub fn tick(&mut self) { self.report_metrics() } /// Called on the first connection between two peers on the default set, after their exchange /// of handshake. /// /// Returns `Ok` if the handshake is accepted and the peer added to the list of peers we sync /// from. fn on_sync_peer_connected( &mut self, who: PeerId, status: BlockAnnouncesHandshake, ) -> Result<(), ()> { trace!(target: "sync", "New peer {} {:?}", who, status); if self.peers.contains_key(&who) { error!(target: "sync", "Called on_sync_peer_connected with already connected peer {}", who); debug_assert!(false); return Err(()) } if status.genesis_hash != self.genesis_hash { log!( target: "sync", if self.important_peers.contains(&who) { Level::Warn } else { Level::Debug }, "Peer is on different chain (our genesis: {} theirs: {})", self.genesis_hash, status.genesis_hash ); self.peerset_handle.report_peer(who, rep::GENESIS_MISMATCH); self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); if self.boot_node_ids.contains(&who) { error!( target: "sync", "Bootnode with peer id `{}` is on a different chain (our genesis: {} theirs: {})", who, self.genesis_hash, status.genesis_hash, ); } return Err(()) } if self.roles.is_light() { // we're not interested in light peers if status.roles.is_light() { debug!(target: "sync", "Peer {} is unable to serve light requests", who); self.peerset_handle.report_peer(who, rep::BAD_ROLE); self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); return Err(()) } // we don't interested in peers that are far behind us let self_best_block = self.chain.info().best_number; let blocks_difference = self_best_block .checked_sub(&status.best_number) .unwrap_or_else(Zero::zero) .saturated_into::(); if blocks_difference > LIGHT_MAXIMAL_BLOCKS_DIFFERENCE { debug!(target: "sync", "Peer {} is far behind us and will unable to serve light requests", who); self.peerset_handle.report_peer(who, rep::PEER_BEHIND_US_LIGHT); self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); return Err(()) } } let no_slot_peer = self.default_peers_set_no_slot_peers.contains(&who); let this_peer_reserved_slot: usize = if no_slot_peer { 1 } else { 0 }; if status.roles.is_full() && self.chain_sync.num_peers() >= self.default_peers_set_num_full + self.default_peers_set_no_slot_connected_peers.len() + this_peer_reserved_slot { debug!(target: "sync", "Too many full nodes, rejecting {}", who); self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); return Err(()) } if status.roles.is_light() && (self.peers.len() - self.chain_sync.num_peers()) >= self.default_peers_set_num_light { // Make sure that not all slots are occupied by light clients. debug!(target: "sync", "Too many light nodes, rejecting {}", who); self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); return Err(()) } let peer = Peer { info: PeerInfo { roles: status.roles, best_hash: status.best_hash, best_number: status.best_number, }, request: None, known_blocks: LruHashSet::new( NonZeroUsize::new(MAX_KNOWN_BLOCKS).expect("Constant is nonzero"), ), }; let req = if peer.info.roles.is_full() { match self.chain_sync.new_peer(who, peer.info.best_hash, peer.info.best_number) { Ok(req) => req, Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu); return Err(()) }, } } else { None }; debug!(target: "sync", "Connected {}", who); self.peers.insert(who, peer); if no_slot_peer { self.default_peers_set_no_slot_connected_peers.insert(who); } self.pending_messages .push_back(CustomMessageOutcome::PeerNewBest(who, status.best_number)); if let Some(req) = req { self.pending_messages.push_back(prepare_block_request( self.chain_sync.as_ref(), &mut self.peers, who, req, )); } Ok(()) } /// Make sure an important block is propagated to peers. /// /// In chain-based consensus, we often need to make sure non-best forks are /// at least temporarily synced. pub fn announce_block(&mut self, hash: B::Hash, data: Option>) { let header = match self.chain.header(BlockId::Hash(hash)) { Ok(Some(header)) => header, Ok(None) => { warn!("Trying to announce unknown block: {}", hash); return }, Err(e) => { warn!("Error reading block header {}: {}", hash, e); return }, }; // don't announce genesis block since it will be ignored if header.number().is_zero() { return } let is_best = self.chain.info().best_hash == hash; debug!(target: "sync", "Reannouncing block {:?} is_best: {}", hash, is_best); let data = data .or_else(|| self.block_announce_data_cache.get(&hash).cloned()) .unwrap_or_default(); for (who, ref mut peer) in self.peers.iter_mut() { let inserted = peer.known_blocks.insert(hash); if inserted { trace!(target: "sync", "Announcing block {:?} to {}", hash, who); let message = BlockAnnounce { header: header.clone(), state: if is_best { Some(BlockState::Best) } else { Some(BlockState::Normal) }, data: Some(data.clone()), }; self.behaviour .write_notification(who, HARDCODED_PEERSETS_SYNC, message.encode()); } } } /// Push a block announce validation. /// /// It is required that [`ChainSync::poll_block_announce_validation`] is /// called later to check for finished validations. The result of the validation /// needs to be passed to [`Protocol::process_block_announce_validation_result`] /// to finish the processing. /// /// # Note /// /// This will internally create a future, but this future will not be registered /// in the task before being polled once. So, it is required to call /// [`ChainSync::poll_block_announce_validation`] to ensure that the future is /// registered properly and will wake up the task when being ready. fn push_block_announce_validation(&mut self, who: PeerId, announce: BlockAnnounce) { let hash = announce.header.hash(); let peer = match self.peers.get_mut(&who) { Some(p) => p, None => { log::error!(target: "sync", "Received block announce from disconnected peer {}", who); debug_assert!(false); return }, }; peer.known_blocks.insert(hash); let is_best = match announce.state.unwrap_or(BlockState::Best) { BlockState::Best => true, BlockState::Normal => false, }; if peer.info.roles.is_full() { self.chain_sync.push_block_announce_validation(who, hash, announce, is_best); } } /// Process the result of the block announce validation. fn process_block_announce_validation_result( &mut self, validation_result: PollBlockAnnounceValidation, ) -> CustomMessageOutcome { let (header, is_best, who) = match validation_result { PollBlockAnnounceValidation::Skip => return CustomMessageOutcome::None, PollBlockAnnounceValidation::Nothing { is_best, who, announce } => { self.update_peer_info(&who); if let Some(data) = announce.data { if !data.is_empty() { self.block_announce_data_cache.put(announce.header.hash(), data); } } // `on_block_announce` returns `OnBlockAnnounce::ImportHeader` // when we have all data required to import the block // in the BlockAnnounce message. This is only when: // 1) we're on light client; // AND // 2) parent block is already imported and not pruned. if is_best { return CustomMessageOutcome::PeerNewBest(who, *announce.header.number()) } else { return CustomMessageOutcome::None } }, PollBlockAnnounceValidation::ImportHeader { announce, is_best, who } => { self.update_peer_info(&who); if let Some(data) = announce.data { if !data.is_empty() { self.block_announce_data_cache.put(announce.header.hash(), data); } } (announce.header, is_best, who) }, PollBlockAnnounceValidation::Failure { who, disconnect } => { if disconnect { self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); } self.report_peer(who, rep::BAD_BLOCK_ANNOUNCEMENT); return CustomMessageOutcome::None }, }; let number = *header.number(); // to import header from announced block let's construct response to request that normally // would have been sent over network (but it is not in our case) let blocks_to_import = self.chain_sync.on_block_data( &who, None, BlockResponse:: { id: 0, blocks: vec![BlockData:: { hash: header.hash(), header: Some(header), body: None, indexed_body: None, receipt: None, message_queue: None, justification: None, justifications: None, }], }, ); if is_best { self.pending_messages.push_back(CustomMessageOutcome::PeerNewBest(who, number)); } match blocks_to_import { Ok(OnBlockData::Import(origin, blocks)) => CustomMessageOutcome::BlockImport(origin, blocks), Ok(OnBlockData::Request(peer, req)) => prepare_block_request(self.chain_sync.as_ref(), &mut self.peers, peer, req), Ok(OnBlockData::Continue) => CustomMessageOutcome::None, Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu); CustomMessageOutcome::None }, } } /// Call this when a block has been finalized. The sync layer may have some additional /// requesting to perform. pub fn on_block_finalized(&mut self, hash: B::Hash, header: &B::Header) { self.chain_sync.on_block_finalized(&hash, *header.number()) } /// Request a justification for the given block. /// /// Uses `protocol` to queue a new justification request and tries to dispatch all pending /// requests. pub fn request_justification(&mut self, hash: &B::Hash, number: NumberFor) { self.chain_sync.request_justification(hash, number) } /// Clear all pending justification requests. pub fn clear_justification_requests(&mut self) { self.chain_sync.clear_justification_requests(); } /// A batch of blocks have been processed, with or without errors. /// Call this when a batch of blocks have been processed by the importqueue, with or without /// errors. pub fn on_blocks_processed( &mut self, imported: usize, count: usize, results: Vec<(Result>, BlockImportError>, B::Hash)>, ) { let results = self.chain_sync.on_blocks_processed(imported, count, results); for result in results { match result { Ok((id, req)) => { self.pending_messages.push_back(prepare_block_request( self.chain_sync.as_ref(), &mut self.peers, id, req, )); }, Err(BadPeer(id, repu)) => { self.behaviour.disconnect_peer(&id, HARDCODED_PEERSETS_SYNC); self.peerset_handle.report_peer(id, repu) }, } } } /// Call this when a justification has been processed by the import queue, with or without /// errors. pub fn justification_import_result( &mut self, who: PeerId, hash: B::Hash, number: NumberFor, success: bool, ) { self.chain_sync.on_justification_import(hash, number, success); if !success { info!("💔 Invalid justification provided by {} for #{}", who, hash); self.behaviour.disconnect_peer(&who, HARDCODED_PEERSETS_SYNC); self.peerset_handle .report_peer(who, sc_peerset::ReputationChange::new_fatal("Invalid justification")); } } /// Set whether the syncing peers set is in reserved-only mode. pub fn set_reserved_only(&self, reserved_only: bool) { self.peerset_handle.set_reserved_only(HARDCODED_PEERSETS_SYNC, reserved_only); } /// Removes a `PeerId` from the list of reserved peers for syncing purposes. pub fn remove_reserved_peer(&self, peer: PeerId) { self.peerset_handle.remove_reserved_peer(HARDCODED_PEERSETS_SYNC, peer); } /// Returns the list of reserved peers. pub fn reserved_peers(&self) -> impl Iterator { self.behaviour.reserved_peers(HARDCODED_PEERSETS_SYNC) } /// Adds a `PeerId` to the list of reserved peers for syncing purposes. pub fn add_reserved_peer(&self, peer: PeerId) { self.peerset_handle.add_reserved_peer(HARDCODED_PEERSETS_SYNC, peer); } /// Sets the list of reserved peers for syncing purposes. pub fn set_reserved_peers(&self, peers: HashSet) { self.peerset_handle.set_reserved_peers(HARDCODED_PEERSETS_SYNC, peers); } /// Sets the list of reserved peers for the given protocol/peerset. pub fn set_reserved_peerset_peers(&self, protocol: ProtocolName, peers: HashSet) { if let Some(index) = self.notification_protocols.iter().position(|p| *p == protocol) { self.peerset_handle.set_reserved_peers(sc_peerset::SetId::from(index), peers); } else { error!( target: "sub-libp2p", "set_reserved_peerset_peers with unknown protocol: {}", protocol ); } } /// Removes a `PeerId` from the list of reserved peers. pub fn remove_set_reserved_peer(&self, protocol: ProtocolName, peer: PeerId) { if let Some(index) = self.notification_protocols.iter().position(|p| *p == protocol) { self.peerset_handle.remove_reserved_peer(sc_peerset::SetId::from(index), peer); } else { error!( target: "sub-libp2p", "remove_set_reserved_peer with unknown protocol: {}", protocol ); } } /// Adds a `PeerId` to the list of reserved peers. pub fn add_set_reserved_peer(&self, protocol: ProtocolName, peer: PeerId) { if let Some(index) = self.notification_protocols.iter().position(|p| *p == protocol) { self.peerset_handle.add_reserved_peer(sc_peerset::SetId::from(index), peer); } else { error!( target: "sub-libp2p", "add_set_reserved_peer with unknown protocol: {}", protocol ); } } /// Notify the protocol that we have learned about the existence of nodes on the default set. /// /// Can be called multiple times with the same `PeerId`s. pub fn add_default_set_discovered_nodes(&mut self, peer_ids: impl Iterator) { for peer_id in peer_ids { self.peerset_handle.add_to_peers_set(HARDCODED_PEERSETS_SYNC, peer_id); } } /// Add a peer to a peers set. pub fn add_to_peers_set(&self, protocol: ProtocolName, peer: PeerId) { if let Some(index) = self.notification_protocols.iter().position(|p| *p == protocol) { self.peerset_handle.add_to_peers_set(sc_peerset::SetId::from(index), peer); } else { error!( target: "sub-libp2p", "add_to_peers_set with unknown protocol: {}", protocol ); } } /// Remove a peer from a peers set. pub fn remove_from_peers_set(&self, protocol: ProtocolName, peer: PeerId) { if let Some(index) = self.notification_protocols.iter().position(|p| *p == protocol) { self.peerset_handle.remove_from_peers_set(sc_peerset::SetId::from(index), peer); } else { error!( target: "sub-libp2p", "remove_from_peers_set with unknown protocol: {}", protocol ); } } /// Encode implementation-specific block request. pub fn encode_block_request(&self, request: &OpaqueBlockRequest) -> Result, String> { self.chain_sync.encode_block_request(request) } /// Encode implementation-specific state request. pub fn encode_state_request(&self, request: &OpaqueStateRequest) -> Result, String> { self.chain_sync.encode_state_request(request) } fn report_metrics(&self) { if let Some(metrics) = &self.metrics { let n = u64::try_from(self.peers.len()).unwrap_or(std::u64::MAX); metrics.peers.set(n); let m = self.chain_sync.metrics(); metrics.fork_targets.set(m.fork_targets.into()); metrics.queued_blocks.set(m.queued_blocks.into()); metrics .justifications .with_label_values(&["pending"]) .set(m.justifications.pending_requests.into()); metrics .justifications .with_label_values(&["active"]) .set(m.justifications.active_requests.into()); metrics .justifications .with_label_values(&["failed"]) .set(m.justifications.failed_requests.into()); metrics .justifications .with_label_values(&["importing"]) .set(m.justifications.importing_requests.into()); } } } fn prepare_block_request( chain_sync: &dyn ChainSync, peers: &mut HashMap>, who: PeerId, request: BlockRequest, ) -> CustomMessageOutcome { let (tx, rx) = oneshot::channel(); if let Some(ref mut peer) = peers.get_mut(&who) { peer.request = Some((PeerRequest::Block(request.clone()), rx)); } let request = chain_sync.create_opaque_block_request(&request); CustomMessageOutcome::BlockRequest { target: who, request, pending_response: tx } } fn prepare_state_request( peers: &mut HashMap>, who: PeerId, request: OpaqueStateRequest, ) -> CustomMessageOutcome { let (tx, rx) = oneshot::channel(); if let Some(ref mut peer) = peers.get_mut(&who) { peer.request = Some((PeerRequest::State, rx)); } CustomMessageOutcome::StateRequest { target: who, request, pending_response: tx } } fn prepare_warp_sync_request( peers: &mut HashMap>, who: PeerId, request: WarpProofRequest, ) -> CustomMessageOutcome { let (tx, rx) = oneshot::channel(); if let Some(ref mut peer) = peers.get_mut(&who) { peer.request = Some((PeerRequest::WarpProof, rx)); } CustomMessageOutcome::WarpSyncRequest { target: who, request, pending_response: tx } } /// Outcome of an incoming custom message. #[derive(Debug)] #[must_use] pub enum CustomMessageOutcome { BlockImport(BlockOrigin, Vec>), JustificationImport(RuntimeOrigin, B::Hash, NumberFor, Justifications), /// Notification protocols have been opened with a remote. NotificationStreamOpened { remote: PeerId, protocol: ProtocolName, /// See [`crate::Event::NotificationStreamOpened::negotiated_fallback`]. negotiated_fallback: Option, roles: Roles, notifications_sink: NotificationsSink, }, /// The [`NotificationsSink`] of some notification protocols need an update. NotificationStreamReplaced { remote: PeerId, protocol: ProtocolName, notifications_sink: NotificationsSink, }, /// Notification protocols have been closed with a remote. NotificationStreamClosed { remote: PeerId, protocol: ProtocolName, }, /// Messages have been received on one or more notifications protocols. NotificationsReceived { remote: PeerId, messages: Vec<(ProtocolName, Bytes)>, }, /// A new block request must be emitted. BlockRequest { target: PeerId, request: OpaqueBlockRequest, pending_response: oneshot::Sender, RequestFailure>>, }, /// A new storage request must be emitted. StateRequest { target: PeerId, request: OpaqueStateRequest, pending_response: oneshot::Sender, RequestFailure>>, }, /// A new warp sync request must be emitted. WarpSyncRequest { target: PeerId, request: WarpProofRequest, pending_response: oneshot::Sender, RequestFailure>>, }, /// Peer has a reported a new head of chain. PeerNewBest(PeerId, NumberFor), /// Now connected to a new peer for syncing purposes. SyncConnected(PeerId), /// No longer connected to a peer for syncing purposes. SyncDisconnected(PeerId), None, } impl NetworkBehaviour for Protocol where B: BlockT, Client: HeaderBackend + 'static, { type ConnectionHandler = ::ConnectionHandler; type OutEvent = CustomMessageOutcome; fn new_handler(&mut self) -> Self::ConnectionHandler { self.behaviour.new_handler() } fn addresses_of_peer(&mut self, peer_id: &PeerId) -> Vec { self.behaviour.addresses_of_peer(peer_id) } fn inject_connection_established( &mut self, peer_id: &PeerId, conn: &ConnectionId, endpoint: &ConnectedPoint, failed_addresses: Option<&Vec>, other_established: usize, ) { self.behaviour.inject_connection_established( peer_id, conn, endpoint, failed_addresses, other_established, ) } fn inject_connection_closed( &mut self, peer_id: &PeerId, conn: &ConnectionId, endpoint: &ConnectedPoint, handler: ::Handler, remaining_established: usize, ) { self.behaviour.inject_connection_closed( peer_id, conn, endpoint, handler, remaining_established, ) } fn inject_event( &mut self, peer_id: PeerId, connection: ConnectionId, event: <::Handler as ConnectionHandler>::OutEvent, ) { self.behaviour.inject_event(peer_id, connection, event) } fn poll( &mut self, cx: &mut std::task::Context, params: &mut impl PollParameters, ) -> Poll> { if let Some(message) = self.pending_messages.pop_front() { return Poll::Ready(NetworkBehaviourAction::GenerateEvent(message)) } // Check for finished outgoing requests. let mut finished_block_requests = Vec::new(); let mut finished_state_requests = Vec::new(); let mut finished_warp_sync_requests = Vec::new(); for (id, peer) in self.peers.iter_mut() { if let Peer { request: Some((_, pending_response)), .. } = peer { match pending_response.poll_unpin(cx) { Poll::Ready(Ok(Ok(resp))) => { let (req, _) = peer.request.take().unwrap(); match req { PeerRequest::Block(req) => { let response = match self.chain_sync.decode_block_response(&resp[..]) { Ok(proto) => proto, Err(e) => { debug!( target: "sync", "Failed to decode block response from peer {:?}: {:?}.", id, e ); self.peerset_handle.report_peer(*id, rep::BAD_MESSAGE); self.behaviour .disconnect_peer(id, HARDCODED_PEERSETS_SYNC); continue }, }; finished_block_requests.push((*id, req, response)); }, PeerRequest::State => { let response = match self.chain_sync.decode_state_response(&resp[..]) { Ok(proto) => proto, Err(e) => { debug!( target: "sync", "Failed to decode state response from peer {:?}: {:?}.", id, e ); self.peerset_handle.report_peer(*id, rep::BAD_MESSAGE); self.behaviour .disconnect_peer(id, HARDCODED_PEERSETS_SYNC); continue }, }; finished_state_requests.push((*id, response)); }, PeerRequest::WarpProof => { finished_warp_sync_requests.push((*id, resp)); }, } }, Poll::Ready(Ok(Err(e))) => { peer.request.take(); debug!(target: "sync", "Request to peer {:?} failed: {:?}.", id, e); match e { RequestFailure::Network(OutboundFailure::Timeout) => { self.peerset_handle.report_peer(*id, rep::TIMEOUT); self.behaviour.disconnect_peer(id, HARDCODED_PEERSETS_SYNC); }, RequestFailure::Network(OutboundFailure::UnsupportedProtocols) => { self.peerset_handle.report_peer(*id, rep::BAD_PROTOCOL); self.behaviour.disconnect_peer(id, HARDCODED_PEERSETS_SYNC); }, RequestFailure::Network(OutboundFailure::DialFailure) => { self.behaviour.disconnect_peer(id, HARDCODED_PEERSETS_SYNC); }, RequestFailure::Refused => { self.peerset_handle.report_peer(*id, rep::REFUSED); self.behaviour.disconnect_peer(id, HARDCODED_PEERSETS_SYNC); }, RequestFailure::Network(OutboundFailure::ConnectionClosed) | RequestFailure::NotConnected => { self.behaviour.disconnect_peer(id, HARDCODED_PEERSETS_SYNC); }, RequestFailure::UnknownProtocol => { debug_assert!( false, "Block request protocol should always be known." ); }, RequestFailure::Obsolete => { debug_assert!( false, "Can not receive `RequestFailure::Obsolete` after dropping the \ response receiver.", ); }, } }, Poll::Ready(Err(oneshot::Canceled)) => { peer.request.take(); trace!( target: "sync", "Request to peer {:?} failed due to oneshot being canceled.", id, ); self.behaviour.disconnect_peer(id, HARDCODED_PEERSETS_SYNC); }, Poll::Pending => {}, } } } for (id, req, response) in finished_block_requests { let ev = self.on_block_response(id, req, response); self.pending_messages.push_back(ev); } for (id, response) in finished_state_requests { let ev = self.on_state_response(id, response); self.pending_messages.push_back(ev); } for (id, response) in finished_warp_sync_requests { let ev = self.on_warp_sync_response(id, EncodedProof(response)); self.pending_messages.push_back(ev); } while let Poll::Ready(Some(())) = self.tick_timeout.poll_next_unpin(cx) { self.tick(); } for (id, request) in self .chain_sync .block_requests() .map(|(peer_id, request)| (peer_id, request)) .collect::>() { let event = prepare_block_request(self.chain_sync.as_ref(), &mut self.peers, id, request); self.pending_messages.push_back(event); } if let Some((id, request)) = self.chain_sync.state_request() { let event = prepare_state_request(&mut self.peers, id, request); self.pending_messages.push_back(event); } for (id, request) in self.chain_sync.justification_requests().collect::>() { let event = prepare_block_request(self.chain_sync.as_ref(), &mut self.peers, id, request); self.pending_messages.push_back(event); } if let Some((id, request)) = self.chain_sync.warp_sync_request() { let event = prepare_warp_sync_request(&mut self.peers, id, request); self.pending_messages.push_back(event); } // Advance the state of `ChainSync` // // Process any received requests received from `NetworkService` and // check if there is any block announcement validation finished. while let Poll::Ready(result) = self.chain_sync.poll(cx) { match self.process_block_announce_validation_result(result) { CustomMessageOutcome::None => {}, outcome => self.pending_messages.push_back(outcome), } } if let Some(message) = self.pending_messages.pop_front() { return Poll::Ready(NetworkBehaviourAction::GenerateEvent(message)) } let event = match self.behaviour.poll(cx, params) { Poll::Pending => return Poll::Pending, Poll::Ready(NetworkBehaviourAction::GenerateEvent(ev)) => ev, Poll::Ready(NetworkBehaviourAction::Dial { opts, handler }) => return Poll::Ready(NetworkBehaviourAction::Dial { opts, handler }), Poll::Ready(NetworkBehaviourAction::NotifyHandler { peer_id, handler, event }) => return Poll::Ready(NetworkBehaviourAction::NotifyHandler { peer_id, handler, event, }), Poll::Ready(NetworkBehaviourAction::ReportObservedAddr { address, score }) => return Poll::Ready(NetworkBehaviourAction::ReportObservedAddr { address, score }), Poll::Ready(NetworkBehaviourAction::CloseConnection { peer_id, connection }) => return Poll::Ready(NetworkBehaviourAction::CloseConnection { peer_id, connection }), }; let outcome = match event { NotificationsOut::CustomProtocolOpen { peer_id, set_id, received_handshake, notifications_sink, negotiated_fallback, } => { // Set number 0 is hardcoded the default set of peers we sync from. if set_id == HARDCODED_PEERSETS_SYNC { // `received_handshake` can be either a `Status` message if received from the // legacy substream ,or a `BlockAnnouncesHandshake` if received from the block // announces substream. match as DecodeAll>::decode_all(&mut &received_handshake[..]) { Ok(GenericMessage::Status(handshake)) => { let handshake = BlockAnnouncesHandshake { roles: handshake.roles, best_number: handshake.best_number, best_hash: handshake.best_hash, genesis_hash: handshake.genesis_hash, }; if self.on_sync_peer_connected(peer_id, handshake).is_ok() { CustomMessageOutcome::SyncConnected(peer_id) } else { CustomMessageOutcome::None } }, Ok(msg) => { debug!( target: "sync", "Expected Status message from {}, but got {:?}", peer_id, msg, ); self.peerset_handle.report_peer(peer_id, rep::BAD_MESSAGE); CustomMessageOutcome::None }, Err(err) => { match as DecodeAll>::decode_all( &mut &received_handshake[..], ) { Ok(handshake) => { if self.on_sync_peer_connected(peer_id, handshake).is_ok() { CustomMessageOutcome::SyncConnected(peer_id) } else { CustomMessageOutcome::None } }, Err(err2) => { debug!( target: "sync", "Couldn't decode handshake sent by {}: {:?}: {} & {}", peer_id, received_handshake, err, err2, ); self.peerset_handle.report_peer(peer_id, rep::BAD_MESSAGE); CustomMessageOutcome::None }, } }, } } else { match ( Roles::decode_all(&mut &received_handshake[..]), self.peers.get(&peer_id), ) { (Ok(roles), _) => CustomMessageOutcome::NotificationStreamOpened { remote: peer_id, protocol: self.notification_protocols[usize::from(set_id)].clone(), negotiated_fallback, roles, notifications_sink, }, (Err(_), Some(peer)) if received_handshake.is_empty() => { // As a convenience, we allow opening substreams for "external" // notification protocols with an empty handshake. This fetches the // roles from the locally-known roles. // TODO: remove this after https://github.com/paritytech/substrate/issues/5685 CustomMessageOutcome::NotificationStreamOpened { remote: peer_id, protocol: self.notification_protocols[usize::from(set_id)].clone(), negotiated_fallback, roles: peer.info.roles, notifications_sink, } }, (Err(err), _) => { debug!(target: "sync", "Failed to parse remote handshake: {}", err); self.bad_handshake_substreams.insert((peer_id, set_id)); self.behaviour.disconnect_peer(&peer_id, set_id); self.peerset_handle.report_peer(peer_id, rep::BAD_MESSAGE); CustomMessageOutcome::None }, } } }, NotificationsOut::CustomProtocolReplaced { peer_id, notifications_sink, set_id } => if set_id == HARDCODED_PEERSETS_SYNC || self.bad_handshake_substreams.contains(&(peer_id, set_id)) { CustomMessageOutcome::None } else { CustomMessageOutcome::NotificationStreamReplaced { remote: peer_id, protocol: self.notification_protocols[usize::from(set_id)].clone(), notifications_sink, } }, NotificationsOut::CustomProtocolClosed { peer_id, set_id } => { // Set number 0 is hardcoded the default set of peers we sync from. if set_id == HARDCODED_PEERSETS_SYNC { if self.on_sync_peer_disconnected(peer_id).is_ok() { CustomMessageOutcome::SyncDisconnected(peer_id) } else { log::trace!( target: "sync", "Disconnected peer which had earlier been refused by on_sync_peer_connected {}", peer_id ); CustomMessageOutcome::None } } else if self.bad_handshake_substreams.remove(&(peer_id, set_id)) { // The substream that has just been closed had been opened with a bad // handshake. The outer layers have never received an opening event about this // substream, and consequently shouldn't receive a closing event either. CustomMessageOutcome::None } else { CustomMessageOutcome::NotificationStreamClosed { remote: peer_id, protocol: self.notification_protocols[usize::from(set_id)].clone(), } } }, NotificationsOut::Notification { peer_id, set_id, message } => match set_id { HARDCODED_PEERSETS_SYNC if self.peers.contains_key(&peer_id) => { if let Ok(announce) = BlockAnnounce::decode(&mut message.as_ref()) { self.push_block_announce_validation(peer_id, announce); // Make sure that the newly added block announce validation future was // polled once to be registered in the task. if let Poll::Ready(res) = self.chain_sync.poll_block_announce_validation(cx) { self.process_block_announce_validation_result(res) } else { CustomMessageOutcome::None } } else { warn!(target: "sub-libp2p", "Failed to decode block announce"); CustomMessageOutcome::None } }, HARDCODED_PEERSETS_SYNC => { trace!( target: "sync", "Received sync for peer earlier refused by sync layer: {}", peer_id ); CustomMessageOutcome::None }, _ if self.bad_handshake_substreams.contains(&(peer_id, set_id)) => CustomMessageOutcome::None, _ => { let protocol_name = self.notification_protocols[usize::from(set_id)].clone(); CustomMessageOutcome::NotificationsReceived { remote: peer_id, messages: vec![(protocol_name, message.freeze())], } }, }, }; if !matches!(outcome, CustomMessageOutcome::::None) { return Poll::Ready(NetworkBehaviourAction::GenerateEvent(outcome)) } if let Some(message) = self.pending_messages.pop_front() { return Poll::Ready(NetworkBehaviourAction::GenerateEvent(message)) } // This block can only be reached if an event was pulled from the behaviour and that // resulted in `CustomMessageOutcome::None`. Since there might be another pending // message from the behaviour, the task is scheduled again. cx.waker().wake_by_ref(); Poll::Pending } fn inject_dial_failure( &mut self, peer_id: Option, handler: Self::ConnectionHandler, error: &libp2p::swarm::DialError, ) { self.behaviour.inject_dial_failure(peer_id, handler, error); } fn inject_new_listener(&mut self, id: ListenerId) { self.behaviour.inject_new_listener(id) } fn inject_new_listen_addr(&mut self, id: ListenerId, addr: &Multiaddr) { self.behaviour.inject_new_listen_addr(id, addr) } fn inject_expired_listen_addr(&mut self, id: ListenerId, addr: &Multiaddr) { self.behaviour.inject_expired_listen_addr(id, addr) } fn inject_new_external_addr(&mut self, addr: &Multiaddr) { self.behaviour.inject_new_external_addr(addr) } fn inject_expired_external_addr(&mut self, addr: &Multiaddr) { self.behaviour.inject_expired_external_addr(addr) } fn inject_listener_error(&mut self, id: ListenerId, err: &(dyn std::error::Error + 'static)) { self.behaviour.inject_listener_error(id, err); } fn inject_listener_closed(&mut self, id: ListenerId, reason: Result<(), &io::Error>) { self.behaviour.inject_listener_closed(id, reason); } }