mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-26 15:47:58 +00:00
80616f6d03
[litep2p](https://github.com/altonen/litep2p) is a libp2p-compatible P2P networking library. It supports all of the features of `rust-libp2p` that are currently being utilized by Polkadot SDK. Compared to `rust-libp2p`, `litep2p` has a quite different architecture which is why the new `litep2p` network backend is only able to use a little of the existing code in `sc-network`. The design has been mainly influenced by how we'd wish to structure our networking-related code in Polkadot SDK: independent higher-levels protocols directly communicating with the network over links that support bidirectional backpressure. A good example would be `NotificationHandle`/`RequestResponseHandle` abstractions which allow, e.g., `SyncingEngine` to directly communicate with peers to announce/request blocks. I've tried running `polkadot --network-backend litep2p` with a few different peer configurations and there is a noticeable reduction in networking CPU usage. For high load (`--out-peers 200`), networking CPU usage goes down from ~110% to ~30% (80 pp) and for normal load (`--out-peers 40`), the usage goes down from ~55% to ~18% (37 pp). These should not be taken as final numbers because: a) there are still some low-hanging optimization fruits, such as enabling [receive window auto-tuning](https://github.com/libp2p/rust-yamux/pull/176), integrating `Peerset` more closely with `litep2p` or improving memory usage of the WebSocket transport b) fixing bugs/instabilities that incorrectly cause `litep2p` to do less work will increase the networking CPU usage c) verification in a more diverse set of tests/conditions is needed Nevertheless, these numbers should give an early estimate for CPU usage of the new networking backend. This PR consists of three separate changes: * introduce a generic `PeerId` (wrapper around `Multihash`) so that we don't have use `NetworkService::PeerId` in every part of the code that uses a `PeerId` * introduce `NetworkBackend` trait, implement it for the libp2p network stack and make Polkadot SDK generic over `NetworkBackend` * implement `NetworkBackend` for litep2p The new library should be considered experimental which is why `rust-libp2p` will remain as the default option for the time being. This PR currently depends on the master branch of `litep2p` but I'll cut a new release for the library once all review comments have been addresses. --------- Signed-off-by: Alexandru Vasile <alexandru.vasile@parity.io> Co-authored-by: Dmitry Markin <dmitry@markin.tech> Co-authored-by: Alexandru Vasile <60601340+lexnv@users.noreply.github.com> Co-authored-by: Alexandru Vasile <alexandru.vasile@parity.io>
662 lines
21 KiB
Rust
662 lines
21 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
//! [`SyncingStrategy`] is a proxy between [`crate::engine::SyncingEngine`]
|
|
//! and specific syncing algorithms.
|
|
|
|
pub mod chain_sync;
|
|
mod state;
|
|
pub mod state_sync;
|
|
pub mod warp;
|
|
|
|
use crate::{
|
|
types::{BadPeer, OpaqueStateRequest, OpaqueStateResponse, SyncStatus},
|
|
LOG_TARGET,
|
|
};
|
|
use chain_sync::{ChainSync, ChainSyncAction, ChainSyncMode};
|
|
use log::{debug, error, info, warn};
|
|
use prometheus_endpoint::Registry;
|
|
use sc_client_api::{BlockBackend, ProofProvider};
|
|
use sc_consensus::{BlockImportError, BlockImportStatus, IncomingBlock};
|
|
use sc_network_common::sync::{
|
|
message::{BlockAnnounce, BlockData, BlockRequest},
|
|
SyncMode,
|
|
};
|
|
use sc_network_types::PeerId;
|
|
use sp_blockchain::{Error as ClientError, HeaderBackend, HeaderMetadata};
|
|
use sp_consensus::BlockOrigin;
|
|
use sp_runtime::{
|
|
traits::{Block as BlockT, Header, NumberFor},
|
|
Justifications,
|
|
};
|
|
use state::{StateStrategy, StateStrategyAction};
|
|
use std::{collections::HashMap, sync::Arc};
|
|
use warp::{EncodedProof, WarpProofRequest, WarpSync, WarpSyncAction, WarpSyncConfig};
|
|
|
|
/// Corresponding `ChainSync` mode.
|
|
fn chain_sync_mode(sync_mode: SyncMode) -> ChainSyncMode {
|
|
match sync_mode {
|
|
SyncMode::Full => ChainSyncMode::Full,
|
|
SyncMode::LightState { skip_proofs, storage_chain_mode } =>
|
|
ChainSyncMode::LightState { skip_proofs, storage_chain_mode },
|
|
SyncMode::Warp => ChainSyncMode::Full,
|
|
}
|
|
}
|
|
|
|
/// Syncing configuration containing data for all strategies.
|
|
#[derive(Clone, Debug)]
|
|
pub struct SyncingConfig {
|
|
/// Syncing mode.
|
|
pub mode: SyncMode,
|
|
/// The number of parallel downloads to guard against slow peers.
|
|
pub max_parallel_downloads: u32,
|
|
/// Maximum number of blocks to request.
|
|
pub max_blocks_per_request: u32,
|
|
/// Prometheus metrics registry.
|
|
pub metrics_registry: Option<Registry>,
|
|
}
|
|
|
|
/// The key identifying a specific strategy for responses routing.
|
|
#[derive(Debug, Clone, Copy, PartialEq, Eq, Hash)]
|
|
pub enum StrategyKey {
|
|
/// Warp sync initiated this request.
|
|
Warp,
|
|
/// State sync initiated this request.
|
|
State,
|
|
/// `ChainSync` initiated this request.
|
|
ChainSync,
|
|
}
|
|
|
|
#[derive(Debug)]
|
|
pub enum SyncingAction<B: BlockT> {
|
|
/// Send block request to peer. Always implies dropping a stale block request to the same peer.
|
|
SendBlockRequest { peer_id: PeerId, key: StrategyKey, request: BlockRequest<B> },
|
|
/// Send state request to peer.
|
|
SendStateRequest { peer_id: PeerId, key: StrategyKey, request: OpaqueStateRequest },
|
|
/// Send warp proof request to peer.
|
|
SendWarpProofRequest { peer_id: PeerId, key: StrategyKey, request: WarpProofRequest<B> },
|
|
/// Drop stale request.
|
|
CancelRequest { peer_id: PeerId, key: StrategyKey },
|
|
/// Peer misbehaved. Disconnect, report it and cancel any requests to it.
|
|
DropPeer(BadPeer),
|
|
/// Import blocks.
|
|
ImportBlocks { origin: BlockOrigin, blocks: Vec<IncomingBlock<B>> },
|
|
/// Import justifications.
|
|
ImportJustifications {
|
|
peer_id: PeerId,
|
|
hash: B::Hash,
|
|
number: NumberFor<B>,
|
|
justifications: Justifications,
|
|
},
|
|
/// Strategy finished. Nothing to do, this is handled by `SyncingStrategy`.
|
|
Finished,
|
|
}
|
|
|
|
impl<B: BlockT> SyncingAction<B> {
|
|
fn is_finished(&self) -> bool {
|
|
matches!(self, SyncingAction::Finished)
|
|
}
|
|
}
|
|
|
|
impl<B: BlockT> From<WarpSyncAction<B>> for SyncingAction<B> {
|
|
fn from(action: WarpSyncAction<B>) -> Self {
|
|
match action {
|
|
WarpSyncAction::SendWarpProofRequest { peer_id, request } =>
|
|
SyncingAction::SendWarpProofRequest { peer_id, key: StrategyKey::Warp, request },
|
|
WarpSyncAction::SendBlockRequest { peer_id, request } =>
|
|
SyncingAction::SendBlockRequest { peer_id, key: StrategyKey::Warp, request },
|
|
WarpSyncAction::DropPeer(bad_peer) => SyncingAction::DropPeer(bad_peer),
|
|
WarpSyncAction::Finished => SyncingAction::Finished,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<B: BlockT> From<StateStrategyAction<B>> for SyncingAction<B> {
|
|
fn from(action: StateStrategyAction<B>) -> Self {
|
|
match action {
|
|
StateStrategyAction::SendStateRequest { peer_id, request } =>
|
|
SyncingAction::SendStateRequest { peer_id, key: StrategyKey::State, request },
|
|
StateStrategyAction::DropPeer(bad_peer) => SyncingAction::DropPeer(bad_peer),
|
|
StateStrategyAction::ImportBlocks { origin, blocks } =>
|
|
SyncingAction::ImportBlocks { origin, blocks },
|
|
StateStrategyAction::Finished => SyncingAction::Finished,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<B: BlockT> From<ChainSyncAction<B>> for SyncingAction<B> {
|
|
fn from(action: ChainSyncAction<B>) -> Self {
|
|
match action {
|
|
ChainSyncAction::SendBlockRequest { peer_id, request } =>
|
|
SyncingAction::SendBlockRequest { peer_id, key: StrategyKey::ChainSync, request },
|
|
ChainSyncAction::SendStateRequest { peer_id, request } =>
|
|
SyncingAction::SendStateRequest { peer_id, key: StrategyKey::ChainSync, request },
|
|
ChainSyncAction::CancelRequest { peer_id } =>
|
|
SyncingAction::CancelRequest { peer_id, key: StrategyKey::ChainSync },
|
|
ChainSyncAction::DropPeer(bad_peer) => SyncingAction::DropPeer(bad_peer),
|
|
ChainSyncAction::ImportBlocks { origin, blocks } =>
|
|
SyncingAction::ImportBlocks { origin, blocks },
|
|
ChainSyncAction::ImportJustifications { peer_id, hash, number, justifications } =>
|
|
SyncingAction::ImportJustifications { peer_id, hash, number, justifications },
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Proxy to specific syncing strategies.
|
|
pub struct SyncingStrategy<B: BlockT, Client> {
|
|
/// Initial syncing configuration.
|
|
config: SyncingConfig,
|
|
/// Client used by syncing strategies.
|
|
client: Arc<Client>,
|
|
/// Warp strategy.
|
|
warp: Option<WarpSync<B, Client>>,
|
|
/// State strategy.
|
|
state: Option<StateStrategy<B>>,
|
|
/// `ChainSync` strategy.`
|
|
chain_sync: Option<ChainSync<B, Client>>,
|
|
/// Connected peers and their best blocks used to seed a new strategy when switching to it in
|
|
/// [`SyncingStrategy::proceed_to_next`].
|
|
peer_best_blocks: HashMap<PeerId, (B::Hash, NumberFor<B>)>,
|
|
}
|
|
|
|
impl<B: BlockT, Client> SyncingStrategy<B, Client>
|
|
where
|
|
B: BlockT,
|
|
Client: HeaderBackend<B>
|
|
+ BlockBackend<B>
|
|
+ HeaderMetadata<B, Error = sp_blockchain::Error>
|
|
+ ProofProvider<B>
|
|
+ Send
|
|
+ Sync
|
|
+ 'static,
|
|
{
|
|
/// Initialize a new syncing strategy.
|
|
pub fn new(
|
|
config: SyncingConfig,
|
|
client: Arc<Client>,
|
|
warp_sync_config: Option<WarpSyncConfig<B>>,
|
|
) -> Result<Self, ClientError> {
|
|
if let SyncMode::Warp = config.mode {
|
|
let warp_sync_config = warp_sync_config
|
|
.expect("Warp sync configuration must be supplied in warp sync mode.");
|
|
let warp_sync = WarpSync::new(client.clone(), warp_sync_config);
|
|
Ok(Self {
|
|
config,
|
|
client,
|
|
warp: Some(warp_sync),
|
|
state: None,
|
|
chain_sync: None,
|
|
peer_best_blocks: Default::default(),
|
|
})
|
|
} else {
|
|
let chain_sync = ChainSync::new(
|
|
chain_sync_mode(config.mode),
|
|
client.clone(),
|
|
config.max_parallel_downloads,
|
|
config.max_blocks_per_request,
|
|
config.metrics_registry.clone(),
|
|
std::iter::empty(),
|
|
)?;
|
|
Ok(Self {
|
|
config,
|
|
client,
|
|
warp: None,
|
|
state: None,
|
|
chain_sync: Some(chain_sync),
|
|
peer_best_blocks: Default::default(),
|
|
})
|
|
}
|
|
}
|
|
|
|
/// Notify that a new peer has connected.
|
|
pub fn add_peer(&mut self, peer_id: PeerId, best_hash: B::Hash, best_number: NumberFor<B>) {
|
|
self.peer_best_blocks.insert(peer_id, (best_hash, best_number));
|
|
|
|
self.warp.as_mut().map(|s| s.add_peer(peer_id, best_hash, best_number));
|
|
self.state.as_mut().map(|s| s.add_peer(peer_id, best_hash, best_number));
|
|
self.chain_sync.as_mut().map(|s| s.add_peer(peer_id, best_hash, best_number));
|
|
}
|
|
|
|
/// Notify that a peer has disconnected.
|
|
pub fn remove_peer(&mut self, peer_id: &PeerId) {
|
|
self.warp.as_mut().map(|s| s.remove_peer(peer_id));
|
|
self.state.as_mut().map(|s| s.remove_peer(peer_id));
|
|
self.chain_sync.as_mut().map(|s| s.remove_peer(peer_id));
|
|
|
|
self.peer_best_blocks.remove(peer_id);
|
|
}
|
|
|
|
/// Submit a validated block announcement.
|
|
///
|
|
/// Returns new best hash & best number of the peer if they are updated.
|
|
pub fn on_validated_block_announce(
|
|
&mut self,
|
|
is_best: bool,
|
|
peer_id: PeerId,
|
|
announce: &BlockAnnounce<B::Header>,
|
|
) -> Option<(B::Hash, NumberFor<B>)> {
|
|
let new_best = if let Some(ref mut warp) = self.warp {
|
|
warp.on_validated_block_announce(is_best, peer_id, announce)
|
|
} else if let Some(ref mut state) = self.state {
|
|
state.on_validated_block_announce(is_best, peer_id, announce)
|
|
} else if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.on_validated_block_announce(is_best, peer_id, announce)
|
|
} else {
|
|
error!(target: LOG_TARGET, "No syncing strategy is active.");
|
|
debug_assert!(false);
|
|
Some((announce.header.hash(), *announce.header.number()))
|
|
};
|
|
|
|
if let Some(new_best) = new_best {
|
|
if let Some(best) = self.peer_best_blocks.get_mut(&peer_id) {
|
|
*best = new_best;
|
|
} else {
|
|
debug!(
|
|
target: LOG_TARGET,
|
|
"Cannot update `peer_best_blocks` as peer {peer_id} is not known to `Strategy` \
|
|
(already disconnected?)",
|
|
);
|
|
}
|
|
}
|
|
|
|
new_best
|
|
}
|
|
|
|
/// Configure an explicit fork sync request in case external code has detected that there is a
|
|
/// stale fork missing.
|
|
pub fn set_sync_fork_request(
|
|
&mut self,
|
|
peers: Vec<PeerId>,
|
|
hash: &B::Hash,
|
|
number: NumberFor<B>,
|
|
) {
|
|
// Fork requests are only handled by `ChainSync`.
|
|
if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.set_sync_fork_request(peers.clone(), hash, number);
|
|
}
|
|
}
|
|
|
|
/// Request extra justification.
|
|
pub fn request_justification(&mut self, hash: &B::Hash, number: NumberFor<B>) {
|
|
// Justifications can only be requested via `ChainSync`.
|
|
if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.request_justification(hash, number);
|
|
}
|
|
}
|
|
|
|
/// Clear extra justification requests.
|
|
pub fn clear_justification_requests(&mut self) {
|
|
// Justification requests can only be cleared by `ChainSync`.
|
|
if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.clear_justification_requests();
|
|
}
|
|
}
|
|
|
|
/// Report a justification import (successful or not).
|
|
pub fn on_justification_import(&mut self, hash: B::Hash, number: NumberFor<B>, success: bool) {
|
|
// Only `ChainSync` is interested in justification import.
|
|
if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.on_justification_import(hash, number, success);
|
|
}
|
|
}
|
|
|
|
/// Process block response.
|
|
pub fn on_block_response(
|
|
&mut self,
|
|
peer_id: PeerId,
|
|
key: StrategyKey,
|
|
request: BlockRequest<B>,
|
|
blocks: Vec<BlockData<B>>,
|
|
) {
|
|
if let (StrategyKey::Warp, Some(ref mut warp)) = (key, &mut self.warp) {
|
|
warp.on_block_response(peer_id, request, blocks);
|
|
} else if let (StrategyKey::ChainSync, Some(ref mut chain_sync)) =
|
|
(key, &mut self.chain_sync)
|
|
{
|
|
chain_sync.on_block_response(peer_id, request, blocks);
|
|
} else {
|
|
error!(
|
|
target: LOG_TARGET,
|
|
"`on_block_response()` called with unexpected key {key:?} \
|
|
or corresponding strategy is not active.",
|
|
);
|
|
debug_assert!(false);
|
|
}
|
|
}
|
|
|
|
/// Process state response.
|
|
pub fn on_state_response(
|
|
&mut self,
|
|
peer_id: PeerId,
|
|
key: StrategyKey,
|
|
response: OpaqueStateResponse,
|
|
) {
|
|
if let (StrategyKey::State, Some(ref mut state)) = (key, &mut self.state) {
|
|
state.on_state_response(peer_id, response);
|
|
} else if let (StrategyKey::ChainSync, Some(ref mut chain_sync)) =
|
|
(key, &mut self.chain_sync)
|
|
{
|
|
chain_sync.on_state_response(peer_id, response);
|
|
} else {
|
|
error!(
|
|
target: LOG_TARGET,
|
|
"`on_state_response()` called with unexpected key {key:?} \
|
|
or corresponding strategy is not active.",
|
|
);
|
|
debug_assert!(false);
|
|
}
|
|
}
|
|
|
|
/// Process warp proof response.
|
|
pub fn on_warp_proof_response(
|
|
&mut self,
|
|
peer_id: &PeerId,
|
|
key: StrategyKey,
|
|
response: EncodedProof,
|
|
) {
|
|
if let (StrategyKey::Warp, Some(ref mut warp)) = (key, &mut self.warp) {
|
|
warp.on_warp_proof_response(peer_id, response);
|
|
} else {
|
|
error!(
|
|
target: LOG_TARGET,
|
|
"`on_warp_proof_response()` called with unexpected key {key:?} \
|
|
or warp strategy is not active",
|
|
);
|
|
debug_assert!(false);
|
|
}
|
|
}
|
|
|
|
/// A batch of blocks have been processed, with or without errors.
|
|
pub fn on_blocks_processed(
|
|
&mut self,
|
|
imported: usize,
|
|
count: usize,
|
|
results: Vec<(Result<BlockImportStatus<NumberFor<B>>, BlockImportError>, B::Hash)>,
|
|
) {
|
|
// Only `StateStrategy` and `ChainSync` are interested in block processing notifications.
|
|
if let Some(ref mut state) = self.state {
|
|
state.on_blocks_processed(imported, count, results);
|
|
} else if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.on_blocks_processed(imported, count, results);
|
|
}
|
|
}
|
|
|
|
/// Notify a syncing strategy that a block has been finalized.
|
|
pub fn on_block_finalized(&mut self, hash: &B::Hash, number: NumberFor<B>) {
|
|
// Only `ChainSync` is interested in block finalization notifications.
|
|
if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.on_block_finalized(hash, number);
|
|
}
|
|
}
|
|
|
|
/// Inform sync about a new best imported block.
|
|
pub fn update_chain_info(&mut self, best_hash: &B::Hash, best_number: NumberFor<B>) {
|
|
// This is relevant to `ChainSync` only.
|
|
if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.update_chain_info(best_hash, best_number);
|
|
}
|
|
}
|
|
|
|
// Are we in major sync mode?
|
|
pub fn is_major_syncing(&self) -> bool {
|
|
self.warp.is_some() ||
|
|
self.state.is_some() ||
|
|
match self.chain_sync {
|
|
Some(ref s) => s.status().state.is_major_syncing(),
|
|
None => unreachable!("At least one syncing strategy is active; qed"),
|
|
}
|
|
}
|
|
|
|
/// Get the number of peers known to the syncing strategy.
|
|
pub fn num_peers(&self) -> usize {
|
|
self.peer_best_blocks.len()
|
|
}
|
|
|
|
/// Returns the current sync status.
|
|
pub fn status(&self) -> SyncStatus<B> {
|
|
// This function presumes that strategies are executed serially and must be refactored
|
|
// once we have parallel strategies.
|
|
if let Some(ref warp) = self.warp {
|
|
warp.status()
|
|
} else if let Some(ref state) = self.state {
|
|
state.status()
|
|
} else if let Some(ref chain_sync) = self.chain_sync {
|
|
chain_sync.status()
|
|
} else {
|
|
unreachable!("At least one syncing strategy is always active; qed")
|
|
}
|
|
}
|
|
|
|
/// Get the total number of downloaded blocks.
|
|
pub fn num_downloaded_blocks(&self) -> usize {
|
|
self.chain_sync
|
|
.as_ref()
|
|
.map_or(0, |chain_sync| chain_sync.num_downloaded_blocks())
|
|
}
|
|
|
|
/// Get an estimate of the number of parallel sync requests.
|
|
pub fn num_sync_requests(&self) -> usize {
|
|
self.chain_sync.as_ref().map_or(0, |chain_sync| chain_sync.num_sync_requests())
|
|
}
|
|
|
|
/// Report Prometheus metrics
|
|
pub fn report_metrics(&self) {
|
|
if let Some(ref chain_sync) = self.chain_sync {
|
|
chain_sync.report_metrics();
|
|
}
|
|
}
|
|
|
|
/// Let `WarpSync` know about target block header
|
|
pub fn set_warp_sync_target_block_header(
|
|
&mut self,
|
|
target_header: B::Header,
|
|
) -> Result<(), ()> {
|
|
match self.config.mode {
|
|
SyncMode::Warp => match self.warp {
|
|
Some(ref mut warp) => {
|
|
warp.set_target_block(target_header);
|
|
Ok(())
|
|
},
|
|
None => {
|
|
// As mode is set to warp sync, but no warp sync strategy is active, this means
|
|
// that warp sync has already finished / was skipped.
|
|
warn!(
|
|
target: LOG_TARGET,
|
|
"Discarding warp sync target, as warp sync was seemingly skipped due \
|
|
to node being (partially) synced.",
|
|
);
|
|
Ok(())
|
|
},
|
|
},
|
|
_ => {
|
|
error!(
|
|
target: LOG_TARGET,
|
|
"Cannot set warp sync target block: not in warp sync mode."
|
|
);
|
|
debug_assert!(false);
|
|
Err(())
|
|
},
|
|
}
|
|
}
|
|
|
|
/// Get actions that should be performed by the owner on the strategy's behalf
|
|
#[must_use]
|
|
pub fn actions(&mut self) -> Result<Vec<SyncingAction<B>>, ClientError> {
|
|
// This function presumes that strategies are executed serially and must be refactored once
|
|
// we have parallel strategies.
|
|
let actions: Vec<_> = if let Some(ref mut warp) = self.warp {
|
|
warp.actions().map(Into::into).collect()
|
|
} else if let Some(ref mut state) = self.state {
|
|
state.actions().map(Into::into).collect()
|
|
} else if let Some(ref mut chain_sync) = self.chain_sync {
|
|
chain_sync.actions().map(Into::into).collect()
|
|
} else {
|
|
unreachable!("At least one syncing strategy is always active; qed")
|
|
};
|
|
|
|
if actions.iter().any(SyncingAction::is_finished) {
|
|
self.proceed_to_next()?;
|
|
}
|
|
|
|
Ok(actions)
|
|
}
|
|
|
|
/// Proceed with the next strategy if the active one finished.
|
|
pub fn proceed_to_next(&mut self) -> Result<(), ClientError> {
|
|
// The strategies are switched as `WarpSync` -> `StateStrategy` -> `ChainSync`.
|
|
if let Some(ref mut warp) = self.warp {
|
|
match warp.take_result() {
|
|
Some(res) => {
|
|
info!(
|
|
target: LOG_TARGET,
|
|
"Warp sync is complete, continuing with state sync."
|
|
);
|
|
let state_sync = StateStrategy::new(
|
|
self.client.clone(),
|
|
res.target_header,
|
|
res.target_body,
|
|
res.target_justifications,
|
|
false,
|
|
self.peer_best_blocks
|
|
.iter()
|
|
.map(|(peer_id, (_, best_number))| (*peer_id, *best_number)),
|
|
);
|
|
|
|
self.warp = None;
|
|
self.state = Some(state_sync);
|
|
Ok(())
|
|
},
|
|
None => {
|
|
error!(
|
|
target: LOG_TARGET,
|
|
"Warp sync failed. Continuing with full sync."
|
|
);
|
|
let chain_sync = match ChainSync::new(
|
|
chain_sync_mode(self.config.mode),
|
|
self.client.clone(),
|
|
self.config.max_parallel_downloads,
|
|
self.config.max_blocks_per_request,
|
|
self.config.metrics_registry.clone(),
|
|
self.peer_best_blocks.iter().map(|(peer_id, (best_hash, best_number))| {
|
|
(*peer_id, *best_hash, *best_number)
|
|
}),
|
|
) {
|
|
Ok(chain_sync) => chain_sync,
|
|
Err(e) => {
|
|
error!(target: LOG_TARGET, "Failed to start `ChainSync`.");
|
|
return Err(e)
|
|
},
|
|
};
|
|
|
|
self.warp = None;
|
|
self.chain_sync = Some(chain_sync);
|
|
Ok(())
|
|
},
|
|
}
|
|
} else if let Some(state) = &self.state {
|
|
if state.is_succeeded() {
|
|
info!(target: LOG_TARGET, "State sync is complete, continuing with block sync.");
|
|
} else {
|
|
error!(target: LOG_TARGET, "State sync failed. Falling back to full sync.");
|
|
}
|
|
let chain_sync = match ChainSync::new(
|
|
chain_sync_mode(self.config.mode),
|
|
self.client.clone(),
|
|
self.config.max_parallel_downloads,
|
|
self.config.max_blocks_per_request,
|
|
self.config.metrics_registry.clone(),
|
|
self.peer_best_blocks.iter().map(|(peer_id, (best_hash, best_number))| {
|
|
(*peer_id, *best_hash, *best_number)
|
|
}),
|
|
) {
|
|
Ok(chain_sync) => chain_sync,
|
|
Err(e) => {
|
|
error!(target: LOG_TARGET, "Failed to start `ChainSync`.");
|
|
return Err(e);
|
|
},
|
|
};
|
|
|
|
self.state = None;
|
|
self.chain_sync = Some(chain_sync);
|
|
Ok(())
|
|
} else {
|
|
unreachable!("Only warp & state strategies can finish; qed")
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod test {
|
|
use super::*;
|
|
use futures::executor::block_on;
|
|
use sc_block_builder::BlockBuilderBuilder;
|
|
use substrate_test_runtime_client::{
|
|
ClientBlockImportExt, ClientExt, DefaultTestClientBuilderExt, TestClientBuilder,
|
|
TestClientBuilderExt,
|
|
};
|
|
|
|
/// Regression test for crash when starting already synced parachain node with `--sync=warp`.
|
|
/// We must remove this after setting of warp sync target block is moved to initialization of
|
|
/// `SyncingEngine` (issue https://github.com/paritytech/polkadot-sdk/issues/3537).
|
|
#[test]
|
|
fn set_target_block_finished_warp_sync() {
|
|
// Populate database with finalized state.
|
|
let mut client = Arc::new(TestClientBuilder::new().build());
|
|
let block = BlockBuilderBuilder::new(&*client)
|
|
.on_parent_block(client.chain_info().best_hash)
|
|
.with_parent_block_number(client.chain_info().best_number)
|
|
.build()
|
|
.unwrap()
|
|
.build()
|
|
.unwrap()
|
|
.block;
|
|
block_on(client.import(BlockOrigin::Own, block.clone())).unwrap();
|
|
let just = (*b"TEST", Vec::new());
|
|
client.finalize_block(block.hash(), Some(just)).unwrap();
|
|
let target_block = BlockBuilderBuilder::new(&*client)
|
|
.on_parent_block(client.chain_info().best_hash)
|
|
.with_parent_block_number(client.chain_info().best_number)
|
|
.build()
|
|
.unwrap()
|
|
.build()
|
|
.unwrap()
|
|
.block;
|
|
|
|
// Initialize syncing strategy.
|
|
let config = SyncingConfig {
|
|
mode: SyncMode::Warp,
|
|
max_parallel_downloads: 3,
|
|
max_blocks_per_request: 64,
|
|
metrics_registry: None,
|
|
};
|
|
let mut strategy =
|
|
SyncingStrategy::new(config, client, Some(WarpSyncConfig::WaitForTarget)).unwrap();
|
|
|
|
// Warp sync instantly finishes as we have finalized state in DB.
|
|
let actions = strategy.actions().unwrap();
|
|
assert_eq!(actions.len(), 1);
|
|
assert!(matches!(actions[0], SyncingAction::Finished));
|
|
assert!(strategy.warp.is_none());
|
|
|
|
// Try setting the target block. We mustn't crash.
|
|
strategy
|
|
.set_warp_sync_target_block_header(target_block.header().clone())
|
|
.unwrap();
|
|
}
|
|
}
|