mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-05-06 11:28:02 +00:00
346 lines
10 KiB
Rust
346 lines
10 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) 2017-2020 Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
//! Chain utilities.
|
|
|
|
use crate::error;
|
|
use crate::builder::{ServiceBuilderCommand, ServiceBuilder};
|
|
use crate::error::Error;
|
|
use sc_chain_spec::ChainSpec;
|
|
use log::{warn, info};
|
|
use futures::{future, prelude::*};
|
|
use sp_runtime::traits::{
|
|
Block as BlockT, NumberFor, One, Zero, Header, SaturatedConversion
|
|
};
|
|
use sp_runtime::generic::{BlockId, SignedBlock};
|
|
use codec::{Decode, Encode, IoReader};
|
|
use crate::client::{Client, LocalCallExecutor};
|
|
use sp_consensus::{
|
|
BlockOrigin,
|
|
import_queue::{IncomingBlock, Link, BlockImportError, BlockImportResult, ImportQueue},
|
|
};
|
|
use sc_executor::{NativeExecutor, NativeExecutionDispatch};
|
|
use sp_core::storage::{StorageKey, well_known_keys, ChildInfo, Storage, StorageChild, StorageMap};
|
|
use sc_client_api::{StorageProvider, BlockBackend, UsageProvider};
|
|
|
|
use std::{io::{Read, Write, Seek}, pin::Pin, collections::HashMap};
|
|
|
|
/// Build a chain spec json
|
|
pub fn build_spec(spec: &dyn ChainSpec, raw: bool) -> error::Result<String> {
|
|
spec.as_json(raw).map_err(Into::into)
|
|
}
|
|
|
|
impl<
|
|
TBl, TRtApi, TBackend,
|
|
TExecDisp, TFchr, TSc, TImpQu, TFprb, TFpp,
|
|
TExPool, TRpc, Backend
|
|
> ServiceBuilderCommand for ServiceBuilder<
|
|
TBl, TRtApi,
|
|
Client<TBackend, LocalCallExecutor<TBackend, NativeExecutor<TExecDisp>>, TBl, TRtApi>,
|
|
TFchr, TSc, TImpQu, TFprb, TFpp, TExPool, TRpc, Backend
|
|
> where
|
|
TBl: BlockT,
|
|
TBackend: 'static + sc_client_api::backend::Backend<TBl> + Send,
|
|
TExecDisp: 'static + NativeExecutionDispatch,
|
|
TImpQu: 'static + ImportQueue<TBl>,
|
|
TRtApi: 'static + Send + Sync,
|
|
Self: Send + 'static,
|
|
{
|
|
type Block = TBl;
|
|
type NativeDispatch = TExecDisp;
|
|
|
|
fn import_blocks(
|
|
mut self,
|
|
input: impl Read + Seek + Send + 'static,
|
|
force: bool,
|
|
) -> Pin<Box<dyn Future<Output = Result<(), Error>> + Send>> {
|
|
struct WaitLink {
|
|
imported_blocks: u64,
|
|
has_error: bool,
|
|
}
|
|
|
|
impl WaitLink {
|
|
fn new() -> WaitLink {
|
|
WaitLink {
|
|
imported_blocks: 0,
|
|
has_error: false,
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<B: BlockT> Link<B> for WaitLink {
|
|
fn blocks_processed(
|
|
&mut self,
|
|
imported: usize,
|
|
_count: usize,
|
|
results: Vec<(Result<BlockImportResult<NumberFor<B>>, BlockImportError>, B::Hash)>
|
|
) {
|
|
self.imported_blocks += imported as u64;
|
|
|
|
for result in results {
|
|
if let (Err(err), hash) = result {
|
|
warn!("There was an error importing block with hash {:?}: {:?}", hash, err);
|
|
self.has_error = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
let mut io_reader_input = IoReader(input);
|
|
let mut count = None::<u64>;
|
|
let mut read_block_count = 0;
|
|
let mut link = WaitLink::new();
|
|
|
|
// Importing blocks is implemented as a future, because we want the operation to be
|
|
// interruptible.
|
|
//
|
|
// Every time we read a block from the input or import a bunch of blocks from the import
|
|
// queue, the `Future` re-schedules itself and returns `Poll::Pending`.
|
|
// This makes it possible either to interleave other operations in-between the block imports,
|
|
// or to stop the operation completely.
|
|
let import = future::poll_fn(move |cx| {
|
|
let client = &self.client;
|
|
let queue = &mut self.import_queue;
|
|
|
|
// Start by reading the number of blocks if not done so already.
|
|
let count = match count {
|
|
Some(c) => c,
|
|
None => {
|
|
let c: u64 = match Decode::decode(&mut io_reader_input) {
|
|
Ok(c) => c,
|
|
Err(err) => {
|
|
let err = format!("Error reading file: {}", err);
|
|
return std::task::Poll::Ready(Err(From::from(err)));
|
|
},
|
|
};
|
|
info!("📦 Importing {} blocks", c);
|
|
count = Some(c);
|
|
c
|
|
}
|
|
};
|
|
|
|
// Read blocks from the input.
|
|
if read_block_count < count {
|
|
match SignedBlock::<Self::Block>::decode(&mut io_reader_input) {
|
|
Ok(signed) => {
|
|
let (header, extrinsics) = signed.block.deconstruct();
|
|
let hash = header.hash();
|
|
// import queue handles verification and importing it into the client
|
|
queue.import_blocks(BlockOrigin::File, vec![
|
|
IncomingBlock::<Self::Block> {
|
|
hash,
|
|
header: Some(header),
|
|
body: Some(extrinsics),
|
|
justification: signed.justification,
|
|
origin: None,
|
|
allow_missing_state: false,
|
|
import_existing: force,
|
|
}
|
|
]);
|
|
}
|
|
Err(e) => {
|
|
warn!("Error reading block data at {}: {}", read_block_count, e);
|
|
return std::task::Poll::Ready(Ok(()));
|
|
}
|
|
}
|
|
|
|
read_block_count += 1;
|
|
if read_block_count % 1000 == 0 {
|
|
info!("#{} blocks were added to the queue", read_block_count);
|
|
}
|
|
|
|
cx.waker().wake_by_ref();
|
|
return std::task::Poll::Pending;
|
|
}
|
|
|
|
let blocks_before = link.imported_blocks;
|
|
queue.poll_actions(cx, &mut link);
|
|
|
|
if link.has_error {
|
|
info!(
|
|
"Stopping after #{} blocks because of an error",
|
|
link.imported_blocks,
|
|
);
|
|
return std::task::Poll::Ready(Ok(()));
|
|
}
|
|
|
|
if link.imported_blocks / 1000 != blocks_before / 1000 {
|
|
info!(
|
|
"#{} blocks were imported (#{} left)",
|
|
link.imported_blocks,
|
|
count - link.imported_blocks
|
|
);
|
|
}
|
|
|
|
if link.imported_blocks >= count {
|
|
info!("🎉 Imported {} blocks. Best: #{}", read_block_count, client.chain_info().best_number);
|
|
return std::task::Poll::Ready(Ok(()));
|
|
|
|
} else {
|
|
// Polling the import queue will re-schedule the task when ready.
|
|
return std::task::Poll::Pending;
|
|
}
|
|
});
|
|
Box::pin(import)
|
|
}
|
|
|
|
fn export_blocks(
|
|
self,
|
|
mut output: impl Write + 'static,
|
|
from: NumberFor<TBl>,
|
|
to: Option<NumberFor<TBl>>,
|
|
binary: bool
|
|
) -> Pin<Box<dyn Future<Output = Result<(), Error>>>> {
|
|
let mut block = from;
|
|
|
|
let last = match to {
|
|
Some(v) if v.is_zero() => One::one(),
|
|
Some(v) => v,
|
|
None => self.client.chain_info().best_number,
|
|
};
|
|
|
|
let mut wrote_header = false;
|
|
|
|
// Exporting blocks is implemented as a future, because we want the operation to be
|
|
// interruptible.
|
|
//
|
|
// Every time we write a block to the output, the `Future` re-schedules itself and returns
|
|
// `Poll::Pending`.
|
|
// This makes it possible either to interleave other operations in-between the block exports,
|
|
// or to stop the operation completely.
|
|
let export = future::poll_fn(move |cx| {
|
|
let client = &self.client;
|
|
|
|
if last < block {
|
|
return std::task::Poll::Ready(Err("Invalid block range specified".into()));
|
|
}
|
|
|
|
if !wrote_header {
|
|
info!("Exporting blocks from #{} to #{}", block, last);
|
|
if binary {
|
|
let last_: u64 = last.saturated_into::<u64>();
|
|
let block_: u64 = block.saturated_into::<u64>();
|
|
let len: u64 = last_ - block_ + 1;
|
|
output.write_all(&len.encode())?;
|
|
}
|
|
wrote_header = true;
|
|
}
|
|
|
|
match client.block(&BlockId::number(block))? {
|
|
Some(block) => {
|
|
if binary {
|
|
output.write_all(&block.encode())?;
|
|
} else {
|
|
serde_json::to_writer(&mut output, &block)
|
|
.map_err(|e| format!("Error writing JSON: {}", e))?;
|
|
}
|
|
},
|
|
// Reached end of the chain.
|
|
None => return std::task::Poll::Ready(Ok(())),
|
|
}
|
|
if (block % 10000.into()).is_zero() {
|
|
info!("#{}", block);
|
|
}
|
|
if block == last {
|
|
return std::task::Poll::Ready(Ok(()));
|
|
}
|
|
block += One::one();
|
|
|
|
// Re-schedule the task in order to continue the operation.
|
|
cx.waker().wake_by_ref();
|
|
std::task::Poll::Pending
|
|
});
|
|
|
|
Box::pin(export)
|
|
}
|
|
|
|
fn revert_chain(
|
|
&self,
|
|
blocks: NumberFor<TBl>
|
|
) -> Result<(), Error> {
|
|
let reverted = self.client.revert(blocks)?;
|
|
let info = self.client.chain_info();
|
|
|
|
if reverted.is_zero() {
|
|
info!("There aren't any non-finalized blocks to revert.");
|
|
} else {
|
|
info!("Reverted {} blocks. Best: #{} ({})", reverted, info.best_number, info.best_hash);
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
fn check_block(
|
|
self,
|
|
block_id: BlockId<TBl>
|
|
) -> Pin<Box<dyn Future<Output = Result<(), Error>> + Send>> {
|
|
match self.client.block(&block_id) {
|
|
Ok(Some(block)) => {
|
|
let mut buf = Vec::new();
|
|
1u64.encode_to(&mut buf);
|
|
block.encode_to(&mut buf);
|
|
let reader = std::io::Cursor::new(buf);
|
|
self.import_blocks(reader, true)
|
|
}
|
|
Ok(None) => Box::pin(future::err("Unknown block".into())),
|
|
Err(e) => Box::pin(future::err(format!("Error reading block: {:?}", e).into())),
|
|
}
|
|
}
|
|
|
|
fn export_raw_state(
|
|
&self,
|
|
block: Option<BlockId<Self::Block>>,
|
|
) -> Result<Storage, Error> {
|
|
let block = block.unwrap_or_else(
|
|
|| BlockId::Hash(self.client.usage_info().chain.best_hash)
|
|
);
|
|
|
|
let empty_key = StorageKey(Vec::new());
|
|
let mut top_storage = self.client.storage_pairs(&block, &empty_key)?;
|
|
let mut children_default = HashMap::new();
|
|
|
|
// Remove all default child storage roots from the top storage and collect the child storage
|
|
// pairs.
|
|
while let Some(pos) = top_storage
|
|
.iter()
|
|
.position(|(k, _)| k.0.starts_with(well_known_keys::DEFAULT_CHILD_STORAGE_KEY_PREFIX)) {
|
|
let (key, _) = top_storage.swap_remove(pos);
|
|
|
|
let key = StorageKey(
|
|
key.0[well_known_keys::DEFAULT_CHILD_STORAGE_KEY_PREFIX.len()..].to_vec(),
|
|
);
|
|
let child_info = ChildInfo::new_default(&key.0);
|
|
|
|
let keys = self.client.child_storage_keys(&block, &child_info, &empty_key)?;
|
|
let mut pairs = StorageMap::new();
|
|
keys.into_iter().try_for_each(|k| {
|
|
if let Some(value) = self.client.child_storage(&block, &child_info, &k)? {
|
|
pairs.insert(k.0, value.0);
|
|
}
|
|
|
|
Ok::<_, Error>(())
|
|
})?;
|
|
|
|
children_default.insert(key.0, StorageChild { child_info, data: pairs });
|
|
}
|
|
|
|
let top = top_storage.into_iter().map(|(k, v)| (k.0, v.0)).collect();
|
|
Ok(Storage { top, children_default })
|
|
}
|
|
}
|