mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-27 01:07:57 +00:00
58cc0992df
* add stub for Client.best_chain_containing_block_hash * add fn blockchain::Backend::leaf_hashes * fix typo * sketch out Client.best_chain_containing_block_hash * fix indent * Blockchain.leaf_hashes -> Blockchain.leaves * add unimplemented! stub impls for Blockchain.leaves * start impl of Blockchain.leaves for in-memory client db * Client.best_chain_containing...: check canonical first and make compile * first rough attempt at maintaining leaf list in in-memory db * fix tab indent * add test best_chain_containing_single_block * add failing test best_chain_containing_with_fork * pub use client::blockchain; in test-client to prevent circular dep in client tests * best_chain_containing_with_single_block: improve and test leaves * far improve in-memory Backend::leaves impl * test blockchain::Backend::leaves more thoroughly * handle more edge cases in blockchain::Backend::leaves impl for in memory * fix test best_chain_containing_with_fork (two distinct test blocks had same hash) * make best_chain_containing_block_hash pass existing tests * improve docstring for Blockchain::leaves * Client.best_chain_containing: some cleanup. support max_block_number * best_chain_containing: remove broken outcommented fast check for best = canonical * remove blank line * best_block_containing: return None if target_hash not found * best_chain_containing: add unreachable! at end of function * improve tests for best_chain_containing * renames * more elaborate test scenario for best_containing * best_containing: fix restriction of search through maybe_max_number * best_containing: tests for restriction of search * get rid of unnecessary clones * replace Client::new_in_mem by new_with_backend which is useful for testing backends * add test_client::new_with_backend for testing different backend impls * add test for in_mem::Backend::leaves * remove unused imports * in_mem test_leaves: simplify * flesh out tests for in_mem leaves impl * remove tests for leaves from client which are now covered in implementing module * improve comment * add Client.new_in_mem again * unwrap in test_client::new_with_backend * make test_client::BlockBuilderExt work not just with in-mem backend * make test client ext not just work with in mem backend * add failing Backend.leaves test for client-db * update Cargo.lock * replace KeccakHasher with Blake2Hasher * refactor address grumble https://github.com/paritytech/substrate/pull/740#discussion_r217822862 * refactor using NumberFor address grumble https://github.com/paritytech/substrate/pull/740#discussion_r217823341 * add test that exposes possible problem * update docstring for Client.best_containing * extract test for Backend.leaves for reuse * improve test blockchain_header_and_hash_return_blocks_from_canonical_chain_given_block_numbers * extract test_blockchain_query_by_number_gets_canonical to easily test multiple impls * remove whitespace * remove todo * Client.best_containing: pre-empt search loop when target in canonical * best_containing: prevent race condition by holding import lock * add todo * extract leaf list update code into function * add comment * client-db: use in-memory-kvdb for tests * use BTreeSet to store leaves for in-mem which is faster and simpler * add docstring * add comments and fix formatting * add initial raw version of LeafSet * remove Client::update_leaves which has been superceded by LeafSet * use LeafSet in in-mem backend * keccak -> blake2 * don't reexport codec traits in primitives addresses https://github.com/paritytech/substrate/pull/740#discussion_r219538185 * fix rebase mistake * improve LeafSet and use it in state-db * correct Transfer nonces to fix ApplyExtinsicFailed(Stale) * use given backend in canoncal test * kill dead tree-route code in util * fix warnings * tests for leafset * reorganizations in in_mem backend * fix reorganization canon block logic * DB commit and safe reversion on write error * fix style nits
1325 lines
47 KiB
Rust
1325 lines
47 KiB
Rust
// Copyright 2017 Parity Technologies (UK) Ltd.
|
|
// This file is part of Substrate.
|
|
|
|
// Substrate is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// Substrate is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with Substrate. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
//! Substrate Client
|
|
|
|
use std::sync::Arc;
|
|
use futures::sync::mpsc;
|
|
use parking_lot::{Mutex, RwLock};
|
|
use primitives::AuthorityId;
|
|
use runtime_primitives::{bft::Justification, generic::{BlockId, SignedBlock, Block as RuntimeBlock}};
|
|
use runtime_primitives::traits::{Block as BlockT, Header as HeaderT, Zero, As, NumberFor, CurrentHeight, BlockNumberToHash};
|
|
use runtime_primitives::BuildStorage;
|
|
use primitives::{Blake2Hasher, H256};
|
|
use primitives::storage::{StorageKey, StorageData};
|
|
use primitives::storage::well_known_keys;
|
|
use codec::{Encode, Decode};
|
|
use state_machine::{
|
|
Backend as StateBackend, CodeExecutor,
|
|
ExecutionStrategy, ExecutionManager, prove_read
|
|
};
|
|
|
|
use backend::{self, BlockImportOperation};
|
|
use blockchain::{self, Info as ChainInfo, Backend as ChainBackend, HeaderBackend as ChainHeaderBackend};
|
|
use call_executor::{CallExecutor, LocalCallExecutor};
|
|
use executor::{RuntimeVersion, RuntimeInfo};
|
|
use notifications::{StorageNotifications, StorageEventStream};
|
|
use {cht, error, in_mem, block_builder, bft, genesis};
|
|
|
|
/// Type that implements `futures::Stream` of block import events.
|
|
pub type ImportNotifications<Block> = mpsc::UnboundedReceiver<BlockImportNotification<Block>>;
|
|
|
|
/// A stream of block finality notifications.
|
|
pub type FinalityNotifications<Block> = mpsc::UnboundedReceiver<FinalityNotification<Block>>;
|
|
|
|
/// Substrate Client
|
|
pub struct Client<B, E, Block> where Block: BlockT {
|
|
backend: Arc<B>,
|
|
executor: E,
|
|
storage_notifications: Mutex<StorageNotifications<Block>>,
|
|
import_notification_sinks: Mutex<Vec<mpsc::UnboundedSender<BlockImportNotification<Block>>>>,
|
|
finality_notification_sinks: Mutex<Vec<mpsc::UnboundedSender<FinalityNotification<Block>>>>,
|
|
import_lock: Mutex<()>,
|
|
importing_block: RwLock<Option<Block::Hash>>, // holds the block hash currently being imported. TODO: replace this with block queue
|
|
execution_strategy: ExecutionStrategy,
|
|
}
|
|
|
|
/// A source of blockchain events.
|
|
pub trait BlockchainEvents<Block: BlockT> {
|
|
/// Get block import event stream. Not guaranteed to be fired for every
|
|
/// imported block.
|
|
fn import_notification_stream(&self) -> ImportNotifications<Block>;
|
|
|
|
/// Get a stream of finality notifications. Not guaranteed to be fired for every
|
|
/// finalized block.
|
|
fn finality_notification_stream(&self) -> FinalityNotifications<Block>;
|
|
|
|
/// Get storage changes event stream.
|
|
///
|
|
/// Passing `None` as `filter_keys` subscribes to all storage changes.
|
|
fn storage_changes_notification_stream(&self, filter_keys: Option<&[StorageKey]>) -> error::Result<StorageEventStream<Block::Hash>>;
|
|
}
|
|
|
|
/// Chain head information.
|
|
pub trait ChainHead<Block: BlockT> {
|
|
/// Get best block header.
|
|
fn best_block_header(&self) -> Result<<Block as BlockT>::Header, error::Error>;
|
|
}
|
|
|
|
/// Fetch block body by ID.
|
|
pub trait BlockBody<Block: BlockT> {
|
|
/// Get block body by ID. Returns `None` if the body is not stored.
|
|
fn block_body(&self, id: &BlockId<Block>) -> error::Result<Option<Vec<<Block as BlockT>::Extrinsic>>>;
|
|
}
|
|
|
|
/// Client info
|
|
// TODO: split queue info from chain info and amalgamate into single struct.
|
|
#[derive(Debug)]
|
|
pub struct ClientInfo<Block: BlockT> {
|
|
/// Best block hash.
|
|
pub chain: ChainInfo<Block>,
|
|
/// Best block number in the queue.
|
|
pub best_queued_number: Option<<<Block as BlockT>::Header as HeaderT>::Number>,
|
|
/// Best queued block hash.
|
|
pub best_queued_hash: Option<Block::Hash>,
|
|
}
|
|
|
|
/// Block import result.
|
|
#[derive(Debug)]
|
|
pub enum ImportResult {
|
|
/// Added to the import queue.
|
|
Queued,
|
|
/// Already in the import queue.
|
|
AlreadyQueued,
|
|
/// Already in the blockchain.
|
|
AlreadyInChain,
|
|
/// Block or parent is known to be bad.
|
|
KnownBad,
|
|
/// Block parent is not in the chain.
|
|
UnknownParent,
|
|
}
|
|
|
|
/// Block status.
|
|
#[derive(Debug, PartialEq, Eq)]
|
|
pub enum BlockStatus {
|
|
/// Added to the import queue.
|
|
Queued,
|
|
/// Already in the blockchain.
|
|
InChain,
|
|
/// Block or parent is known to be bad.
|
|
KnownBad,
|
|
/// Not in the queue or the blockchain.
|
|
Unknown,
|
|
}
|
|
|
|
/// Block data origin.
|
|
#[derive(Debug, PartialEq, Eq, Clone, Copy)]
|
|
pub enum BlockOrigin {
|
|
/// Genesis block built into the client.
|
|
Genesis,
|
|
/// Block is part of the initial sync with the network.
|
|
NetworkInitialSync,
|
|
/// Block was broadcasted on the network.
|
|
NetworkBroadcast,
|
|
/// Block that was received from the network and validated in the consensus process.
|
|
ConsensusBroadcast,
|
|
/// Block that was collated by this node.
|
|
Own,
|
|
/// Block was imported from a file.
|
|
File,
|
|
}
|
|
|
|
/// Summary of an imported block
|
|
#[derive(Clone, Debug)]
|
|
pub struct BlockImportNotification<Block: BlockT> {
|
|
/// Imported block header hash.
|
|
pub hash: Block::Hash,
|
|
/// Imported block origin.
|
|
pub origin: BlockOrigin,
|
|
/// Imported block header.
|
|
pub header: Block::Header,
|
|
/// Is this the new best block.
|
|
pub is_new_best: bool,
|
|
}
|
|
|
|
/// Summary of a finalized block.
|
|
#[derive(Clone, Debug)]
|
|
pub struct FinalityNotification<Block: BlockT> {
|
|
/// Imported block header hash.
|
|
pub hash: Block::Hash,
|
|
/// Imported block header.
|
|
pub header: Block::Header,
|
|
}
|
|
|
|
/// A header paired with a justification which has already been checked.
|
|
#[derive(Debug, PartialEq, Eq, Clone)]
|
|
pub struct JustifiedHeader<Block: BlockT> {
|
|
header: <Block as BlockT>::Header,
|
|
justification: ::bft::Justification<Block::Hash>,
|
|
authorities: Vec<AuthorityId>,
|
|
}
|
|
|
|
impl<Block: BlockT> JustifiedHeader<Block> {
|
|
/// Deconstruct the justified header into parts.
|
|
pub fn into_inner(self) -> (<Block as BlockT>::Header, ::bft::Justification<Block::Hash>, Vec<AuthorityId>) {
|
|
(self.header, self.justification, self.authorities)
|
|
}
|
|
}
|
|
|
|
/// Create an instance of in-memory client.
|
|
pub fn new_in_mem<E, Block, S>(
|
|
executor: E,
|
|
genesis_storage: S,
|
|
) -> error::Result<Client<in_mem::Backend<Block, Blake2Hasher>, LocalCallExecutor<in_mem::Backend<Block, Blake2Hasher>, E>, Block>>
|
|
where
|
|
E: CodeExecutor<Blake2Hasher> + RuntimeInfo,
|
|
S: BuildStorage,
|
|
Block: BlockT,
|
|
H256: From<Block::Hash>,
|
|
{
|
|
new_with_backend(Arc::new(in_mem::Backend::new()), executor, genesis_storage)
|
|
}
|
|
|
|
/// Create a client with the explicitely provided backend.
|
|
/// This is useful for testing backend implementations.
|
|
pub fn new_with_backend<B, E, Block, S>(
|
|
backend: Arc<B>,
|
|
executor: E,
|
|
build_genesis_storage: S,
|
|
) -> error::Result<Client<B, LocalCallExecutor<B, E>, Block>>
|
|
where
|
|
E: CodeExecutor<Blake2Hasher> + RuntimeInfo,
|
|
S: BuildStorage,
|
|
Block: BlockT,
|
|
H256: From<Block::Hash>,
|
|
B: backend::LocalBackend<Block, Blake2Hasher>
|
|
{
|
|
let call_executor = LocalCallExecutor::new(backend.clone(), executor);
|
|
Client::new(backend, call_executor, build_genesis_storage, ExecutionStrategy::NativeWhenPossible)
|
|
}
|
|
|
|
impl<B, E, Block> Client<B, E, Block> where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher>,
|
|
Block: BlockT,
|
|
{
|
|
/// Creates new Substrate Client with given blockchain and code executor.
|
|
pub fn new<S: BuildStorage>(
|
|
backend: Arc<B>,
|
|
executor: E,
|
|
build_genesis_storage: S,
|
|
execution_strategy: ExecutionStrategy,
|
|
) -> error::Result<Self> {
|
|
if backend.blockchain().header(BlockId::Number(Zero::zero()))?.is_none() {
|
|
let genesis_storage = build_genesis_storage.build_storage()?;
|
|
let genesis_block = genesis::construct_genesis_block::<Block>(&genesis_storage);
|
|
info!("Initialising Genesis block/state (state: {}, header-hash: {})", genesis_block.header().state_root(), genesis_block.header().hash());
|
|
let mut op = backend.begin_operation(BlockId::Hash(Default::default()))?;
|
|
op.reset_storage(genesis_storage.into_iter())?;
|
|
op.set_block_data(
|
|
genesis_block.deconstruct().0,
|
|
Some(vec![]),
|
|
None,
|
|
::backend::NewBlockState::Final
|
|
)?;
|
|
backend.commit_operation(op)?;
|
|
}
|
|
Ok(Client {
|
|
backend,
|
|
executor,
|
|
storage_notifications: Default::default(),
|
|
import_notification_sinks: Default::default(),
|
|
finality_notification_sinks: Default::default(),
|
|
import_lock: Default::default(),
|
|
importing_block: Default::default(),
|
|
execution_strategy,
|
|
})
|
|
}
|
|
|
|
/// Get a reference to the state at a given block.
|
|
pub fn state_at(&self, block: &BlockId<Block>) -> error::Result<B::State> {
|
|
self.backend.state_at(*block)
|
|
}
|
|
|
|
/// Expose backend reference. To be used in tests only
|
|
pub fn backend(&self) -> &Arc<B> {
|
|
&self.backend
|
|
}
|
|
|
|
/// Return single storage entry of contract under given address in state in a block of given hash.
|
|
pub fn storage(&self, id: &BlockId<Block>, key: &StorageKey) -> error::Result<Option<StorageData>> {
|
|
Ok(self.state_at(id)?
|
|
.storage(&key.0).map_err(|e| error::Error::from_state(Box::new(e)))?
|
|
.map(StorageData))
|
|
}
|
|
|
|
/// Get the code at a given block.
|
|
pub fn code_at(&self, id: &BlockId<Block>) -> error::Result<Vec<u8>> {
|
|
Ok(self.storage(id, &StorageKey(well_known_keys::CODE.to_vec()))?
|
|
.expect("None is returned if there's no value stored for the given key; ':code' key is always defined; qed").0)
|
|
}
|
|
|
|
/// Get the set of authorities at a given block.
|
|
pub fn authorities_at(&self, id: &BlockId<Block>) -> error::Result<Vec<AuthorityId>> {
|
|
match self.backend.blockchain().cache().and_then(|cache| cache.authorities_at(*id)) {
|
|
Some(cached_value) => Ok(cached_value),
|
|
None => self.executor.call(id, "authorities",&[])
|
|
.and_then(|r| Vec::<AuthorityId>::decode(&mut &r.return_data[..])
|
|
.ok_or(error::ErrorKind::InvalidAuthoritiesSet.into()))
|
|
}
|
|
}
|
|
|
|
/// Get the RuntimeVersion at a given block.
|
|
pub fn runtime_version_at(&self, id: &BlockId<Block>) -> error::Result<RuntimeVersion> {
|
|
// TODO: Post Poc-2 return an error if version is missing
|
|
self.executor.runtime_version(id)
|
|
}
|
|
|
|
/// Get call executor reference.
|
|
pub fn executor(&self) -> &E {
|
|
&self.executor
|
|
}
|
|
|
|
/// Returns the runtime metadata.
|
|
pub fn metadata(&self, id: &BlockId<Block>) -> error::Result<Vec<u8>> {
|
|
self.executor.call(id, "metadata",&[]).map(|v| v.return_data)
|
|
}
|
|
|
|
/// Reads storage value at a given block + key, returning read proof.
|
|
pub fn read_proof(&self, id: &BlockId<Block>, key: &[u8]) -> error::Result<Vec<Vec<u8>>> {
|
|
self.state_at(id)
|
|
.and_then(|state| prove_read(state, key)
|
|
.map(|(_, proof)| proof)
|
|
.map_err(Into::into))
|
|
}
|
|
|
|
/// Execute a call to a contract on top of state in a block of given hash
|
|
/// AND returning execution proof.
|
|
///
|
|
/// No changes are made.
|
|
pub fn execution_proof(&self, id: &BlockId<Block>, method: &str, call_data: &[u8]) -> error::Result<(Vec<u8>, Vec<Vec<u8>>)> {
|
|
self.state_at(id).and_then(|state| self.executor.prove_at_state(state, &mut Default::default(), method, call_data))
|
|
}
|
|
|
|
/// Reads given header and generates CHT-based header proof.
|
|
pub fn header_proof(&self, id: &BlockId<Block>) -> error::Result<(Block::Header, Vec<Vec<u8>>)> {
|
|
self.header_proof_with_cht_size(id, cht::SIZE)
|
|
}
|
|
|
|
/// Reads given header and generates CHT-based header proof for CHT of given size.
|
|
pub fn header_proof_with_cht_size(&self, id: &BlockId<Block>, cht_size: u64) -> error::Result<(Block::Header, Vec<Vec<u8>>)> {
|
|
let proof_error = || error::ErrorKind::Backend(format!("Failed to generate header proof for {:?}", id));
|
|
let header = self.header(id)?.ok_or_else(|| error::ErrorKind::UnknownBlock(format!("{:?}", id)))?;
|
|
let block_num = *header.number();
|
|
let cht_num = cht::block_to_cht_number(cht_size, block_num).ok_or_else(proof_error)?;
|
|
let cht_start = cht::start_number(cht_size, cht_num);
|
|
let headers = (cht_start.as_()..).map(|num| self.block_hash(As::sa(num)).unwrap_or_default());
|
|
let proof = cht::build_proof::<Block::Header, Blake2Hasher, _>(cht_size, cht_num, block_num, headers)
|
|
.ok_or_else(proof_error)?;
|
|
Ok((header, proof))
|
|
}
|
|
|
|
/// Create a new block, built on the head of the chain.
|
|
pub fn new_block(&self) -> error::Result<block_builder::BlockBuilder<B, E, Block, Blake2Hasher>>
|
|
where E: Clone
|
|
{
|
|
block_builder::BlockBuilder::new(self)
|
|
}
|
|
|
|
/// Create a new block, built on top of `parent`.
|
|
pub fn new_block_at(&self, parent: &BlockId<Block>) -> error::Result<block_builder::BlockBuilder<B, E, Block, Blake2Hasher>>
|
|
where E: Clone
|
|
{
|
|
block_builder::BlockBuilder::at_block(parent, &self)
|
|
}
|
|
|
|
/// Set up the native execution environment to call into a native runtime code.
|
|
pub fn call_api<A, R>(&self, function: &'static str, args: &A) -> error::Result<R>
|
|
where A: Encode, R: Decode
|
|
{
|
|
self.call_api_at(&BlockId::Number(self.info()?.chain.best_number), function, args)
|
|
}
|
|
|
|
/// Call a runtime function at given block.
|
|
pub fn call_api_at<A, R>(&self, at: &BlockId<Block>, function: &'static str, args: &A) -> error::Result<R>
|
|
where A: Encode, R: Decode
|
|
{
|
|
let parent = at;
|
|
let header = <<Block as BlockT>::Header as HeaderT>::new(
|
|
self.block_number_from_id(&parent)?
|
|
.ok_or_else(|| error::ErrorKind::UnknownBlock(format!("{:?}", parent)))? + As::sa(1),
|
|
Default::default(),
|
|
Default::default(),
|
|
self.block_hash_from_id(&parent)?
|
|
.ok_or_else(|| error::ErrorKind::UnknownBlock(format!("{:?}", parent)))?,
|
|
Default::default()
|
|
);
|
|
self.state_at(&parent).and_then(|state| {
|
|
let mut overlay = Default::default();
|
|
let execution_manager = || ExecutionManager::Both(|wasm_result, native_result| {
|
|
warn!("Consensus error between wasm and native runtime execution at block {:?}", at);
|
|
warn!(" Function {:?}", function);
|
|
warn!(" Native result {:?}", native_result);
|
|
warn!(" Wasm result {:?}", wasm_result);
|
|
wasm_result
|
|
});
|
|
self.executor().call_at_state(
|
|
&state,
|
|
&mut overlay,
|
|
"initialise_block",
|
|
&header.encode(),
|
|
execution_manager()
|
|
)?;
|
|
let (r, _, _) = args.using_encoded(|input|
|
|
self.executor().call_at_state(
|
|
&state,
|
|
&mut overlay,
|
|
function,
|
|
input,
|
|
execution_manager()
|
|
))?;
|
|
Ok(R::decode(&mut &r[..])
|
|
.ok_or_else(|| error::Error::from(error::ErrorKind::CallResultDecode(function)))?)
|
|
})
|
|
}
|
|
|
|
/// Check a header's justification.
|
|
pub fn check_justification(
|
|
&self,
|
|
header: <Block as BlockT>::Header,
|
|
justification: ::bft::UncheckedJustification<Block::Hash>,
|
|
) -> error::Result<JustifiedHeader<Block>> {
|
|
let parent_hash = header.parent_hash().clone();
|
|
let authorities = self.authorities_at(&BlockId::Hash(parent_hash))?;
|
|
let just = ::bft::check_justification::<Block>(&authorities[..], parent_hash, justification)
|
|
.map_err(|_|
|
|
error::ErrorKind::BadJustification(
|
|
format!("{}", header.hash())
|
|
)
|
|
)?;
|
|
Ok(JustifiedHeader {
|
|
header,
|
|
justification: just,
|
|
authorities,
|
|
})
|
|
}
|
|
|
|
/// Queue a block for import.
|
|
pub fn import_block(
|
|
&self,
|
|
origin: BlockOrigin,
|
|
header: JustifiedHeader<Block>,
|
|
body: Option<Vec<<Block as BlockT>::Extrinsic>>,
|
|
finalized: bool,
|
|
) -> error::Result<ImportResult> {
|
|
let (header, justification, authorities) = header.into_inner();
|
|
let parent_hash = header.parent_hash().clone();
|
|
match self.backend.blockchain().status(BlockId::Hash(parent_hash))? {
|
|
blockchain::BlockStatus::InChain => {},
|
|
blockchain::BlockStatus::Unknown => return Ok(ImportResult::UnknownParent),
|
|
}
|
|
let hash = header.hash();
|
|
let _import_lock = self.import_lock.lock();
|
|
let height: u64 = header.number().as_();
|
|
*self.importing_block.write() = Some(hash);
|
|
|
|
let result = self.execute_and_import_block(
|
|
origin,
|
|
hash,
|
|
header,
|
|
justification,
|
|
body,
|
|
authorities,
|
|
finalized
|
|
);
|
|
|
|
*self.importing_block.write() = None;
|
|
telemetry!("block.import";
|
|
"height" => height,
|
|
"best" => ?hash,
|
|
"origin" => ?origin
|
|
);
|
|
result
|
|
}
|
|
|
|
fn execute_and_import_block(
|
|
&self,
|
|
origin: BlockOrigin,
|
|
hash: Block::Hash,
|
|
header: Block::Header,
|
|
justification: bft::Justification<Block::Hash>,
|
|
body: Option<Vec<Block::Extrinsic>>,
|
|
authorities: Vec<AuthorityId>,
|
|
finalized: bool,
|
|
) -> error::Result<ImportResult> {
|
|
let parent_hash = header.parent_hash().clone();
|
|
match self.backend.blockchain().status(BlockId::Hash(hash))? {
|
|
blockchain::BlockStatus::InChain => return Ok(ImportResult::AlreadyInChain),
|
|
blockchain::BlockStatus::Unknown => {},
|
|
}
|
|
|
|
let (last_best, last_best_number) = {
|
|
let info = self.backend.blockchain().info()?;
|
|
(info.best_hash, info.best_number)
|
|
};
|
|
|
|
// this is a fairly arbitrary choice of where to draw the line on making notifications,
|
|
// but the general goal is to only make notifications when we are already fully synced
|
|
// and get a new chain head.
|
|
let make_notifications = match origin {
|
|
BlockOrigin::NetworkBroadcast | BlockOrigin::Own | BlockOrigin::ConsensusBroadcast => true,
|
|
BlockOrigin::Genesis | BlockOrigin::NetworkInitialSync | BlockOrigin::File => false,
|
|
};
|
|
|
|
// ensure parent block is finalized to maintain invariant that
|
|
// finality is called sequentially.
|
|
if finalized {
|
|
self.apply_finality(parent_hash, last_best, make_notifications)?;
|
|
}
|
|
|
|
let mut transaction = self.backend.begin_operation(BlockId::Hash(parent_hash))?;
|
|
let (storage_update, changes_update, storage_changes) = match transaction.state()? {
|
|
Some(transaction_state) => {
|
|
let mut overlay = Default::default();
|
|
let mut r = self.executor.call_at_state(
|
|
transaction_state,
|
|
&mut overlay,
|
|
"execute_block",
|
|
&<Block as BlockT>::new(header.clone(), body.clone().unwrap_or_default()).encode(),
|
|
match (origin, self.execution_strategy) {
|
|
(BlockOrigin::NetworkInitialSync, _) | (_, ExecutionStrategy::NativeWhenPossible) =>
|
|
ExecutionManager::NativeWhenPossible,
|
|
(_, ExecutionStrategy::AlwaysWasm) => ExecutionManager::AlwaysWasm,
|
|
_ => ExecutionManager::Both(|wasm_result, native_result| {
|
|
warn!("Consensus error between wasm and native block execution at block {}", hash);
|
|
warn!(" Header {:?}", header);
|
|
warn!(" Native result {:?}", native_result);
|
|
warn!(" Wasm result {:?}", wasm_result);
|
|
telemetry!("block.execute.consensus_failure";
|
|
"hash" => ?hash,
|
|
"origin" => ?origin,
|
|
"header" => ?header
|
|
);
|
|
wasm_result
|
|
}),
|
|
},
|
|
);
|
|
let (_, storage_update, changes_update) = r?;
|
|
overlay.commit_prospective();
|
|
(Some(storage_update), Some(changes_update), Some(overlay.into_committed()))
|
|
},
|
|
None => (None, None, None)
|
|
};
|
|
|
|
// TODO: non longest-chain rule.
|
|
let is_new_best = finalized || header.number() > &last_best_number;
|
|
let leaf_state = if finalized {
|
|
::backend::NewBlockState::Final
|
|
} else if is_new_best {
|
|
::backend::NewBlockState::Best
|
|
} else {
|
|
::backend::NewBlockState::Normal
|
|
};
|
|
|
|
trace!("Imported {}, (#{}), best={}, origin={:?}", hash, header.number(), is_new_best, origin);
|
|
let unchecked: bft::UncheckedJustification<_> = justification.uncheck().into();
|
|
|
|
transaction.set_block_data(
|
|
header.clone(),
|
|
body,
|
|
Some(unchecked.into()),
|
|
leaf_state,
|
|
)?;
|
|
|
|
transaction.update_authorities(authorities);
|
|
if let Some(storage_update) = storage_update {
|
|
transaction.update_storage(storage_update)?;
|
|
}
|
|
if let Some(Some(changes_update)) = changes_update {
|
|
transaction.update_changes_trie(changes_update)?;
|
|
}
|
|
self.backend.commit_operation(transaction)?;
|
|
|
|
if make_notifications {
|
|
if let Some(storage_changes) = storage_changes {
|
|
// TODO [ToDr] How to handle re-orgs? Should we re-emit all storage changes?
|
|
self.storage_notifications.lock()
|
|
.trigger(&hash, storage_changes);
|
|
}
|
|
|
|
if finalized {
|
|
let notification = FinalityNotification::<Block> {
|
|
hash,
|
|
header: header.clone(),
|
|
};
|
|
|
|
self.finality_notification_sinks.lock()
|
|
.retain(|sink| sink.unbounded_send(notification.clone()).is_ok());
|
|
}
|
|
|
|
let notification = BlockImportNotification::<Block> {
|
|
hash,
|
|
origin,
|
|
header,
|
|
is_new_best,
|
|
};
|
|
|
|
self.import_notification_sinks.lock()
|
|
.retain(|sink| sink.unbounded_send(notification.clone()).is_ok());
|
|
}
|
|
|
|
Ok(ImportResult::Queued)
|
|
}
|
|
|
|
/// Finalizes all blocks up to given.
|
|
fn apply_finality(&self, block: Block::Hash, best_block: Block::Hash, notify: bool) -> error::Result<()> {
|
|
// find tree route from last finalized to given block.
|
|
let last_finalized = self.backend.blockchain().last_finalized()?;
|
|
|
|
if block == last_finalized { return Ok(()) }
|
|
let route_from_finalized = ::blockchain::tree_route(
|
|
self.backend.blockchain(),
|
|
BlockId::Hash(last_finalized),
|
|
BlockId::Hash(block),
|
|
)?;
|
|
|
|
if let Some(retracted) = route_from_finalized.retracted().get(0) {
|
|
warn!("Safety violation: attempted to revert finalized block {:?} which is not in the \
|
|
same chain as last finalized {:?}", retracted, last_finalized);
|
|
|
|
bail!(error::ErrorKind::NotInFinalizedChain);
|
|
}
|
|
|
|
let route_from_best = ::blockchain::tree_route(
|
|
self.backend.blockchain(),
|
|
BlockId::Hash(best_block),
|
|
BlockId::Hash(block),
|
|
)?;
|
|
|
|
// if the block is not a direct ancestor of the current best chain,
|
|
// then some other block is the common ancestor.
|
|
if route_from_best.common_block().hash != block {
|
|
// TODO: reorganize best block to be the best chain containing
|
|
// `block`.
|
|
}
|
|
|
|
for finalize_new in route_from_finalized.enacted() {
|
|
self.backend.finalize_block(BlockId::Hash(finalize_new.hash))?;
|
|
}
|
|
|
|
if notify {
|
|
// sometimes when syncing, tons of blocks can be finalized at once.
|
|
// we'll send notifications spuriously in that case.
|
|
const MAX_TO_NOTIFY: usize = 256;
|
|
let enacted = route_from_finalized.enacted();
|
|
let start = enacted.len() - ::std::cmp::min(enacted.len(), MAX_TO_NOTIFY);
|
|
let mut sinks = self.finality_notification_sinks.lock();
|
|
for finalized in &enacted[start..] {
|
|
let header = self.header(&BlockId::Hash(finalized.hash))?
|
|
.expect("header already known to exist in DB because it is indicated in the tree route; qed");
|
|
let notification = FinalityNotification {
|
|
header,
|
|
hash: finalized.hash,
|
|
};
|
|
|
|
sinks.retain(|sink| sink.unbounded_send(notification.clone()).is_ok());
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Finalize a block. This will implicitly finalize all blocks up to it and
|
|
/// fire finality notifications.
|
|
///
|
|
/// Pass a flag to indicate whether finality notifications should be propagated.
|
|
/// This is usually tied to some synchronization state, where we don't send notifications
|
|
/// while performing major synchronization work.
|
|
pub fn finalize_block(&self, id: BlockId<Block>, notify: bool) -> error::Result<()> {
|
|
let last_best = self.backend.blockchain().info()?.best_hash;
|
|
let to_finalize_hash = match id {
|
|
BlockId::Hash(h) => h,
|
|
BlockId::Number(n) => self.backend.blockchain().hash(n)?
|
|
.ok_or_else(|| error::ErrorKind::UnknownBlock(format!("No block with number {:?}", n)))?,
|
|
};
|
|
|
|
self.apply_finality(to_finalize_hash, last_best, notify)
|
|
}
|
|
|
|
/// Attempts to revert the chain by `n` blocks. Returns the number of blocks that were
|
|
/// successfully reverted.
|
|
pub fn revert(&self, n: NumberFor<Block>) -> error::Result<NumberFor<Block>> {
|
|
Ok(self.backend.revert(n)?)
|
|
}
|
|
|
|
/// Get blockchain info.
|
|
pub fn info(&self) -> error::Result<ClientInfo<Block>> {
|
|
let info = self.backend.blockchain().info().map_err(|e| error::Error::from_blockchain(Box::new(e)))?;
|
|
Ok(ClientInfo {
|
|
chain: info,
|
|
best_queued_hash: None,
|
|
best_queued_number: None,
|
|
})
|
|
}
|
|
|
|
/// Get block status.
|
|
pub fn block_status(&self, id: &BlockId<Block>) -> error::Result<BlockStatus> {
|
|
// TODO: more efficient implementation
|
|
if let BlockId::Hash(ref h) = id {
|
|
if self.importing_block.read().as_ref().map_or(false, |importing| h == importing) {
|
|
return Ok(BlockStatus::Queued);
|
|
}
|
|
}
|
|
match self.backend.blockchain().header(*id).map_err(|e| error::Error::from_blockchain(Box::new(e)))?.is_some() {
|
|
true => Ok(BlockStatus::InChain),
|
|
false => Ok(BlockStatus::Unknown),
|
|
}
|
|
}
|
|
|
|
/// Get block hash by number.
|
|
pub fn block_hash(&self, block_number: <<Block as BlockT>::Header as HeaderT>::Number) -> error::Result<Option<Block::Hash>> {
|
|
self.backend.blockchain().hash(block_number)
|
|
}
|
|
|
|
/// Convert an arbitrary block ID into a block hash.
|
|
pub fn block_hash_from_id(&self, id: &BlockId<Block>) -> error::Result<Option<Block::Hash>> {
|
|
match *id {
|
|
BlockId::Hash(h) => Ok(Some(h)),
|
|
BlockId::Number(n) => self.block_hash(n),
|
|
}
|
|
}
|
|
|
|
/// Convert an arbitrary block ID into a block hash.
|
|
pub fn block_number_from_id(&self, id: &BlockId<Block>) -> error::Result<Option<<<Block as BlockT>::Header as HeaderT>::Number>> {
|
|
match *id {
|
|
BlockId::Hash(_) => Ok(self.header(id)?.map(|h| h.number().clone())),
|
|
BlockId::Number(n) => Ok(Some(n)),
|
|
}
|
|
}
|
|
|
|
/// Get block header by id.
|
|
pub fn header(&self, id: &BlockId<Block>) -> error::Result<Option<<Block as BlockT>::Header>> {
|
|
self.backend.blockchain().header(*id)
|
|
}
|
|
|
|
/// Get block body by id.
|
|
pub fn body(&self, id: &BlockId<Block>) -> error::Result<Option<Vec<<Block as BlockT>::Extrinsic>>> {
|
|
self.backend.blockchain().body(*id)
|
|
}
|
|
|
|
/// Get block justification set by id.
|
|
pub fn justification(&self, id: &BlockId<Block>) -> error::Result<Option<Justification<Block::Hash>>> {
|
|
self.backend.blockchain().justification(*id)
|
|
}
|
|
|
|
/// Get full block by id.
|
|
pub fn block(&self, id: &BlockId<Block>) -> error::Result<Option<SignedBlock<Block::Header, Block::Extrinsic, Block::Hash>>> {
|
|
Ok(match (self.header(id)?, self.body(id)?, self.justification(id)?) {
|
|
(Some(header), Some(extrinsics), Some(justification)) =>
|
|
Some(SignedBlock { block: RuntimeBlock { header, extrinsics }, justification }),
|
|
_ => None,
|
|
})
|
|
}
|
|
|
|
/// Get best block header.
|
|
pub fn best_block_header(&self) -> error::Result<<Block as BlockT>::Header> {
|
|
let info = self.backend.blockchain().info().map_err(|e| error::Error::from_blockchain(Box::new(e)))?;
|
|
Ok(self.header(&BlockId::Hash(info.best_hash))?.expect("Best block header must always exist"))
|
|
}
|
|
|
|
/// Get the most recent block hash of the best (longest) chains
|
|
/// that contain block with the given `target_hash`.
|
|
/// If `maybe_max_block_number` is `Some(max_block_number)`
|
|
/// the search is limited to block `numbers <= max_block_number`.
|
|
/// in other words as if there were no blocks greater `max_block_number`.
|
|
/// TODO [snd] possibly implement this on blockchain::Backend and just redirect here
|
|
/// Returns `Ok(None)` if `target_hash` is not found in search space.
|
|
/// TODO [snd] write down time complexity
|
|
pub fn best_containing(&self, target_hash: Block::Hash, maybe_max_number: Option<NumberFor<Block>>) -> error::Result<Option<Block::Hash>> {
|
|
let target_header = {
|
|
match self.backend.blockchain().header(BlockId::Hash(target_hash))? {
|
|
Some(x) => x,
|
|
// target not in blockchain
|
|
None => { return Ok(None); },
|
|
}
|
|
};
|
|
|
|
if let Some(max_number) = maybe_max_number {
|
|
// target outside search range
|
|
if target_header.number() > &max_number {
|
|
return Ok(None);
|
|
}
|
|
}
|
|
|
|
let (leaves, best_already_checked) = {
|
|
// ensure no blocks are imported during this code block.
|
|
// an import could trigger a reorg which could change the canonical chain.
|
|
// we depend on the canonical chain staying the same during this code block.
|
|
let _import_lock = self.import_lock.lock();
|
|
|
|
let info = self.backend.blockchain().info()?;
|
|
|
|
let canon_hash = self.backend.blockchain().hash(*target_header.number())?
|
|
.ok_or_else(|| error::Error::from(format!("failed to get hash for block number {}", target_header.number())))?;
|
|
|
|
if canon_hash == target_hash {
|
|
if let Some(max_number) = maybe_max_number {
|
|
// something has to guarantee that max_number is in chain
|
|
return Ok(Some(self.backend.blockchain().hash(max_number)?.ok_or_else(|| error::Error::from(format!("failed to get hash for block number {}", max_number)))?));
|
|
} else {
|
|
return Ok(Some(info.best_hash));
|
|
}
|
|
}
|
|
(self.backend.blockchain().leaves()?, info.best_hash)
|
|
};
|
|
|
|
// for each chain. longest chain first. shortest last
|
|
for leaf_hash in leaves {
|
|
// ignore canonical chain which we already checked above
|
|
if leaf_hash == best_already_checked {
|
|
continue;
|
|
}
|
|
|
|
// start at the leaf
|
|
let mut current_hash = leaf_hash;
|
|
|
|
// if search is not restricted then the leaf is the best
|
|
let mut best_hash = leaf_hash;
|
|
|
|
// go backwards entering the search space
|
|
// waiting until we are <= max_number
|
|
if let Some(max_number) = maybe_max_number {
|
|
loop {
|
|
// TODO [snd] this should be a panic
|
|
let current_header = self.backend.blockchain().header(BlockId::Hash(current_hash.clone()))?
|
|
.ok_or_else(|| error::Error::from(format!("failed to get header for hash {}", current_hash)))?;
|
|
|
|
if current_header.number() <= &max_number {
|
|
best_hash = current_header.hash();
|
|
break;
|
|
}
|
|
|
|
current_hash = *current_header.parent_hash();
|
|
}
|
|
}
|
|
|
|
// go backwards through the chain (via parent links)
|
|
loop {
|
|
// until we find target
|
|
if current_hash == target_hash {
|
|
return Ok(Some(best_hash));
|
|
}
|
|
|
|
// TODO [snd] this should be a panic
|
|
let current_header = self.backend.blockchain().header(BlockId::Hash(current_hash.clone()))?
|
|
.ok_or_else(|| error::Error::from(format!("failed to get header for hash {}", current_hash)))?;
|
|
|
|
// stop search in this chain once we go below the target's block number
|
|
if current_header.number() < target_header.number() {
|
|
break;
|
|
}
|
|
|
|
current_hash = *current_header.parent_hash();
|
|
}
|
|
}
|
|
|
|
unreachable!("this is a bug. `target_hash` is in blockchain but wasn't found following all leaves backwards");
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> CurrentHeight for Client<B, E, Block> where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher> + Clone,
|
|
Block: BlockT,
|
|
{
|
|
type BlockNumber = <Block::Header as HeaderT>::Number;
|
|
fn current_height(&self) -> Self::BlockNumber {
|
|
self.backend.blockchain().info().map(|i| i.best_number).unwrap_or_else(|_| Zero::zero())
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> BlockNumberToHash for Client<B, E, Block> where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher> + Clone,
|
|
Block: BlockT,
|
|
{
|
|
type BlockNumber = <Block::Header as HeaderT>::Number;
|
|
type Hash = Block::Hash;
|
|
fn block_number_to_hash(&self, n: Self::BlockNumber) -> Option<Self::Hash> {
|
|
self.block_hash(n).unwrap_or(None)
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> bft::BlockImport<Block> for Client<B, E, Block>
|
|
where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher>,
|
|
Block: BlockT,
|
|
{
|
|
fn import_block(
|
|
&self,
|
|
block: Block,
|
|
justification: ::bft::Justification<Block::Hash>,
|
|
authorities: &[AuthorityId],
|
|
) -> bool {
|
|
let (header, extrinsics) = block.deconstruct();
|
|
let justified_header = JustifiedHeader {
|
|
header: header,
|
|
justification,
|
|
authorities: authorities.to_vec(),
|
|
};
|
|
|
|
// TODO [rob]: non-instant finality.
|
|
self.import_block(
|
|
BlockOrigin::ConsensusBroadcast,
|
|
justified_header,
|
|
Some(extrinsics),
|
|
true
|
|
).is_ok()
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> bft::Authorities<Block> for Client<B, E, Block>
|
|
where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher>,
|
|
Block: BlockT,
|
|
{
|
|
fn authorities(&self, at: &BlockId<Block>) -> Result<Vec<AuthorityId>, bft::Error> {
|
|
let on_chain_version: Result<_, bft::Error> = self.runtime_version_at(at)
|
|
.map_err(|e| { trace!("Error getting runtime version {:?}", e); bft::ErrorKind::RuntimeVersionMissing.into() });
|
|
let on_chain_version = on_chain_version?;
|
|
let native_version: Result<_, bft::Error> = self.executor.native_runtime_version()
|
|
.ok_or_else(|| bft::ErrorKind::NativeRuntimeMissing.into());
|
|
let native_version = native_version?;
|
|
if !on_chain_version.can_author_with(&native_version) {
|
|
return Err(bft::ErrorKind::IncompatibleAuthoringRuntime(on_chain_version, native_version).into())
|
|
}
|
|
self.authorities_at(at).map_err(|_| {
|
|
let descriptor = format!("{:?}", at);
|
|
bft::ErrorKind::StateUnavailable(descriptor).into()
|
|
})
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> BlockchainEvents<Block> for Client<B, E, Block>
|
|
where
|
|
E: CallExecutor<Block, Blake2Hasher>,
|
|
Block: BlockT,
|
|
{
|
|
/// Get block import event stream.
|
|
fn import_notification_stream(&self) -> ImportNotifications<Block> {
|
|
let (sink, stream) = mpsc::unbounded();
|
|
self.import_notification_sinks.lock().push(sink);
|
|
stream
|
|
}
|
|
|
|
fn finality_notification_stream(&self) -> FinalityNotifications<Block> {
|
|
let (sink, stream) = mpsc::unbounded();
|
|
self.finality_notification_sinks.lock().push(sink);
|
|
stream
|
|
}
|
|
|
|
/// Get storage changes event stream.
|
|
fn storage_changes_notification_stream(&self, filter_keys: Option<&[StorageKey]>) -> error::Result<StorageEventStream<Block::Hash>> {
|
|
Ok(self.storage_notifications.lock().listen(filter_keys))
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> ChainHead<Block> for Client<B, E, Block>
|
|
where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher>,
|
|
Block: BlockT,
|
|
{
|
|
fn best_block_header(&self) -> error::Result<<Block as BlockT>::Header> {
|
|
Client::best_block_header(self)
|
|
}
|
|
}
|
|
|
|
impl<B, E, Block> BlockBody<Block> for Client<B, E, Block>
|
|
where
|
|
B: backend::Backend<Block, Blake2Hasher>,
|
|
E: CallExecutor<Block, Blake2Hasher>,
|
|
Block: BlockT,
|
|
{
|
|
fn block_body(&self, id: &BlockId<Block>) -> error::Result<Option<Vec<<Block as BlockT>::Extrinsic>>> {
|
|
self.body(id)
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
use keyring::Keyring;
|
|
use test_client::{self, TestClient};
|
|
use test_client::client::BlockOrigin;
|
|
use test_client::client::backend::Backend as TestBackend;
|
|
use test_client::BlockBuilderExt;
|
|
use test_client::runtime::Transfer;
|
|
|
|
#[test]
|
|
fn client_initialises_from_genesis_ok() {
|
|
let client = test_client::new();
|
|
|
|
assert_eq!(client.call_api::<_, u64>("balance_of", &Keyring::Alice.to_raw_public()).unwrap(), 1000);
|
|
assert_eq!(client.call_api::<_, u64>("balance_of", &Keyring::Ferdie.to_raw_public()).unwrap(), 0);
|
|
}
|
|
|
|
#[test]
|
|
fn authorities_call_works() {
|
|
let client = test_client::new();
|
|
|
|
assert_eq!(client.info().unwrap().chain.best_number, 0);
|
|
assert_eq!(client.authorities_at(&BlockId::Number(0)).unwrap(), vec![
|
|
Keyring::Alice.to_raw_public().into(),
|
|
Keyring::Bob.to_raw_public().into(),
|
|
Keyring::Charlie.to_raw_public().into()
|
|
]);
|
|
}
|
|
|
|
#[test]
|
|
fn block_builder_works_with_no_transactions() {
|
|
let client = test_client::new();
|
|
|
|
let builder = client.new_block().unwrap();
|
|
|
|
client.justify_and_import(BlockOrigin::Own, builder.bake().unwrap()).unwrap();
|
|
|
|
assert_eq!(client.info().unwrap().chain.best_number, 1);
|
|
}
|
|
|
|
#[test]
|
|
fn block_builder_works_with_transactions() {
|
|
let client = test_client::new();
|
|
|
|
let mut builder = client.new_block().unwrap();
|
|
|
|
builder.push_transfer(Transfer {
|
|
from: Keyring::Alice.to_raw_public().into(),
|
|
to: Keyring::Ferdie.to_raw_public().into(),
|
|
amount: 42,
|
|
nonce: 0,
|
|
}).unwrap();
|
|
|
|
client.justify_and_import(BlockOrigin::Own, builder.bake().unwrap()).unwrap();
|
|
|
|
assert_eq!(client.info().unwrap().chain.best_number, 1);
|
|
assert!(client.state_at(&BlockId::Number(1)).unwrap() != client.state_at(&BlockId::Number(0)).unwrap());
|
|
assert_eq!(client.call_api::<_, u64>("balance_of", &Keyring::Alice.to_raw_public()).unwrap(), 958);
|
|
assert_eq!(client.call_api::<_, u64>("balance_of", &Keyring::Ferdie.to_raw_public()).unwrap(), 42);
|
|
}
|
|
|
|
#[test]
|
|
fn client_uses_authorities_from_blockchain_cache() {
|
|
let client = test_client::new();
|
|
test_client::client::in_mem::cache_authorities_at(
|
|
client.backend().blockchain(),
|
|
Default::default(),
|
|
Some(vec![[1u8; 32].into()]));
|
|
assert_eq!(client.authorities_at(
|
|
&BlockId::Hash(Default::default())).unwrap(),
|
|
vec![[1u8; 32].into()]);
|
|
}
|
|
|
|
#[test]
|
|
fn block_builder_does_not_include_invalid() {
|
|
let client = test_client::new();
|
|
|
|
let mut builder = client.new_block().unwrap();
|
|
|
|
builder.push_transfer(Transfer {
|
|
from: Keyring::Alice.to_raw_public().into(),
|
|
to: Keyring::Ferdie.to_raw_public().into(),
|
|
amount: 42,
|
|
nonce: 0,
|
|
}).unwrap();
|
|
|
|
assert!(builder.push_transfer(Transfer {
|
|
from: Keyring::Eve.to_raw_public().into(),
|
|
to: Keyring::Alice.to_raw_public().into(),
|
|
amount: 42,
|
|
nonce: 0,
|
|
}).is_err());
|
|
|
|
client.justify_and_import(BlockOrigin::Own, builder.bake().unwrap()).unwrap();
|
|
|
|
assert_eq!(client.info().unwrap().chain.best_number, 1);
|
|
assert!(client.state_at(&BlockId::Number(1)).unwrap() != client.state_at(&BlockId::Number(0)).unwrap());
|
|
assert_eq!(client.body(&BlockId::Number(1)).unwrap().unwrap().len(), 1)
|
|
}
|
|
|
|
#[test]
|
|
fn best_containing_with_genesis_block() {
|
|
// block tree:
|
|
// G
|
|
|
|
let client = test_client::new();
|
|
|
|
let genesis_hash = client.info().unwrap().chain.genesis_hash;
|
|
|
|
assert_eq!(genesis_hash.clone(), client.best_containing(genesis_hash.clone(), None).unwrap().unwrap());
|
|
}
|
|
|
|
#[test]
|
|
fn best_containing_with_hash_not_found() {
|
|
// block tree:
|
|
// G
|
|
|
|
let client = test_client::new();
|
|
|
|
let uninserted_block = client.new_block().unwrap().bake().unwrap();
|
|
|
|
assert_eq!(None, client.best_containing(uninserted_block.hash().clone(), None).unwrap());
|
|
}
|
|
|
|
#[test]
|
|
fn best_containing_with_single_chain_3_blocks() {
|
|
// block tree:
|
|
// G -> A1 -> A2
|
|
|
|
let client = test_client::new();
|
|
|
|
// G -> A1
|
|
let a1 = client.new_block().unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a1.clone()).unwrap();
|
|
|
|
// A1 -> A2
|
|
let a2 = client.new_block().unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a2.clone()).unwrap();
|
|
|
|
let genesis_hash = client.info().unwrap().chain.genesis_hash;
|
|
|
|
assert_eq!(a2.hash(), client.best_containing(genesis_hash, None).unwrap().unwrap());
|
|
assert_eq!(a2.hash(), client.best_containing(a1.hash(), None).unwrap().unwrap());
|
|
assert_eq!(a2.hash(), client.best_containing(a2.hash(), None).unwrap().unwrap());
|
|
}
|
|
|
|
#[test]
|
|
fn best_containing_with_multiple_forks() {
|
|
// NOTE: we use the version of the trait from `test_client`
|
|
// because that is actually different than the version linked to
|
|
// in the test facade crate.
|
|
use test_client::blockchain::Backend as BlockchainBackendT;
|
|
|
|
// block tree:
|
|
// G -> A1 -> A2 -> A3 -> A4 -> A5
|
|
// A1 -> B2 -> B3 -> B4
|
|
// B2 -> C3
|
|
// A1 -> D2
|
|
let client = test_client::new();
|
|
|
|
// G -> A1
|
|
let a1 = client.new_block().unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a1.clone()).unwrap();
|
|
|
|
// A1 -> A2
|
|
let a2 = client.new_block_at(&BlockId::Hash(a1.hash())).unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a2.clone()).unwrap();
|
|
|
|
// A2 -> A3
|
|
let a3 = client.new_block_at(&BlockId::Hash(a2.hash())).unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a3.clone()).unwrap();
|
|
|
|
// A3 -> A4
|
|
let a4 = client.new_block_at(&BlockId::Hash(a3.hash())).unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a4.clone()).unwrap();
|
|
|
|
// A4 -> A5
|
|
let a5 = client.new_block_at(&BlockId::Hash(a4.hash())).unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, a5.clone()).unwrap();
|
|
|
|
// A1 -> B2
|
|
let mut builder = client.new_block_at(&BlockId::Hash(a1.hash())).unwrap();
|
|
// this push is required as otherwise B2 has the same hash as A2 and won't get imported
|
|
builder.push_transfer(Transfer {
|
|
from: Keyring::Alice.to_raw_public().into(),
|
|
to: Keyring::Ferdie.to_raw_public().into(),
|
|
amount: 41,
|
|
nonce: 0,
|
|
}).unwrap();
|
|
let b2 = builder.bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, b2.clone()).unwrap();
|
|
|
|
// B2 -> B3
|
|
let b3 = client.new_block_at(&BlockId::Hash(b2.hash())).unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, b3.clone()).unwrap();
|
|
|
|
// B3 -> B4
|
|
let b4 = client.new_block_at(&BlockId::Hash(b3.hash())).unwrap().bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, b4.clone()).unwrap();
|
|
|
|
// // B2 -> C3
|
|
let mut builder = client.new_block_at(&BlockId::Hash(b2.hash())).unwrap();
|
|
// this push is required as otherwise C3 has the same hash as B3 and won't get imported
|
|
builder.push_transfer(Transfer {
|
|
from: Keyring::Alice.to_raw_public().into(),
|
|
to: Keyring::Ferdie.to_raw_public().into(),
|
|
amount: 1,
|
|
nonce: 1,
|
|
}).unwrap();
|
|
let c3 = builder.bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, c3.clone()).unwrap();
|
|
|
|
// A1 -> D2
|
|
let mut builder = client.new_block_at(&BlockId::Hash(a1.hash())).unwrap();
|
|
// this push is required as otherwise D2 has the same hash as B2 and won't get imported
|
|
builder.push_transfer(Transfer {
|
|
from: Keyring::Alice.to_raw_public().into(),
|
|
to: Keyring::Ferdie.to_raw_public().into(),
|
|
amount: 1,
|
|
nonce: 0,
|
|
}).unwrap();
|
|
let d2 = builder.bake().unwrap();
|
|
client.justify_and_import(BlockOrigin::Own, d2.clone()).unwrap();
|
|
|
|
assert_eq!(client.info().unwrap().chain.best_hash, a5.hash());
|
|
|
|
let genesis_hash = client.info().unwrap().chain.genesis_hash;
|
|
let leaves = BlockchainBackendT::leaves(client.backend().blockchain()).unwrap();
|
|
|
|
assert!(leaves.contains(&a5.hash()));
|
|
assert!(leaves.contains(&b4.hash()));
|
|
assert!(leaves.contains(&c3.hash()));
|
|
assert!(leaves.contains(&d2.hash()));
|
|
assert_eq!(leaves.len(), 4);
|
|
|
|
// search without restriction
|
|
|
|
assert_eq!(a5.hash(), client.best_containing(genesis_hash, None).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a1.hash(), None).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a2.hash(), None).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a3.hash(), None).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a4.hash(), None).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a5.hash(), None).unwrap().unwrap());
|
|
|
|
assert_eq!(b4.hash(), client.best_containing(b2.hash(), None).unwrap().unwrap());
|
|
assert_eq!(b4.hash(), client.best_containing(b3.hash(), None).unwrap().unwrap());
|
|
assert_eq!(b4.hash(), client.best_containing(b4.hash(), None).unwrap().unwrap());
|
|
|
|
assert_eq!(c3.hash(), client.best_containing(c3.hash(), None).unwrap().unwrap());
|
|
|
|
assert_eq!(d2.hash(), client.best_containing(d2.hash(), None).unwrap().unwrap());
|
|
|
|
|
|
// search only blocks with number <= 5. equivalent to without restriction for this scenario
|
|
|
|
assert_eq!(a5.hash(), client.best_containing(genesis_hash, Some(5)).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a1.hash(), Some(5)).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a2.hash(), Some(5)).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a3.hash(), Some(5)).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a4.hash(), Some(5)).unwrap().unwrap());
|
|
assert_eq!(a5.hash(), client.best_containing(a5.hash(), Some(5)).unwrap().unwrap());
|
|
|
|
assert_eq!(b4.hash(), client.best_containing(b2.hash(), Some(5)).unwrap().unwrap());
|
|
assert_eq!(b4.hash(), client.best_containing(b3.hash(), Some(5)).unwrap().unwrap());
|
|
assert_eq!(b4.hash(), client.best_containing(b4.hash(), Some(5)).unwrap().unwrap());
|
|
|
|
assert_eq!(c3.hash(), client.best_containing(c3.hash(), Some(5)).unwrap().unwrap());
|
|
|
|
assert_eq!(d2.hash(), client.best_containing(d2.hash(), Some(5)).unwrap().unwrap());
|
|
|
|
|
|
// search only blocks with number <= 4
|
|
|
|
assert_eq!(a4.hash(), client.best_containing(genesis_hash, Some(4)).unwrap().unwrap());
|
|
assert_eq!(a4.hash(), client.best_containing(a1.hash(), Some(4)).unwrap().unwrap());
|
|
assert_eq!(a4.hash(), client.best_containing(a2.hash(), Some(4)).unwrap().unwrap());
|
|
assert_eq!(a4.hash(), client.best_containing(a3.hash(), Some(4)).unwrap().unwrap());
|
|
assert_eq!(a4.hash(), client.best_containing(a4.hash(), Some(4)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(a5.hash(), Some(4)).unwrap());
|
|
|
|
assert_eq!(b4.hash(), client.best_containing(b2.hash(), Some(4)).unwrap().unwrap());
|
|
assert_eq!(b4.hash(), client.best_containing(b3.hash(), Some(4)).unwrap().unwrap());
|
|
assert_eq!(b4.hash(), client.best_containing(b4.hash(), Some(4)).unwrap().unwrap());
|
|
|
|
assert_eq!(c3.hash(), client.best_containing(c3.hash(), Some(4)).unwrap().unwrap());
|
|
|
|
assert_eq!(d2.hash(), client.best_containing(d2.hash(), Some(4)).unwrap().unwrap());
|
|
|
|
|
|
// search only blocks with number <= 3
|
|
|
|
assert_eq!(a3.hash(), client.best_containing(genesis_hash, Some(3)).unwrap().unwrap());
|
|
assert_eq!(a3.hash(), client.best_containing(a1.hash(), Some(3)).unwrap().unwrap());
|
|
assert_eq!(a3.hash(), client.best_containing(a2.hash(), Some(3)).unwrap().unwrap());
|
|
assert_eq!(a3.hash(), client.best_containing(a3.hash(), Some(3)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(a4.hash(), Some(3)).unwrap());
|
|
assert_eq!(None, client.best_containing(a5.hash(), Some(3)).unwrap());
|
|
|
|
assert_eq!(b3.hash(), client.best_containing(b2.hash(), Some(3)).unwrap().unwrap());
|
|
assert_eq!(b3.hash(), client.best_containing(b3.hash(), Some(3)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(b4.hash(), Some(3)).unwrap());
|
|
|
|
assert_eq!(c3.hash(), client.best_containing(c3.hash(), Some(3)).unwrap().unwrap());
|
|
|
|
assert_eq!(d2.hash(), client.best_containing(d2.hash(), Some(3)).unwrap().unwrap());
|
|
|
|
|
|
// search only blocks with number <= 2
|
|
|
|
assert_eq!(a2.hash(), client.best_containing(genesis_hash, Some(2)).unwrap().unwrap());
|
|
assert_eq!(a2.hash(), client.best_containing(a1.hash(), Some(2)).unwrap().unwrap());
|
|
assert_eq!(a2.hash(), client.best_containing(a2.hash(), Some(2)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(a3.hash(), Some(2)).unwrap());
|
|
assert_eq!(None, client.best_containing(a4.hash(), Some(2)).unwrap());
|
|
assert_eq!(None, client.best_containing(a5.hash(), Some(2)).unwrap());
|
|
|
|
assert_eq!(b2.hash(), client.best_containing(b2.hash(), Some(2)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(b3.hash(), Some(2)).unwrap());
|
|
assert_eq!(None, client.best_containing(b4.hash(), Some(2)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(c3.hash(), Some(2)).unwrap());
|
|
|
|
assert_eq!(d2.hash(), client.best_containing(d2.hash(), Some(2)).unwrap().unwrap());
|
|
|
|
|
|
// search only blocks with number <= 1
|
|
|
|
assert_eq!(a1.hash(), client.best_containing(genesis_hash, Some(1)).unwrap().unwrap());
|
|
assert_eq!(a1.hash(), client.best_containing(a1.hash(), Some(1)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(a2.hash(), Some(1)).unwrap());
|
|
assert_eq!(None, client.best_containing(a3.hash(), Some(1)).unwrap());
|
|
assert_eq!(None, client.best_containing(a4.hash(), Some(1)).unwrap());
|
|
assert_eq!(None, client.best_containing(a5.hash(), Some(1)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(b2.hash(), Some(1)).unwrap());
|
|
assert_eq!(None, client.best_containing(b3.hash(), Some(1)).unwrap());
|
|
assert_eq!(None, client.best_containing(b4.hash(), Some(1)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(c3.hash(), Some(1)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(d2.hash(), Some(1)).unwrap());
|
|
|
|
// search only blocks with number <= 0
|
|
|
|
assert_eq!(genesis_hash, client.best_containing(genesis_hash, Some(0)).unwrap().unwrap());
|
|
assert_eq!(None, client.best_containing(a1.hash(), Some(0)).unwrap());
|
|
assert_eq!(None, client.best_containing(a2.hash(), Some(0)).unwrap());
|
|
assert_eq!(None, client.best_containing(a3.hash(), Some(0)).unwrap());
|
|
assert_eq!(None, client.best_containing(a4.hash(), Some(0)).unwrap());
|
|
assert_eq!(None, client.best_containing(a5.hash(), Some(0)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(b2.hash(), Some(0)).unwrap());
|
|
assert_eq!(None, client.best_containing(b3.hash(), Some(0)).unwrap());
|
|
assert_eq!(None, client.best_containing(b4.hash(), Some(0)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(c3.hash().clone(), Some(0)).unwrap());
|
|
|
|
assert_eq!(None, client.best_containing(d2.hash().clone(), Some(0)).unwrap());
|
|
}
|
|
}
|