mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-29 05:37:58 +00:00
3f69a2e1ff
* remote ext: make it possible to pass ws client * grumbles: introduce `map_uri`
1169 lines
32 KiB
Rust
1169 lines
32 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) 2020-2022 Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: Apache-2.0
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
//! # Remote Externalities
|
|
//!
|
|
//! An equivalent of `sp_io::TestExternalities` that can load its state from a remote substrate
|
|
//! based chain, or a local state snapshot file.
|
|
|
|
use codec::{Decode, Encode};
|
|
|
|
use jsonrpsee::{
|
|
core::{client::ClientT, Error as RpcError},
|
|
proc_macros::rpc,
|
|
rpc_params,
|
|
ws_client::{WsClient, WsClientBuilder},
|
|
};
|
|
|
|
use log::*;
|
|
use serde::de::DeserializeOwned;
|
|
use sp_core::{
|
|
hashing::twox_128,
|
|
hexdisplay::HexDisplay,
|
|
storage::{
|
|
well_known_keys::{is_default_child_storage_key, DEFAULT_CHILD_STORAGE_KEY_PREFIX},
|
|
ChildInfo, ChildType, PrefixedStorageKey, StorageData, StorageKey,
|
|
},
|
|
};
|
|
pub use sp_io::TestExternalities;
|
|
use sp_runtime::traits::Block as BlockT;
|
|
use std::{
|
|
fs,
|
|
path::{Path, PathBuf},
|
|
sync::Arc,
|
|
};
|
|
|
|
pub mod rpc_api;
|
|
|
|
type KeyValue = (StorageKey, StorageData);
|
|
type TopKeyValues = Vec<KeyValue>;
|
|
type ChildKeyValues = Vec<(ChildInfo, Vec<KeyValue>)>;
|
|
|
|
const LOG_TARGET: &str = "remote-ext";
|
|
const DEFAULT_TARGET: &str = "wss://rpc.polkadot.io:443";
|
|
const BATCH_SIZE: usize = 1000;
|
|
|
|
#[rpc(client)]
|
|
pub trait RpcApi<Hash> {
|
|
#[method(name = "childstate_getKeys")]
|
|
fn child_get_keys(
|
|
&self,
|
|
child_key: PrefixedStorageKey,
|
|
prefix: StorageKey,
|
|
hash: Option<Hash>,
|
|
) -> Result<Vec<StorageKey>, RpcError>;
|
|
|
|
#[method(name = "childstate_getStorage")]
|
|
fn child_get_storage(
|
|
&self,
|
|
child_key: PrefixedStorageKey,
|
|
prefix: StorageKey,
|
|
hash: Option<Hash>,
|
|
) -> Result<StorageData, RpcError>;
|
|
|
|
#[method(name = "state_getStorage")]
|
|
fn get_storage(&self, prefix: StorageKey, hash: Option<Hash>) -> Result<StorageData, RpcError>;
|
|
|
|
#[method(name = "state_getKeysPaged")]
|
|
fn get_keys_paged(
|
|
&self,
|
|
prefix: Option<StorageKey>,
|
|
count: u32,
|
|
start_key: Option<StorageKey>,
|
|
hash: Option<Hash>,
|
|
) -> Result<Vec<StorageKey>, RpcError>;
|
|
|
|
#[method(name = "chain_getFinalizedHead")]
|
|
fn finalized_head(&self) -> Result<Hash, RpcError>;
|
|
}
|
|
|
|
/// The execution mode.
|
|
#[derive(Clone)]
|
|
pub enum Mode<B: BlockT> {
|
|
/// Online. Potentially writes to a cache file.
|
|
Online(OnlineConfig<B>),
|
|
/// Offline. Uses a state snapshot file and needs not any client config.
|
|
Offline(OfflineConfig),
|
|
/// Prefer using a cache file if it exists, else use a remote server.
|
|
OfflineOrElseOnline(OfflineConfig, OnlineConfig<B>),
|
|
}
|
|
|
|
impl<B: BlockT> Default for Mode<B> {
|
|
fn default() -> Self {
|
|
Mode::Online(OnlineConfig::default())
|
|
}
|
|
}
|
|
|
|
/// Configuration of the offline execution.
|
|
///
|
|
/// A state snapshot config must be present.
|
|
#[derive(Clone)]
|
|
pub struct OfflineConfig {
|
|
/// The configuration of the state snapshot file to use. It must be present.
|
|
pub state_snapshot: SnapshotConfig,
|
|
}
|
|
|
|
impl<P: Into<PathBuf>> From<P> for SnapshotConfig {
|
|
fn from(p: P) -> Self {
|
|
Self { path: p.into() }
|
|
}
|
|
}
|
|
|
|
/// Description of the transport protocol (for online execution).
|
|
#[derive(Debug, Clone)]
|
|
pub enum Transport {
|
|
/// Use the `URI` to open a new WebSocket connection.
|
|
Uri(String),
|
|
/// Use existing WebSocket connection.
|
|
RemoteClient(Arc<WsClient>),
|
|
}
|
|
|
|
impl Transport {
|
|
fn as_client(&self) -> Option<&WsClient> {
|
|
match self {
|
|
Self::RemoteClient(client) => Some(&*client),
|
|
_ => None,
|
|
}
|
|
}
|
|
|
|
// Open a new WebSocket connection if it's not connected.
|
|
async fn map_uri(&mut self) -> Result<(), &'static str> {
|
|
if let Self::Uri(uri) = self {
|
|
log::debug!(target: LOG_TARGET, "initializing remote client to {:?}", uri);
|
|
|
|
let ws_client = WsClientBuilder::default()
|
|
.max_request_body_size(u32::MAX)
|
|
.build(&uri)
|
|
.await
|
|
.map_err(|e| {
|
|
log::error!(target: LOG_TARGET, "error: {:?}", e);
|
|
"failed to build ws client"
|
|
})?;
|
|
|
|
*self = Self::RemoteClient(Arc::new(ws_client))
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
impl From<String> for Transport {
|
|
fn from(uri: String) -> Self {
|
|
Transport::Uri(uri)
|
|
}
|
|
}
|
|
|
|
impl From<Arc<WsClient>> for Transport {
|
|
fn from(client: Arc<WsClient>) -> Self {
|
|
Transport::RemoteClient(client)
|
|
}
|
|
}
|
|
|
|
/// Configuration of the online execution.
|
|
///
|
|
/// A state snapshot config may be present and will be written to in that case.
|
|
#[derive(Clone)]
|
|
pub struct OnlineConfig<B: BlockT> {
|
|
/// The block hash at which to get the runtime state. Will be latest finalized head if not
|
|
/// provided.
|
|
pub at: Option<B::Hash>,
|
|
/// An optional state snapshot file to WRITE to, not for reading. Not written if set to `None`.
|
|
pub state_snapshot: Option<SnapshotConfig>,
|
|
/// The pallets to scrape. If empty, entire chain state will be scraped.
|
|
pub pallets: Vec<String>,
|
|
/// Transport config.
|
|
pub transport: Transport,
|
|
}
|
|
|
|
impl<B: BlockT> OnlineConfig<B> {
|
|
/// Return rpc (ws) client.
|
|
fn rpc_client(&self) -> &WsClient {
|
|
self.transport
|
|
.as_client()
|
|
.expect("ws client must have been initialized by now; qed.")
|
|
}
|
|
}
|
|
|
|
impl<B: BlockT> Default for OnlineConfig<B> {
|
|
fn default() -> Self {
|
|
Self {
|
|
transport: Transport::Uri(DEFAULT_TARGET.to_owned()),
|
|
at: None,
|
|
state_snapshot: None,
|
|
pallets: vec![],
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Configuration of the state snapshot.
|
|
#[derive(Clone)]
|
|
pub struct SnapshotConfig {
|
|
/// The path to the snapshot file.
|
|
pub path: PathBuf,
|
|
}
|
|
|
|
impl SnapshotConfig {
|
|
pub fn new<P: Into<PathBuf>>(path: P) -> Self {
|
|
Self { path: path.into() }
|
|
}
|
|
}
|
|
|
|
impl Default for SnapshotConfig {
|
|
fn default() -> Self {
|
|
Self { path: Path::new("SNAPSHOT").into() }
|
|
}
|
|
}
|
|
|
|
/// Builder for remote-externalities.
|
|
pub struct Builder<B: BlockT> {
|
|
/// Custom key-pairs to be injected into the externalities. The *hashed* keys and values must
|
|
/// be given.
|
|
hashed_key_values: Vec<KeyValue>,
|
|
/// Storage entry key prefixes to be injected into the externalities. The *hashed* prefix must
|
|
/// be given.
|
|
hashed_prefixes: Vec<Vec<u8>>,
|
|
/// Storage entry keys to be injected into the externalities. The *hashed* key must be given.
|
|
hashed_keys: Vec<Vec<u8>>,
|
|
/// The keys that will be excluded from the final externality. The *hashed* key must be given.
|
|
hashed_blacklist: Vec<Vec<u8>>,
|
|
/// connectivity mode, online or offline.
|
|
mode: Mode<B>,
|
|
}
|
|
|
|
// NOTE: ideally we would use `DefaultNoBound` here, but not worth bringing in frame-support for
|
|
// that.
|
|
impl<B: BlockT + DeserializeOwned> Default for Builder<B> {
|
|
fn default() -> Self {
|
|
Self {
|
|
mode: Default::default(),
|
|
hashed_key_values: Default::default(),
|
|
hashed_prefixes: Default::default(),
|
|
hashed_keys: Default::default(),
|
|
hashed_blacklist: Default::default(),
|
|
}
|
|
}
|
|
}
|
|
|
|
// Mode methods
|
|
impl<B: BlockT + DeserializeOwned> Builder<B> {
|
|
fn as_online(&self) -> &OnlineConfig<B> {
|
|
match &self.mode {
|
|
Mode::Online(config) => &config,
|
|
Mode::OfflineOrElseOnline(_, config) => &config,
|
|
_ => panic!("Unexpected mode: Online"),
|
|
}
|
|
}
|
|
|
|
fn as_online_mut(&mut self) -> &mut OnlineConfig<B> {
|
|
match &mut self.mode {
|
|
Mode::Online(config) => config,
|
|
Mode::OfflineOrElseOnline(_, config) => config,
|
|
_ => panic!("Unexpected mode: Online"),
|
|
}
|
|
}
|
|
}
|
|
|
|
// RPC methods
|
|
impl<B: BlockT + DeserializeOwned> Builder<B> {
|
|
async fn rpc_get_storage(
|
|
&self,
|
|
key: StorageKey,
|
|
maybe_at: Option<B::Hash>,
|
|
) -> Result<StorageData, &'static str> {
|
|
trace!(target: LOG_TARGET, "rpc: get_storage");
|
|
self.as_online().rpc_client().get_storage(key, maybe_at).await.map_err(|e| {
|
|
error!(target: LOG_TARGET, "Error = {:?}", e);
|
|
"rpc get_storage failed."
|
|
})
|
|
}
|
|
|
|
/// Get the latest finalized head.
|
|
async fn rpc_get_head(&self) -> Result<B::Hash, &'static str> {
|
|
trace!(target: LOG_TARGET, "rpc: finalized_head");
|
|
self.as_online().rpc_client().finalized_head().await.map_err(|e| {
|
|
error!(target: LOG_TARGET, "Error = {:?}", e);
|
|
"rpc finalized_head failed."
|
|
})
|
|
}
|
|
|
|
/// Get all the keys at `prefix` at `hash` using the paged, safe RPC methods.
|
|
async fn rpc_get_keys_paged(
|
|
&self,
|
|
prefix: StorageKey,
|
|
at: B::Hash,
|
|
) -> Result<Vec<StorageKey>, &'static str> {
|
|
const PAGE: u32 = 512;
|
|
let mut last_key: Option<StorageKey> = None;
|
|
let mut all_keys: Vec<StorageKey> = vec![];
|
|
let keys = loop {
|
|
let page = self
|
|
.as_online()
|
|
.rpc_client()
|
|
.get_keys_paged(Some(prefix.clone()), PAGE, last_key.clone(), Some(at))
|
|
.await
|
|
.map_err(|e| {
|
|
error!(target: LOG_TARGET, "Error = {:?}", e);
|
|
"rpc get_keys failed"
|
|
})?;
|
|
let page_len = page.len();
|
|
all_keys.extend(page);
|
|
|
|
if page_len < PAGE as usize {
|
|
log::debug!(target: LOG_TARGET, "last page received: {}", page_len);
|
|
break all_keys
|
|
} else {
|
|
let new_last_key =
|
|
all_keys.last().expect("all_keys is populated; has .last(); qed");
|
|
log::debug!(
|
|
target: LOG_TARGET,
|
|
"new total = {}, full page received: {}",
|
|
all_keys.len(),
|
|
HexDisplay::from(new_last_key)
|
|
);
|
|
last_key = Some(new_last_key.clone());
|
|
}
|
|
};
|
|
|
|
Ok(keys)
|
|
}
|
|
|
|
/// Synonym of `rpc_get_pairs_unsafe` that uses paged queries to first get the keys, and then
|
|
/// map them to values one by one.
|
|
///
|
|
/// This can work with public nodes. But, expect it to be darn slow.
|
|
pub(crate) async fn rpc_get_pairs_paged(
|
|
&self,
|
|
prefix: StorageKey,
|
|
at: B::Hash,
|
|
) -> Result<Vec<KeyValue>, &'static str> {
|
|
let keys = self.rpc_get_keys_paged(prefix, at).await?;
|
|
let keys_count = keys.len();
|
|
log::debug!(target: LOG_TARGET, "Querying a total of {} keys", keys.len());
|
|
|
|
let mut key_values: Vec<KeyValue> = vec![];
|
|
let client = self.as_online().rpc_client();
|
|
for chunk_keys in keys.chunks(BATCH_SIZE) {
|
|
let batch = chunk_keys
|
|
.iter()
|
|
.cloned()
|
|
.map(|key| ("state_getStorage", rpc_params![key, at]))
|
|
.collect::<Vec<_>>();
|
|
let values = client.batch_request::<Option<StorageData>>(batch).await.map_err(|e| {
|
|
log::error!(
|
|
target: LOG_TARGET,
|
|
"failed to execute batch: {:?}. Error: {:?}",
|
|
chunk_keys,
|
|
e
|
|
);
|
|
"batch failed."
|
|
})?;
|
|
|
|
assert_eq!(chunk_keys.len(), values.len());
|
|
|
|
for (idx, key) in chunk_keys.into_iter().enumerate() {
|
|
let maybe_value = values[idx].clone();
|
|
let value = maybe_value.unwrap_or_else(|| {
|
|
log::warn!(target: LOG_TARGET, "key {:?} had none corresponding value.", &key);
|
|
StorageData(vec![])
|
|
});
|
|
key_values.push((key.clone(), value));
|
|
if key_values.len() % (10 * BATCH_SIZE) == 0 {
|
|
let ratio: f64 = key_values.len() as f64 / keys_count as f64;
|
|
log::debug!(
|
|
target: LOG_TARGET,
|
|
"progress = {:.2} [{} / {}]",
|
|
ratio,
|
|
key_values.len(),
|
|
keys_count,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(key_values)
|
|
}
|
|
|
|
/// Get the values corresponding to `child_keys` at the given `prefixed_top_key`.
|
|
pub(crate) async fn rpc_child_get_storage_paged(
|
|
&self,
|
|
prefixed_top_key: &StorageKey,
|
|
child_keys: Vec<StorageKey>,
|
|
at: B::Hash,
|
|
) -> Result<Vec<KeyValue>, &'static str> {
|
|
let mut child_kv_inner = vec![];
|
|
for batch_child_key in child_keys.chunks(BATCH_SIZE) {
|
|
let batch_request = batch_child_key
|
|
.iter()
|
|
.cloned()
|
|
.map(|key| {
|
|
(
|
|
"childstate_getStorage",
|
|
rpc_params![
|
|
PrefixedStorageKey::new(prefixed_top_key.as_ref().to_vec()),
|
|
key,
|
|
at
|
|
],
|
|
)
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
let batch_response = self
|
|
.as_online()
|
|
.rpc_client()
|
|
.batch_request::<Option<StorageData>>(batch_request)
|
|
.await
|
|
.map_err(|e| {
|
|
log::error!(
|
|
target: LOG_TARGET,
|
|
"failed to execute batch: {:?}. Error: {:?}",
|
|
batch_child_key,
|
|
e
|
|
);
|
|
"batch failed."
|
|
})?;
|
|
|
|
assert_eq!(batch_child_key.len(), batch_response.len());
|
|
|
|
for (idx, key) in batch_child_key.into_iter().enumerate() {
|
|
let maybe_value = batch_response[idx].clone();
|
|
let value = maybe_value.unwrap_or_else(|| {
|
|
log::warn!(target: LOG_TARGET, "key {:?} had none corresponding value.", &key);
|
|
StorageData(vec![])
|
|
});
|
|
child_kv_inner.push((key.clone(), value));
|
|
}
|
|
}
|
|
|
|
Ok(child_kv_inner)
|
|
}
|
|
|
|
pub(crate) async fn rpc_child_get_keys(
|
|
&self,
|
|
prefixed_top_key: &StorageKey,
|
|
child_prefix: StorageKey,
|
|
at: B::Hash,
|
|
) -> Result<Vec<StorageKey>, &'static str> {
|
|
let child_keys = self
|
|
.as_online()
|
|
.rpc_client()
|
|
.child_get_keys(
|
|
PrefixedStorageKey::new(prefixed_top_key.as_ref().to_vec()),
|
|
child_prefix,
|
|
Some(at),
|
|
)
|
|
.await
|
|
.map_err(|e| {
|
|
error!(target: LOG_TARGET, "Error = {:?}", e);
|
|
"rpc child_get_keys failed."
|
|
})?;
|
|
|
|
debug!(
|
|
target: LOG_TARGET,
|
|
"scraped {} child-keys of the child-bearing top key: {}",
|
|
child_keys.len(),
|
|
HexDisplay::from(prefixed_top_key)
|
|
);
|
|
|
|
Ok(child_keys)
|
|
}
|
|
}
|
|
|
|
// Internal methods
|
|
impl<B: BlockT + DeserializeOwned> Builder<B> {
|
|
/// Save the given data to the top keys snapshot.
|
|
fn save_top_snapshot(&self, data: &[KeyValue], path: &PathBuf) -> Result<(), &'static str> {
|
|
let mut path = path.clone();
|
|
let encoded = data.encode();
|
|
path.set_extension("top");
|
|
debug!(
|
|
target: LOG_TARGET,
|
|
"writing {} bytes to state snapshot file {:?}",
|
|
encoded.len(),
|
|
path
|
|
);
|
|
fs::write(path, encoded).map_err(|_| "fs::write failed.")?;
|
|
Ok(())
|
|
}
|
|
|
|
/// Save the given data to the child keys snapshot.
|
|
fn save_child_snapshot(
|
|
&self,
|
|
data: &ChildKeyValues,
|
|
path: &PathBuf,
|
|
) -> Result<(), &'static str> {
|
|
let mut path = path.clone();
|
|
path.set_extension("child");
|
|
let encoded = data.encode();
|
|
debug!(
|
|
target: LOG_TARGET,
|
|
"writing {} bytes to state snapshot file {:?}",
|
|
encoded.len(),
|
|
path
|
|
);
|
|
fs::write(path, encoded).map_err(|_| "fs::write failed.")?;
|
|
Ok(())
|
|
}
|
|
|
|
fn load_top_snapshot(&self, path: &PathBuf) -> Result<TopKeyValues, &'static str> {
|
|
let mut path = path.clone();
|
|
path.set_extension("top");
|
|
info!(target: LOG_TARGET, "loading top key-pairs from snapshot {:?}", path);
|
|
let bytes = fs::read(path).map_err(|_| "fs::read failed.")?;
|
|
Decode::decode(&mut &*bytes).map_err(|e| {
|
|
log::error!(target: LOG_TARGET, "{:?}", e);
|
|
"decode failed"
|
|
})
|
|
}
|
|
|
|
fn load_child_snapshot(&self, path: &PathBuf) -> Result<ChildKeyValues, &'static str> {
|
|
let mut path = path.clone();
|
|
path.set_extension("child");
|
|
info!(target: LOG_TARGET, "loading child key-pairs from snapshot {:?}", path);
|
|
let bytes = fs::read(path).map_err(|_| "fs::read failed.")?;
|
|
Decode::decode(&mut &*bytes).map_err(|e| {
|
|
log::error!(target: LOG_TARGET, "{:?}", e);
|
|
"decode failed"
|
|
})
|
|
}
|
|
|
|
/// Load all the `top` keys from the remote config, and maybe write then to cache.
|
|
async fn load_top_remote_and_maybe_save(&self) -> Result<TopKeyValues, &'static str> {
|
|
let top_kv = self.load_top_remote().await?;
|
|
if let Some(c) = &self.as_online().state_snapshot {
|
|
self.save_top_snapshot(&top_kv, &c.path)?;
|
|
}
|
|
Ok(top_kv)
|
|
}
|
|
|
|
/// Load all of the child keys from the remote config, given the already scraped list of top key
|
|
/// pairs.
|
|
///
|
|
/// Stores all values to cache as well, if provided.
|
|
async fn load_child_remote_and_maybe_save(
|
|
&self,
|
|
top_kv: &[KeyValue],
|
|
) -> Result<ChildKeyValues, &'static str> {
|
|
let child_kv = self.load_child_remote(&top_kv).await?;
|
|
if let Some(c) = &self.as_online().state_snapshot {
|
|
self.save_child_snapshot(&child_kv, &c.path)?;
|
|
}
|
|
Ok(child_kv)
|
|
}
|
|
|
|
/// Load all of the child keys from the remote config, given the already scraped list of top key
|
|
/// pairs.
|
|
///
|
|
/// `top_kv` need not be only child-bearing top keys. It should be all of the top keys that are
|
|
/// included thus far.
|
|
async fn load_child_remote(&self, top_kv: &[KeyValue]) -> Result<ChildKeyValues, &'static str> {
|
|
let child_roots = top_kv
|
|
.iter()
|
|
.filter_map(|(k, _)| is_default_child_storage_key(k.as_ref()).then(|| k))
|
|
.collect::<Vec<_>>();
|
|
|
|
info!(
|
|
target: LOG_TARGET,
|
|
"👩👦 scraping child-tree data from {} top keys",
|
|
child_roots.len()
|
|
);
|
|
|
|
let mut child_kv = vec![];
|
|
for prefixed_top_key in child_roots {
|
|
let at = self.as_online().at.expect("at must be initialized in online mode.");
|
|
let child_keys =
|
|
self.rpc_child_get_keys(prefixed_top_key, StorageKey(vec![]), at).await?;
|
|
let child_kv_inner =
|
|
self.rpc_child_get_storage_paged(prefixed_top_key, child_keys, at).await?;
|
|
|
|
let prefixed_top_key = PrefixedStorageKey::new(prefixed_top_key.clone().0);
|
|
let un_prefixed = match ChildType::from_prefixed_key(&prefixed_top_key) {
|
|
Some((ChildType::ParentKeyId, storage_key)) => storage_key,
|
|
None => {
|
|
log::error!(target: LOG_TARGET, "invalid key: {:?}", prefixed_top_key);
|
|
return Err("Invalid child key")
|
|
},
|
|
};
|
|
|
|
child_kv.push((ChildInfo::new_default(&un_prefixed), child_kv_inner));
|
|
}
|
|
|
|
Ok(child_kv)
|
|
}
|
|
|
|
/// Build `Self` from a network node denoted by `uri`.
|
|
async fn load_top_remote(&self) -> Result<TopKeyValues, &'static str> {
|
|
let config = self.as_online();
|
|
let at = self
|
|
.as_online()
|
|
.at
|
|
.expect("online config must be initialized by this point; qed.")
|
|
.clone();
|
|
log::info!(target: LOG_TARGET, "scraping key-pairs from remote @ {:?}", at);
|
|
|
|
let mut keys_and_values = if config.pallets.len() > 0 {
|
|
let mut filtered_kv = vec![];
|
|
for p in config.pallets.iter() {
|
|
let hashed_prefix = StorageKey(twox_128(p.as_bytes()).to_vec());
|
|
let pallet_kv = self.rpc_get_pairs_paged(hashed_prefix.clone(), at).await?;
|
|
log::info!(
|
|
target: LOG_TARGET,
|
|
"downloaded data for module {} (count: {} / prefix: {}).",
|
|
p,
|
|
pallet_kv.len(),
|
|
HexDisplay::from(&hashed_prefix),
|
|
);
|
|
filtered_kv.extend(pallet_kv);
|
|
}
|
|
filtered_kv
|
|
} else {
|
|
log::info!(target: LOG_TARGET, "downloading data for all pallets.");
|
|
self.rpc_get_pairs_paged(StorageKey(vec![]), at).await?
|
|
};
|
|
|
|
for prefix in &self.hashed_prefixes {
|
|
log::info!(
|
|
target: LOG_TARGET,
|
|
"adding data for hashed prefix: {:?}",
|
|
HexDisplay::from(prefix)
|
|
);
|
|
let additional_key_values =
|
|
self.rpc_get_pairs_paged(StorageKey(prefix.to_vec()), at).await?;
|
|
keys_and_values.extend(additional_key_values);
|
|
}
|
|
|
|
for key in &self.hashed_keys {
|
|
let key = StorageKey(key.to_vec());
|
|
log::info!(
|
|
target: LOG_TARGET,
|
|
"adding data for hashed key: {:?}",
|
|
HexDisplay::from(&key)
|
|
);
|
|
let value = self.rpc_get_storage(key.clone(), Some(at)).await?;
|
|
keys_and_values.push((key, value));
|
|
}
|
|
|
|
Ok(keys_and_values)
|
|
}
|
|
|
|
pub(crate) async fn init_remote_client(&mut self) -> Result<(), &'static str> {
|
|
// First, initialize the ws client.
|
|
self.as_online_mut().transport.map_uri().await?;
|
|
|
|
// Then, if `at` is not set, set it.
|
|
if self.as_online().at.is_none() {
|
|
let at = self.rpc_get_head().await?;
|
|
self.as_online_mut().at = Some(at);
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
pub(crate) async fn pre_build(
|
|
mut self,
|
|
) -> Result<(TopKeyValues, ChildKeyValues), &'static str> {
|
|
let mut top_kv = match self.mode.clone() {
|
|
Mode::Offline(config) => self.load_top_snapshot(&config.state_snapshot.path)?,
|
|
Mode::Online(_) => {
|
|
self.init_remote_client().await?;
|
|
self.load_top_remote_and_maybe_save().await?
|
|
},
|
|
Mode::OfflineOrElseOnline(offline_config, _) => {
|
|
if let Ok(kv) = self.load_top_snapshot(&offline_config.state_snapshot.path) {
|
|
kv
|
|
} else {
|
|
self.init_remote_client().await?;
|
|
self.load_top_remote_and_maybe_save().await?
|
|
}
|
|
},
|
|
};
|
|
|
|
// inject manual key values.
|
|
if !self.hashed_key_values.is_empty() {
|
|
log::debug!(
|
|
target: LOG_TARGET,
|
|
"extending externalities with {} manually injected key-values",
|
|
self.hashed_key_values.len()
|
|
);
|
|
top_kv.extend(self.hashed_key_values.clone());
|
|
}
|
|
|
|
// exclude manual key values.
|
|
if !self.hashed_blacklist.is_empty() {
|
|
log::debug!(
|
|
target: LOG_TARGET,
|
|
"excluding externalities from {} keys",
|
|
self.hashed_blacklist.len()
|
|
);
|
|
top_kv.retain(|(k, _)| !self.hashed_blacklist.contains(&k.0))
|
|
}
|
|
|
|
let child_kv = match self.mode.clone() {
|
|
Mode::Online(_) => self.load_child_remote_and_maybe_save(&top_kv).await?,
|
|
Mode::OfflineOrElseOnline(offline_config, _) =>
|
|
if let Ok(kv) = self.load_child_snapshot(&offline_config.state_snapshot.path) {
|
|
kv
|
|
} else {
|
|
self.load_child_remote_and_maybe_save(&top_kv).await?
|
|
},
|
|
Mode::Offline(ref config) => self
|
|
.load_child_snapshot(&config.state_snapshot.path)
|
|
.map_err(|why| {
|
|
log::warn!(
|
|
target: LOG_TARGET,
|
|
"failed to load child-key file due to {:?}.",
|
|
why
|
|
)
|
|
})
|
|
.unwrap_or_default(),
|
|
};
|
|
|
|
Ok((top_kv, child_kv))
|
|
}
|
|
}
|
|
|
|
// Public methods
|
|
impl<B: BlockT + DeserializeOwned> Builder<B> {
|
|
/// Create a new builder.
|
|
pub fn new() -> Self {
|
|
Default::default()
|
|
}
|
|
|
|
/// Inject a manual list of key and values to the storage.
|
|
pub fn inject_hashed_key_value(mut self, injections: &[KeyValue]) -> Self {
|
|
for i in injections {
|
|
self.hashed_key_values.push(i.clone());
|
|
}
|
|
self
|
|
}
|
|
|
|
/// Inject a hashed prefix. This is treated as-is, and should be pre-hashed.
|
|
///
|
|
/// This should be used to inject a "PREFIX", like a storage (double) map.
|
|
pub fn inject_hashed_prefix(mut self, hashed: &[u8]) -> Self {
|
|
self.hashed_prefixes.push(hashed.to_vec());
|
|
self
|
|
}
|
|
|
|
/// Just a utility wrapper of [`Self::inject_hashed_prefix`] that injects
|
|
/// [`DEFAULT_CHILD_STORAGE_KEY_PREFIX`] as a prefix.
|
|
///
|
|
/// If set, this will guarantee that the child-tree data of ALL pallets will be downloaded.
|
|
///
|
|
/// This is not needed if the entire state is being downloaded.
|
|
///
|
|
/// Otherwise, the only other way to make sure a child-tree is manually included is to inject
|
|
/// its root (`DEFAULT_CHILD_STORAGE_KEY_PREFIX`, plus some other postfix) into
|
|
/// [`Self::inject_hashed_key`]. Unfortunately, there's no federated way of managing child tree
|
|
/// roots as of now and each pallet does its own thing. Therefore, it is not possible for this
|
|
/// library to automatically include child trees of pallet X, when its top keys are included.
|
|
pub fn inject_default_child_tree_prefix(self) -> Self {
|
|
self.inject_hashed_prefix(DEFAULT_CHILD_STORAGE_KEY_PREFIX)
|
|
}
|
|
|
|
/// Inject a hashed key to scrape. This is treated as-is, and should be pre-hashed.
|
|
///
|
|
/// This should be used to inject a "KEY", like a storage value.
|
|
pub fn inject_hashed_key(mut self, hashed: &[u8]) -> Self {
|
|
self.hashed_keys.push(hashed.to_vec());
|
|
self
|
|
}
|
|
|
|
/// Blacklist this hashed key from the final externalities. This is treated as-is, and should be
|
|
/// pre-hashed.
|
|
pub fn blacklist_hashed_key(mut self, hashed: &[u8]) -> Self {
|
|
self.hashed_blacklist.push(hashed.to_vec());
|
|
self
|
|
}
|
|
|
|
/// Configure a state snapshot to be used.
|
|
pub fn mode(mut self, mode: Mode<B>) -> Self {
|
|
self.mode = mode;
|
|
self
|
|
}
|
|
|
|
/// overwrite the `at` value, if `mode` is set to [`Mode::Online`].
|
|
///
|
|
/// noop if `mode` is [`Mode::Offline`]
|
|
pub fn overwrite_online_at(mut self, at: B::Hash) -> Self {
|
|
if let Mode::Online(mut online) = self.mode.clone() {
|
|
online.at = Some(at);
|
|
self.mode = Mode::Online(online);
|
|
}
|
|
self
|
|
}
|
|
|
|
/// Build the test externalities.
|
|
pub async fn build(self) -> Result<TestExternalities, &'static str> {
|
|
let (top_kv, child_kv) = self.pre_build().await?;
|
|
let mut ext = TestExternalities::new_with_code(Default::default(), Default::default());
|
|
|
|
info!(target: LOG_TARGET, "injecting a total of {} top keys", top_kv.len());
|
|
for (k, v) in top_kv {
|
|
// skip writing the child root data.
|
|
if is_default_child_storage_key(k.as_ref()) {
|
|
continue
|
|
}
|
|
ext.insert(k.0, v.0);
|
|
}
|
|
|
|
info!(
|
|
target: LOG_TARGET,
|
|
"injecting a total of {} child keys",
|
|
child_kv.iter().map(|(_, kv)| kv).flatten().count()
|
|
);
|
|
|
|
for (info, key_values) in child_kv {
|
|
for (k, v) in key_values {
|
|
ext.insert_child(info.clone(), k.0, v.0);
|
|
}
|
|
}
|
|
|
|
ext.commit_all().unwrap();
|
|
info!(
|
|
target: LOG_TARGET,
|
|
"initialized state externalities with storage root {:?}",
|
|
ext.as_backend().root()
|
|
);
|
|
|
|
Ok(ext)
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod test_prelude {
|
|
pub(crate) use super::*;
|
|
pub(crate) use sp_runtime::testing::{Block as RawBlock, ExtrinsicWrapper, H256 as Hash};
|
|
|
|
pub(crate) type Block = RawBlock<ExtrinsicWrapper<Hash>>;
|
|
|
|
pub(crate) fn init_logger() {
|
|
let _ = env_logger::Builder::from_default_env()
|
|
.format_module_path(true)
|
|
.format_level(true)
|
|
.filter_module(LOG_TARGET, log::LevelFilter::Debug)
|
|
.try_init();
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::test_prelude::*;
|
|
|
|
#[tokio::test]
|
|
async fn can_load_state_snapshot() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Offline(OfflineConfig {
|
|
state_snapshot: SnapshotConfig::new("test_data/proxy_test"),
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect("Can't read state snapshot file")
|
|
.execute_with(|| {});
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_exclude_from_cache() {
|
|
init_logger();
|
|
|
|
// get the first key from the cache file.
|
|
let some_key = Builder::<Block>::new()
|
|
.mode(Mode::Offline(OfflineConfig {
|
|
state_snapshot: SnapshotConfig::new("test_data/proxy_test"),
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect("Can't read state snapshot file")
|
|
.execute_with(|| {
|
|
let key =
|
|
sp_io::storage::next_key(&[]).expect("some key must exist in the snapshot");
|
|
assert!(sp_io::storage::get(&key).is_some());
|
|
key
|
|
});
|
|
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Offline(OfflineConfig {
|
|
state_snapshot: SnapshotConfig::new("test_data/proxy_test"),
|
|
}))
|
|
.blacklist_hashed_key(&some_key)
|
|
.build()
|
|
.await
|
|
.expect("Can't read state snapshot file")
|
|
.execute_with(|| assert!(sp_io::storage::get(&some_key).is_none()));
|
|
}
|
|
}
|
|
|
|
#[cfg(all(test, feature = "remote-test"))]
|
|
mod remote_tests {
|
|
use super::test_prelude::*;
|
|
const REMOTE_INACCESSIBLE: &'static str = "Can't reach the remote node. Is it running?";
|
|
|
|
#[tokio::test]
|
|
async fn offline_else_online_works() {
|
|
init_logger();
|
|
// this shows that in the second run, we use the remote and create a cache.
|
|
Builder::<Block>::new()
|
|
.mode(Mode::OfflineOrElseOnline(
|
|
OfflineConfig {
|
|
state_snapshot: SnapshotConfig::new("offline_else_online_works_data"),
|
|
},
|
|
OnlineConfig {
|
|
pallets: vec!["Proxy".to_owned()],
|
|
state_snapshot: Some(SnapshotConfig::new("offline_else_online_works_data")),
|
|
..Default::default()
|
|
},
|
|
))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
// this shows that in the second run, we are not using the remote
|
|
Builder::<Block>::new()
|
|
.mode(Mode::OfflineOrElseOnline(
|
|
OfflineConfig {
|
|
state_snapshot: SnapshotConfig::new("offline_else_online_works_data"),
|
|
},
|
|
OnlineConfig {
|
|
pallets: vec!["Proxy".to_owned()],
|
|
state_snapshot: Some(SnapshotConfig::new("offline_else_online_works_data")),
|
|
transport: "ws://non-existent:666".to_owned().into(),
|
|
..Default::default()
|
|
},
|
|
))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
let to_delete = std::fs::read_dir(Path::new("."))
|
|
.unwrap()
|
|
.into_iter()
|
|
.map(|d| d.unwrap())
|
|
.filter(|p| {
|
|
p.path().file_name().unwrap_or_default() == "offline_else_online_works_data" ||
|
|
p.path().extension().unwrap_or_default() == "top" ||
|
|
p.path().extension().unwrap_or_default() == "child"
|
|
})
|
|
.collect::<Vec<_>>();
|
|
assert!(to_delete.len() > 0);
|
|
for d in to_delete {
|
|
std::fs::remove_file(d.path()).unwrap();
|
|
}
|
|
}
|
|
|
|
#[tokio::test]
|
|
#[ignore = "too slow"]
|
|
async fn can_build_one_big_pallet() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
pallets: vec!["System".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_build_one_small_pallet() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
transport: "wss://kusama-rpc.polkadot.io:443".to_owned().into(),
|
|
pallets: vec!["Council".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
transport: "wss://rpc.polkadot.io:443".to_owned().into(),
|
|
pallets: vec!["Council".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_build_few_pallet() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
transport: "wss://kusama-rpc.polkadot.io:443".to_owned().into(),
|
|
pallets: vec!["Proxy".to_owned(), "Multisig".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
transport: "wss://rpc.polkadot.io:443".to_owned().into(),
|
|
pallets: vec!["Proxy".to_owned(), "Multisig".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_create_top_snapshot() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
state_snapshot: Some(SnapshotConfig::new("can_create_top_snapshot_data")),
|
|
pallets: vec!["Proxy".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
let to_delete = std::fs::read_dir(Path::new("."))
|
|
.unwrap()
|
|
.into_iter()
|
|
.map(|d| d.unwrap())
|
|
.filter(|p| {
|
|
p.path().file_name().unwrap_or_default() == "can_create_top_snapshot_data" ||
|
|
p.path().extension().unwrap_or_default() == "top" ||
|
|
p.path().extension().unwrap_or_default() == "child"
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
assert!(to_delete.len() > 0);
|
|
|
|
for d in to_delete {
|
|
use std::os::unix::fs::MetadataExt;
|
|
if d.path().extension().unwrap_or_default() == "top" {
|
|
// if this is the top snapshot it must not be empty.
|
|
assert!(std::fs::metadata(d.path()).unwrap().size() > 1);
|
|
} else {
|
|
// the child is empty for this pallet.
|
|
assert!(std::fs::metadata(d.path()).unwrap().size() == 1);
|
|
}
|
|
std::fs::remove_file(d.path()).unwrap();
|
|
}
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_build_child_tree() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
transport: "wss://rpc.polkadot.io:443".to_owned().into(),
|
|
pallets: vec!["Crowdloan".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_create_child_snapshot() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
state_snapshot: Some(SnapshotConfig::new("can_create_child_snapshot_data")),
|
|
pallets: vec!["Crowdloan".to_owned()],
|
|
..Default::default()
|
|
}))
|
|
.inject_default_child_tree_prefix()
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
let to_delete = std::fs::read_dir(Path::new("."))
|
|
.unwrap()
|
|
.into_iter()
|
|
.map(|d| d.unwrap())
|
|
.filter(|p| {
|
|
p.path().file_name().unwrap_or_default() == "can_create_child_snapshot_data" ||
|
|
p.path().extension().unwrap_or_default() == "top" ||
|
|
p.path().extension().unwrap_or_default() == "child"
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
assert!(to_delete.len() > 0);
|
|
|
|
for d in to_delete {
|
|
use std::os::unix::fs::MetadataExt;
|
|
// if this is the top snapshot it must not be empty
|
|
if d.path().extension().unwrap_or_default() == "child" {
|
|
assert!(std::fs::metadata(d.path()).unwrap().size() > 1);
|
|
} else {
|
|
assert!(std::fs::metadata(d.path()).unwrap().size() > 1);
|
|
}
|
|
std::fs::remove_file(d.path()).unwrap();
|
|
}
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn can_fetch_all() {
|
|
init_logger();
|
|
Builder::<Block>::new()
|
|
.mode(Mode::Online(OnlineConfig {
|
|
state_snapshot: Some(SnapshotConfig::new("can_fetch_all_data")),
|
|
..Default::default()
|
|
}))
|
|
.build()
|
|
.await
|
|
.expect(REMOTE_INACCESSIBLE)
|
|
.execute_with(|| {});
|
|
|
|
let to_delete = std::fs::read_dir(Path::new("."))
|
|
.unwrap()
|
|
.into_iter()
|
|
.map(|d| d.unwrap())
|
|
.filter(|p| {
|
|
p.path().file_name().unwrap_or_default() == "can_fetch_all_data" ||
|
|
p.path().extension().unwrap_or_default() == "top" ||
|
|
p.path().extension().unwrap_or_default() == "child"
|
|
})
|
|
.collect::<Vec<_>>();
|
|
|
|
assert!(to_delete.len() > 0);
|
|
|
|
for d in to_delete {
|
|
use std::os::unix::fs::MetadataExt;
|
|
// if we download everything, child tree must also be filled.
|
|
if d.path().extension().unwrap_or_default() == "child" {
|
|
assert!(std::fs::metadata(d.path()).unwrap().size() > 1);
|
|
} else {
|
|
assert!(std::fs::metadata(d.path()).unwrap().size() > 1);
|
|
}
|
|
std::fs::remove_file(d.path()).unwrap();
|
|
}
|
|
}
|
|
}
|