mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-26 15:47:58 +00:00
002d9260f9
**Update:** Pushed additional changes based on the review comments. **This pull request fixes various spelling mistakes in this repository.** Most of the changes are contained in the first **3** commits: - `Fix spelling mistakes in comments and docs` - `Fix spelling mistakes in test names` - `Fix spelling mistakes in error messages, panic messages, logs and tracing` Other source code spelling mistakes are separated into individual commits for easier reviewing: - `Fix the spelling of 'authority'` - `Fix the spelling of 'REASONABLE_HEADERS_IN_JUSTIFICATION_ANCESTRY'` - `Fix the spelling of 'prev_enqueud_messages'` - `Fix the spelling of 'endpoint'` - `Fix the spelling of 'children'` - `Fix the spelling of 'PenpalSiblingSovereignAccount'` - `Fix the spelling of 'PenpalSudoAccount'` - `Fix the spelling of 'insufficient'` - `Fix the spelling of 'PalletXcmExtrinsicsBenchmark'` - `Fix the spelling of 'subtracted'` - `Fix the spelling of 'CandidatePendingAvailability'` - `Fix the spelling of 'exclusive'` - `Fix the spelling of 'until'` - `Fix the spelling of 'discriminator'` - `Fix the spelling of 'nonexistent'` - `Fix the spelling of 'subsystem'` - `Fix the spelling of 'indices'` - `Fix the spelling of 'committed'` - `Fix the spelling of 'topology'` - `Fix the spelling of 'response'` - `Fix the spelling of 'beneficiary'` - `Fix the spelling of 'formatted'` - `Fix the spelling of 'UNKNOWN_PROOF_REQUEST'` - `Fix the spelling of 'succeeded'` - `Fix the spelling of 'reopened'` - `Fix the spelling of 'proposer'` - `Fix the spelling of 'InstantiationNonce'` - `Fix the spelling of 'depositor'` - `Fix the spelling of 'expiration'` - `Fix the spelling of 'phantom'` - `Fix the spelling of 'AggregatedKeyValue'` - `Fix the spelling of 'randomness'` - `Fix the spelling of 'defendant'` - `Fix the spelling of 'AquaticMammal'` - `Fix the spelling of 'transactions'` - `Fix the spelling of 'PassingTracingSubscriber'` - `Fix the spelling of 'TxSignaturePayload'` - `Fix the spelling of 'versioning'` - `Fix the spelling of 'descendant'` - `Fix the spelling of 'overridden'` - `Fix the spelling of 'network'` Let me know if this structure is adequate. **Note:** The usage of the words `Merkle`, `Merkelize`, `Merklization`, `Merkelization`, `Merkleization`, is somewhat inconsistent but I left it as it is. ~~**Note:** In some places the term `Receival` is used to refer to message reception, IMO `Reception` is the correct word here, but I left it as it is.~~ ~~**Note:** In some places the term `Overlayed` is used instead of the more acceptable version `Overlaid` but I also left it as it is.~~ ~~**Note:** In some places the term `Applyable` is used instead of the correct version `Applicable` but I also left it as it is.~~ **Note:** Some usage of British vs American english e.g. `judgement` vs `judgment`, `initialise` vs `initialize`, `optimise` vs `optimize` etc. are both present in different places, but I suppose that's understandable given the number of contributors. ~~**Note:** There is a spelling mistake in `.github/CODEOWNERS` but it triggers errors in CI when I make changes to it, so I left it as it is.~~
517 lines
16 KiB
Rust
517 lines
16 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: Apache-2.0
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
//! Test implementation for Externalities.
|
|
|
|
use std::{
|
|
any::{Any, TypeId},
|
|
panic::{AssertUnwindSafe, UnwindSafe},
|
|
};
|
|
|
|
use crate::{
|
|
backend::Backend, ext::Ext, InMemoryBackend, OverlayedChanges, StorageKey, StorageValue,
|
|
TrieBackendBuilder,
|
|
};
|
|
|
|
use hash_db::{HashDB, Hasher};
|
|
use sp_core::{
|
|
offchain::testing::TestPersistentOffchainDB,
|
|
storage::{
|
|
well_known_keys::{is_child_storage_key, CODE},
|
|
StateVersion, Storage,
|
|
},
|
|
};
|
|
use sp_externalities::{Extension, ExtensionStore, Extensions};
|
|
use sp_trie::{PrefixedMemoryDB, StorageProof};
|
|
|
|
/// Simple HashMap-based Externalities impl.
|
|
pub struct TestExternalities<H>
|
|
where
|
|
H: Hasher + 'static,
|
|
H::Out: codec::Codec + Ord,
|
|
{
|
|
/// The overlay changed storage.
|
|
overlay: OverlayedChanges<H>,
|
|
offchain_db: TestPersistentOffchainDB,
|
|
/// Storage backend.
|
|
pub backend: InMemoryBackend<H>,
|
|
/// Extensions.
|
|
pub extensions: Extensions,
|
|
/// State version to use during tests.
|
|
pub state_version: StateVersion,
|
|
}
|
|
|
|
impl<H> TestExternalities<H>
|
|
where
|
|
H: Hasher + 'static,
|
|
H::Out: Ord + 'static + codec::Codec,
|
|
{
|
|
/// Get externalities implementation.
|
|
pub fn ext(&mut self) -> Ext<H, InMemoryBackend<H>> {
|
|
Ext::new(&mut self.overlay, &self.backend, Some(&mut self.extensions))
|
|
}
|
|
|
|
/// Create a new instance of `TestExternalities` with storage.
|
|
pub fn new(storage: Storage) -> Self {
|
|
Self::new_with_code_and_state(&[], storage, Default::default())
|
|
}
|
|
|
|
/// Create a new instance of `TestExternalities` with storage for a given state version.
|
|
pub fn new_with_state_version(storage: Storage, state_version: StateVersion) -> Self {
|
|
Self::new_with_code_and_state(&[], storage, state_version)
|
|
}
|
|
|
|
/// New empty test externalities.
|
|
pub fn new_empty() -> Self {
|
|
Self::new_with_code_and_state(&[], Storage::default(), Default::default())
|
|
}
|
|
|
|
/// Create a new instance of `TestExternalities` with code and storage.
|
|
pub fn new_with_code(code: &[u8], storage: Storage) -> Self {
|
|
Self::new_with_code_and_state(code, storage, Default::default())
|
|
}
|
|
|
|
/// Create a new instance of `TestExternalities` with code and storage for a given state
|
|
/// version.
|
|
pub fn new_with_code_and_state(
|
|
code: &[u8],
|
|
mut storage: Storage,
|
|
state_version: StateVersion,
|
|
) -> Self {
|
|
assert!(storage.top.keys().all(|key| !is_child_storage_key(key)));
|
|
|
|
storage.top.insert(CODE.to_vec(), code.to_vec());
|
|
|
|
let offchain_db = TestPersistentOffchainDB::new();
|
|
|
|
let backend = (storage, state_version).into();
|
|
|
|
TestExternalities {
|
|
overlay: OverlayedChanges::default(),
|
|
offchain_db,
|
|
extensions: Default::default(),
|
|
backend,
|
|
state_version,
|
|
}
|
|
}
|
|
|
|
/// Returns the overlayed changes.
|
|
pub fn overlayed_changes(&self) -> &OverlayedChanges<H> {
|
|
&self.overlay
|
|
}
|
|
|
|
/// Move offchain changes from overlay to the persistent store.
|
|
pub fn persist_offchain_overlay(&mut self) {
|
|
self.offchain_db.apply_offchain_changes(self.overlay.offchain_drain_committed());
|
|
}
|
|
|
|
/// A shared reference type around the offchain worker storage.
|
|
pub fn offchain_db(&self) -> TestPersistentOffchainDB {
|
|
self.offchain_db.clone()
|
|
}
|
|
|
|
/// Batch insert key/values into backend
|
|
pub fn batch_insert<I>(&mut self, kvs: I)
|
|
where
|
|
I: IntoIterator<Item = (StorageKey, StorageValue)>,
|
|
{
|
|
self.backend.insert(
|
|
Some((None, kvs.into_iter().map(|(k, v)| (k, Some(v))).collect())),
|
|
self.state_version,
|
|
);
|
|
}
|
|
|
|
/// Insert key/value into backend
|
|
pub fn insert(&mut self, k: StorageKey, v: StorageValue) {
|
|
self.backend.insert(vec![(None, vec![(k, Some(v))])], self.state_version);
|
|
}
|
|
|
|
/// Insert key/value into backend.
|
|
///
|
|
/// This only supports inserting keys in child tries.
|
|
pub fn insert_child(&mut self, c: sp_core::storage::ChildInfo, k: StorageKey, v: StorageValue) {
|
|
self.backend.insert(vec![(Some(c), vec![(k, Some(v))])], self.state_version);
|
|
}
|
|
|
|
/// Registers the given extension for this instance.
|
|
pub fn register_extension<E: Any + Extension>(&mut self, ext: E) {
|
|
self.extensions.register(ext);
|
|
}
|
|
|
|
/// Sets raw storage key/values and a root.
|
|
///
|
|
/// This can be used as a fast way to restore the storage state from a backup because the trie
|
|
/// does not need to be computed.
|
|
pub fn from_raw_snapshot(
|
|
raw_storage: Vec<(Vec<u8>, (Vec<u8>, i32))>,
|
|
storage_root: H::Out,
|
|
state_version: StateVersion,
|
|
) -> Self {
|
|
let mut backend = PrefixedMemoryDB::default();
|
|
|
|
for (key, (v, ref_count)) in raw_storage {
|
|
let mut hash = H::Out::default();
|
|
let hash_len = hash.as_ref().len();
|
|
|
|
if key.len() < hash_len {
|
|
log::warn!("Invalid key in `from_raw_snapshot`: {key:?}");
|
|
continue
|
|
}
|
|
|
|
hash.as_mut().copy_from_slice(&key[(key.len() - hash_len)..]);
|
|
|
|
// Each time .emplace is called the internal MemoryDb ref count increments.
|
|
// Repeatedly call emplace to initialise the ref count to the correct value.
|
|
for _ in 0..ref_count {
|
|
backend.emplace(hash, (&key[..(key.len() - hash_len)], None), v.clone());
|
|
}
|
|
}
|
|
|
|
Self {
|
|
backend: TrieBackendBuilder::new(backend, storage_root).build(),
|
|
overlay: Default::default(),
|
|
offchain_db: Default::default(),
|
|
extensions: Default::default(),
|
|
state_version,
|
|
}
|
|
}
|
|
|
|
/// Drains the underlying raw storage key/values and returns the root hash.
|
|
///
|
|
/// Useful for backing up the storage in a format that can be quickly re-loaded.
|
|
pub fn into_raw_snapshot(mut self) -> (Vec<(Vec<u8>, (Vec<u8>, i32))>, H::Out) {
|
|
let raw_key_values = self
|
|
.backend
|
|
.backend_storage_mut()
|
|
.drain()
|
|
.into_iter()
|
|
.filter(|(_, (_, r))| *r > 0)
|
|
.collect::<Vec<(Vec<u8>, (Vec<u8>, i32))>>();
|
|
|
|
(raw_key_values, *self.backend.root())
|
|
}
|
|
|
|
/// Return a new backend with all pending changes.
|
|
///
|
|
/// In contrast to [`commit_all`](Self::commit_all) this will not panic if there are open
|
|
/// transactions.
|
|
pub fn as_backend(&self) -> InMemoryBackend<H> {
|
|
let top: Vec<_> =
|
|
self.overlay.changes().map(|(k, v)| (k.clone(), v.value().cloned())).collect();
|
|
let mut transaction = vec![(None, top)];
|
|
|
|
for (child_changes, child_info) in self.overlay.children() {
|
|
transaction.push((
|
|
Some(child_info.clone()),
|
|
child_changes.map(|(k, v)| (k.clone(), v.value().cloned())).collect(),
|
|
))
|
|
}
|
|
|
|
self.backend.update(transaction, self.state_version)
|
|
}
|
|
|
|
/// Commit all pending changes to the underlying backend.
|
|
///
|
|
/// # Panic
|
|
///
|
|
/// This will panic if there are still open transactions.
|
|
pub fn commit_all(&mut self) -> Result<(), String> {
|
|
let changes = self.overlay.drain_storage_changes(&self.backend, self.state_version)?;
|
|
|
|
self.backend
|
|
.apply_transaction(changes.transaction_storage_root, changes.transaction);
|
|
Ok(())
|
|
}
|
|
|
|
/// Execute the given closure while `self` is set as externalities.
|
|
///
|
|
/// Returns the result of the given closure.
|
|
pub fn execute_with<R>(&mut self, execute: impl FnOnce() -> R) -> R {
|
|
let mut ext = self.ext();
|
|
sp_externalities::set_and_run_with_externalities(&mut ext, execute)
|
|
}
|
|
|
|
/// Execute the given closure while `self`, with `proving_backend` as backend, is set as
|
|
/// externalities.
|
|
///
|
|
/// This implementation will wipe the proof recorded in between calls. Consecutive calls will
|
|
/// get their own proof from scratch.
|
|
pub fn execute_and_prove<R>(&mut self, execute: impl FnOnce() -> R) -> (R, StorageProof) {
|
|
let proving_backend = TrieBackendBuilder::wrap(&self.backend)
|
|
.with_recorder(Default::default())
|
|
.build();
|
|
let mut proving_ext =
|
|
Ext::new(&mut self.overlay, &proving_backend, Some(&mut self.extensions));
|
|
|
|
let outcome = sp_externalities::set_and_run_with_externalities(&mut proving_ext, execute);
|
|
let proof = proving_backend.extract_proof().expect("Failed to extract storage proof");
|
|
|
|
(outcome, proof)
|
|
}
|
|
|
|
/// Execute the given closure while `self` is set as externalities.
|
|
///
|
|
/// Returns the result of the given closure, if no panics occurred.
|
|
/// Otherwise, returns `Err`.
|
|
pub fn execute_with_safe<R>(
|
|
&mut self,
|
|
f: impl FnOnce() -> R + UnwindSafe,
|
|
) -> Result<R, String> {
|
|
let mut ext = AssertUnwindSafe(self.ext());
|
|
std::panic::catch_unwind(move || {
|
|
sp_externalities::set_and_run_with_externalities(&mut *ext, f)
|
|
})
|
|
.map_err(|e| format!("Closure panicked: {:?}", e))
|
|
}
|
|
}
|
|
|
|
impl<H: Hasher> std::fmt::Debug for TestExternalities<H>
|
|
where
|
|
H::Out: Ord + codec::Codec,
|
|
{
|
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
|
let pairs: Vec<_> = self
|
|
.backend
|
|
.pairs(Default::default())
|
|
.expect("creating an iterator over all of the pairs doesn't fail in tests")
|
|
.collect();
|
|
write!(f, "overlay: {:?}\nbackend: {:?}", self.overlay, pairs)
|
|
}
|
|
}
|
|
|
|
impl<H: Hasher> PartialEq for TestExternalities<H>
|
|
where
|
|
H::Out: Ord + 'static + codec::Codec,
|
|
{
|
|
/// This doesn't test if they are in the same state, only if they contains the
|
|
/// same data at this state
|
|
fn eq(&self, other: &TestExternalities<H>) -> bool {
|
|
self.as_backend().eq(&other.as_backend())
|
|
}
|
|
}
|
|
|
|
impl<H: Hasher> Default for TestExternalities<H>
|
|
where
|
|
H::Out: Ord + 'static + codec::Codec,
|
|
{
|
|
fn default() -> Self {
|
|
// default to default version.
|
|
Self::new_with_state_version(Storage::default(), Default::default())
|
|
}
|
|
}
|
|
|
|
impl<H: Hasher> From<Storage> for TestExternalities<H>
|
|
where
|
|
H::Out: Ord + 'static + codec::Codec,
|
|
{
|
|
fn from(storage: Storage) -> Self {
|
|
Self::new_with_state_version(storage, Default::default())
|
|
}
|
|
}
|
|
|
|
impl<H: Hasher> From<(Storage, StateVersion)> for TestExternalities<H>
|
|
where
|
|
H::Out: Ord + 'static + codec::Codec,
|
|
{
|
|
fn from((storage, state_version): (Storage, StateVersion)) -> Self {
|
|
Self::new_with_state_version(storage, state_version)
|
|
}
|
|
}
|
|
|
|
impl<H> sp_externalities::ExtensionStore for TestExternalities<H>
|
|
where
|
|
H: Hasher,
|
|
H::Out: Ord + codec::Codec,
|
|
{
|
|
fn extension_by_type_id(&mut self, type_id: TypeId) -> Option<&mut dyn Any> {
|
|
self.extensions.get_mut(type_id)
|
|
}
|
|
|
|
fn register_extension_with_type_id(
|
|
&mut self,
|
|
type_id: TypeId,
|
|
extension: Box<dyn Extension>,
|
|
) -> Result<(), sp_externalities::Error> {
|
|
self.extensions.register_with_type_id(type_id, extension)
|
|
}
|
|
|
|
fn deregister_extension_by_type_id(
|
|
&mut self,
|
|
type_id: TypeId,
|
|
) -> Result<(), sp_externalities::Error> {
|
|
if self.extensions.deregister(type_id) {
|
|
Ok(())
|
|
} else {
|
|
Err(sp_externalities::Error::ExtensionIsNotRegistered(type_id))
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<H> sp_externalities::ExternalitiesExt for TestExternalities<H>
|
|
where
|
|
H: Hasher,
|
|
H::Out: Ord + codec::Codec,
|
|
{
|
|
fn extension<T: Any + Extension>(&mut self) -> Option<&mut T> {
|
|
self.extension_by_type_id(TypeId::of::<T>()).and_then(<dyn Any>::downcast_mut)
|
|
}
|
|
|
|
fn register_extension<T: Extension>(&mut self, ext: T) -> Result<(), sp_externalities::Error> {
|
|
self.register_extension_with_type_id(TypeId::of::<T>(), Box::new(ext))
|
|
}
|
|
|
|
fn deregister_extension<T: Extension>(&mut self) -> Result<(), sp_externalities::Error> {
|
|
self.deregister_extension_by_type_id(TypeId::of::<T>())
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
use sp_core::{storage::ChildInfo, traits::Externalities, H256};
|
|
use sp_runtime::traits::BlakeTwo256;
|
|
|
|
#[test]
|
|
fn commit_should_work() {
|
|
let storage = Storage::default(); // avoid adding the trie threshold.
|
|
let mut ext = TestExternalities::<BlakeTwo256>::from((storage, Default::default()));
|
|
let mut ext = ext.ext();
|
|
ext.set_storage(b"doe".to_vec(), b"reindeer".to_vec());
|
|
ext.set_storage(b"dog".to_vec(), b"puppy".to_vec());
|
|
ext.set_storage(b"dogglesworth".to_vec(), b"cat".to_vec());
|
|
let root = array_bytes::hex_n_into_unchecked::<_, H256, 32>(
|
|
"ed4d8c799d996add422395a6abd7545491d40bd838d738afafa1b8a4de625489",
|
|
);
|
|
assert_eq!(H256::from_slice(ext.storage_root(Default::default()).as_slice()), root);
|
|
}
|
|
|
|
#[test]
|
|
fn raw_storage_drain_and_restore() {
|
|
// Create a TestExternalities with some data in it.
|
|
let mut original_ext =
|
|
TestExternalities::<BlakeTwo256>::from((Default::default(), Default::default()));
|
|
original_ext.insert(b"doe".to_vec(), b"reindeer".to_vec());
|
|
original_ext.insert(b"dog".to_vec(), b"puppy".to_vec());
|
|
original_ext.insert(b"dogglesworth".to_vec(), b"cat".to_vec());
|
|
let child_info = ChildInfo::new_default(&b"test_child"[..]);
|
|
original_ext.insert_child(child_info.clone(), b"cattytown".to_vec(), b"is_dark".to_vec());
|
|
original_ext.insert_child(child_info.clone(), b"doggytown".to_vec(), b"is_sunny".to_vec());
|
|
|
|
// Apply the backend to itself again to increase the ref count of all nodes.
|
|
original_ext.backend.apply_transaction(
|
|
*original_ext.backend.root(),
|
|
original_ext.backend.clone().into_storage(),
|
|
);
|
|
|
|
// Ensure all have the correct ref count
|
|
assert!(original_ext.backend.backend_storage().keys().values().all(|r| *r == 2));
|
|
|
|
// Drain the raw storage and root.
|
|
let root = *original_ext.backend.root();
|
|
let (raw_storage, storage_root) = original_ext.into_raw_snapshot();
|
|
|
|
// Load the raw storage and root into a new TestExternalities.
|
|
let recovered_ext = TestExternalities::<BlakeTwo256>::from_raw_snapshot(
|
|
raw_storage,
|
|
storage_root,
|
|
Default::default(),
|
|
);
|
|
|
|
// Check the storage root is the same as the original
|
|
assert_eq!(root, *recovered_ext.backend.root());
|
|
|
|
// Check the original storage key/values were recovered correctly
|
|
assert_eq!(recovered_ext.backend.storage(b"doe").unwrap(), Some(b"reindeer".to_vec()));
|
|
assert_eq!(recovered_ext.backend.storage(b"dog").unwrap(), Some(b"puppy".to_vec()));
|
|
assert_eq!(recovered_ext.backend.storage(b"dogglesworth").unwrap(), Some(b"cat".to_vec()));
|
|
|
|
// Check the original child storage key/values were recovered correctly
|
|
assert_eq!(
|
|
recovered_ext.backend.child_storage(&child_info, b"cattytown").unwrap(),
|
|
Some(b"is_dark".to_vec())
|
|
);
|
|
assert_eq!(
|
|
recovered_ext.backend.child_storage(&child_info, b"doggytown").unwrap(),
|
|
Some(b"is_sunny".to_vec())
|
|
);
|
|
|
|
// Ensure all have the correct ref count after importing
|
|
assert!(recovered_ext.backend.backend_storage().keys().values().all(|r| *r == 2));
|
|
}
|
|
|
|
#[test]
|
|
fn set_and_retrieve_code() {
|
|
let mut ext = TestExternalities::<BlakeTwo256>::default();
|
|
let mut ext = ext.ext();
|
|
|
|
let code = vec![1, 2, 3];
|
|
ext.set_storage(CODE.to_vec(), code.clone());
|
|
|
|
assert_eq!(&ext.storage(CODE).unwrap(), &code);
|
|
}
|
|
|
|
#[test]
|
|
fn check_send() {
|
|
fn assert_send<T: Send>() {}
|
|
assert_send::<TestExternalities<BlakeTwo256>>();
|
|
}
|
|
|
|
#[test]
|
|
fn commit_all_and_kill_child_storage() {
|
|
let mut ext = TestExternalities::<BlakeTwo256>::default();
|
|
let child_info = ChildInfo::new_default(&b"test_child"[..]);
|
|
|
|
{
|
|
let mut ext = ext.ext();
|
|
ext.place_child_storage(&child_info, b"doe".to_vec(), Some(b"reindeer".to_vec()));
|
|
ext.place_child_storage(&child_info, b"dog".to_vec(), Some(b"puppy".to_vec()));
|
|
ext.place_child_storage(&child_info, b"dog2".to_vec(), Some(b"puppy2".to_vec()));
|
|
}
|
|
|
|
ext.commit_all().unwrap();
|
|
|
|
{
|
|
let mut ext = ext.ext();
|
|
|
|
assert!(
|
|
ext.kill_child_storage(&child_info, Some(2), None).maybe_cursor.is_some(),
|
|
"Should not delete all keys"
|
|
);
|
|
|
|
assert!(ext.child_storage(&child_info, &b"doe"[..]).is_none());
|
|
assert!(ext.child_storage(&child_info, &b"dog"[..]).is_none());
|
|
assert!(ext.child_storage(&child_info, &b"dog2"[..]).is_some());
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn as_backend_generates_same_backend_as_commit_all() {
|
|
let mut ext = TestExternalities::<BlakeTwo256>::default();
|
|
{
|
|
let mut ext = ext.ext();
|
|
ext.set_storage(b"doe".to_vec(), b"reindeer".to_vec());
|
|
ext.set_storage(b"dog".to_vec(), b"puppy".to_vec());
|
|
ext.set_storage(b"dogglesworth".to_vec(), b"cat".to_vec());
|
|
}
|
|
|
|
let backend = ext.as_backend();
|
|
|
|
ext.commit_all().unwrap();
|
|
assert!(ext.backend.eq(&backend), "Both backend should be equal.");
|
|
}
|
|
}
|