mirror of
https://github.com/pezkuwichain/revive-differential-tests.git
synced 2026-04-22 21:57:58 +00:00
Compare commits
29 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| e7ebe4fa2f | |||
| 5c957e5ac1 | |||
| a0248b58f3 | |||
| 6a9acea748 | |||
| c31c7d94f9 | |||
| dcc43d4ea6 | |||
| cb13c3c2cd | |||
| 7cda3416f0 | |||
| d7bc4f1fab | |||
| b6db597a57 | |||
| d7136d9a3d | |||
| e5a3f0aee9 | |||
| 3cdf57f7c3 | |||
| dab8ffe520 | |||
| c913a8222f | |||
| c8cef4834f | |||
| ca59a1f6a9 | |||
| adc0c44cde | |||
| 811e17136b | |||
| ba32bad6b3 | |||
| bb754cba4f | |||
| c858bbe66d | |||
| 906878f06a | |||
| 9a71369e8a | |||
| 84ab873b46 | |||
| 2ef6f7ba63 | |||
| 38e6140a7c | |||
| ca6c5529e2 | |||
| 038a2db53c |
Generated
+5
-16
@@ -3948,22 +3948,10 @@ dependencies = [
|
|||||||
"serde_stacker",
|
"serde_stacker",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
|
||||||
name = "revive-dt-common"
|
|
||||||
version = "0.1.0"
|
|
||||||
dependencies = [
|
|
||||||
"anyhow",
|
|
||||||
"futures",
|
|
||||||
"once_cell",
|
|
||||||
"tokio",
|
|
||||||
"tracing",
|
|
||||||
]
|
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "revive-dt-compiler"
|
name = "revive-dt-compiler"
|
||||||
version = "0.1.0"
|
version = "0.1.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"alloy-primitives",
|
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"revive-common",
|
"revive-common",
|
||||||
"revive-dt-config",
|
"revive-dt-config",
|
||||||
@@ -3994,7 +3982,6 @@ dependencies = [
|
|||||||
"clap",
|
"clap",
|
||||||
"indexmap 2.10.0",
|
"indexmap 2.10.0",
|
||||||
"rayon",
|
"rayon",
|
||||||
"revive-dt-common",
|
|
||||||
"revive-dt-compiler",
|
"revive-dt-compiler",
|
||||||
"revive-dt-config",
|
"revive-dt-config",
|
||||||
"revive-dt-format",
|
"revive-dt-format",
|
||||||
@@ -4016,7 +4003,7 @@ dependencies = [
|
|||||||
"alloy-primitives",
|
"alloy-primitives",
|
||||||
"alloy-sol-types",
|
"alloy-sol-types",
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"revive-dt-common",
|
"revive-dt-node-interaction",
|
||||||
"semver 1.0.26",
|
"semver 1.0.26",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
@@ -4029,9 +4016,7 @@ version = "0.1.0"
|
|||||||
dependencies = [
|
dependencies = [
|
||||||
"alloy",
|
"alloy",
|
||||||
"anyhow",
|
"anyhow",
|
||||||
"revive-dt-common",
|
|
||||||
"revive-dt-config",
|
"revive-dt-config",
|
||||||
"revive-dt-format",
|
|
||||||
"revive-dt-node-interaction",
|
"revive-dt-node-interaction",
|
||||||
"serde",
|
"serde",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
@@ -4048,6 +4033,10 @@ version = "0.1.0"
|
|||||||
dependencies = [
|
dependencies = [
|
||||||
"alloy",
|
"alloy",
|
||||||
"anyhow",
|
"anyhow",
|
||||||
|
"futures",
|
||||||
|
"once_cell",
|
||||||
|
"tokio",
|
||||||
|
"tracing",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
|
|||||||
@@ -11,7 +11,6 @@ repository = "https://github.com/paritytech/revive-differential-testing.git"
|
|||||||
rust-version = "1.85.0"
|
rust-version = "1.85.0"
|
||||||
|
|
||||||
[workspace.dependencies]
|
[workspace.dependencies]
|
||||||
revive-dt-common = { version = "0.1.0", path = "crates/common" }
|
|
||||||
revive-dt-compiler = { version = "0.1.0", path = "crates/compiler" }
|
revive-dt-compiler = { version = "0.1.0", path = "crates/compiler" }
|
||||||
revive-dt-config = { version = "0.1.0", path = "crates/config" }
|
revive-dt-config = { version = "0.1.0", path = "crates/config" }
|
||||||
revive-dt-core = { version = "0.1.0", path = "crates/core" }
|
revive-dt-core = { version = "0.1.0", path = "crates/core" }
|
||||||
|
|||||||
@@ -1,16 +0,0 @@
|
|||||||
[package]
|
|
||||||
name = "revive-dt-common"
|
|
||||||
description = "A library containing common concepts that other crates in the workspace can rely on"
|
|
||||||
version.workspace = true
|
|
||||||
authors.workspace = true
|
|
||||||
license.workspace = true
|
|
||||||
edition.workspace = true
|
|
||||||
repository.workspace = true
|
|
||||||
rust-version.workspace = true
|
|
||||||
|
|
||||||
[dependencies]
|
|
||||||
anyhow = { workspace = true }
|
|
||||||
futures = { workspace = true }
|
|
||||||
tracing = { workspace = true }
|
|
||||||
once_cell = { workspace = true }
|
|
||||||
tokio = { workspace = true }
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
mod blocking_executor;
|
|
||||||
|
|
||||||
pub use blocking_executor::*;
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
mod files_with_extension_iterator;
|
|
||||||
|
|
||||||
pub use files_with_extension_iterator::*;
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
//! This crate provides common concepts, functionality, types, macros, and more that other crates in
|
|
||||||
//! the workspace can benefit from.
|
|
||||||
|
|
||||||
pub mod concepts;
|
|
||||||
pub mod iterators;
|
|
||||||
pub mod macros;
|
|
||||||
@@ -1,3 +0,0 @@
|
|||||||
mod define_wrapper_type;
|
|
||||||
|
|
||||||
pub use define_wrapper_type::*;
|
|
||||||
@@ -9,13 +9,11 @@ repository.workspace = true
|
|||||||
rust-version.workspace = true
|
rust-version.workspace = true
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
anyhow = { workspace = true }
|
||||||
revive-solc-json-interface = { workspace = true }
|
revive-solc-json-interface = { workspace = true }
|
||||||
revive-dt-config = { workspace = true }
|
revive-dt-config = { workspace = true }
|
||||||
revive-dt-solc-binaries = { workspace = true }
|
revive-dt-solc-binaries = { workspace = true }
|
||||||
revive-common = { workspace = true }
|
revive-common = { workspace = true }
|
||||||
|
|
||||||
alloy-primitives = { workspace = true }
|
|
||||||
anyhow = { workspace = true }
|
|
||||||
semver = { workspace = true }
|
semver = { workspace = true }
|
||||||
serde_json = { workspace = true }
|
serde_json = { workspace = true }
|
||||||
tracing = { workspace = true }
|
tracing = { workspace = true }
|
||||||
|
|||||||
@@ -9,7 +9,6 @@ use std::{
|
|||||||
path::{Path, PathBuf},
|
path::{Path, PathBuf},
|
||||||
};
|
};
|
||||||
|
|
||||||
use alloy_primitives::Address;
|
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
|
|
||||||
use revive_common::EVMVersion;
|
use revive_common::EVMVersion;
|
||||||
@@ -159,26 +158,6 @@ where
|
|||||||
self
|
self
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn with_library(
|
|
||||||
mut self,
|
|
||||||
scope: impl AsRef<Path>,
|
|
||||||
library_ident: impl AsRef<str>,
|
|
||||||
library_address: Address,
|
|
||||||
) -> Self {
|
|
||||||
self.input
|
|
||||||
.settings
|
|
||||||
.libraries
|
|
||||||
.get_or_insert_with(Default::default)
|
|
||||||
.entry(scope.as_ref().display().to_string())
|
|
||||||
.or_default()
|
|
||||||
.insert(
|
|
||||||
library_ident.as_ref().to_owned(),
|
|
||||||
library_address.to_string(),
|
|
||||||
);
|
|
||||||
|
|
||||||
self
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn try_build(self, solc_path: PathBuf) -> anyhow::Result<CompilerOutput<T::Options>> {
|
pub fn try_build(self, solc_path: PathBuf) -> anyhow::Result<CompilerOutput<T::Options>> {
|
||||||
T::new(solc_path).build(CompilerInput {
|
T::new(solc_path).build(CompilerInput {
|
||||||
extra_options: self.extra_options,
|
extra_options: self.extra_options,
|
||||||
|
|||||||
@@ -10,10 +10,6 @@ use crate::{CompilerInput, CompilerOutput, SolidityCompiler};
|
|||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_solc_json_interface::SolcStandardJsonOutput;
|
use revive_solc_json_interface::SolcStandardJsonOutput;
|
||||||
|
|
||||||
// TODO: I believe that we need to also pass the solc compiler to resolc so that resolc uses the
|
|
||||||
// specified solc compiler. I believe that currently we completely ignore the specified solc binary
|
|
||||||
// when invoking resolc which doesn't seem right if we're using solc as a compiler frontend.
|
|
||||||
|
|
||||||
/// A wrapper around the `resolc` binary, emitting PVM-compatible bytecode.
|
/// A wrapper around the `resolc` binary, emitting PVM-compatible bytecode.
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
pub struct Resolc {
|
pub struct Resolc {
|
||||||
|
|||||||
@@ -73,12 +73,6 @@ pub struct Arguments {
|
|||||||
)]
|
)]
|
||||||
pub account: String,
|
pub account: String,
|
||||||
|
|
||||||
/// This argument controls which private keys the nodes should have access to and be added to
|
|
||||||
/// its wallet signers. With a value of N, private keys (0, N] will be added to the signer set
|
|
||||||
/// of the node.
|
|
||||||
#[arg(long = "private-keys-count", default_value_t = 30)]
|
|
||||||
pub private_keys_to_add: usize,
|
|
||||||
|
|
||||||
/// The differential testing leader node implementation.
|
/// The differential testing leader node implementation.
|
||||||
#[arg(short, long = "leader", default_value = "geth")]
|
#[arg(short, long = "leader", default_value = "geth")]
|
||||||
pub leader: TestingPlatform,
|
pub leader: TestingPlatform,
|
||||||
|
|||||||
@@ -13,7 +13,6 @@ name = "retester"
|
|||||||
path = "src/main.rs"
|
path = "src/main.rs"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
revive-dt-common = { workspace = true }
|
|
||||||
revive-dt-compiler = { workspace = true }
|
revive-dt-compiler = { workspace = true }
|
||||||
revive-dt-config = { workspace = true }
|
revive-dt-config = { workspace = true }
|
||||||
revive-dt-format = { workspace = true }
|
revive-dt-format = { workspace = true }
|
||||||
|
|||||||
+164
-320
@@ -1,11 +1,11 @@
|
|||||||
//! The test driver handles the compilation and execution of the test cases.
|
//! The test driver handles the compilation and execution of the test cases.
|
||||||
|
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::fmt::Debug;
|
|
||||||
use std::marker::PhantomData;
|
use std::marker::PhantomData;
|
||||||
|
|
||||||
use alloy::json_abi::JsonAbi;
|
use alloy::json_abi::JsonAbi;
|
||||||
use alloy::network::{Ethereum, TransactionBuilder};
|
use alloy::network::{Ethereum, TransactionBuilder};
|
||||||
|
use alloy::primitives::Bytes;
|
||||||
use alloy::rpc::types::TransactionReceipt;
|
use alloy::rpc::types::TransactionReceipt;
|
||||||
use alloy::rpc::types::trace::geth::{
|
use alloy::rpc::types::trace::geth::{
|
||||||
CallFrame, GethDebugBuiltInTracerType, GethDebugTracerType, GethDebugTracingOptions, GethTrace,
|
CallFrame, GethDebugBuiltInTracerType, GethDebugTracerType, GethDebugTracingOptions, GethTrace,
|
||||||
@@ -20,21 +20,21 @@ use alloy::{
|
|||||||
};
|
};
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use indexmap::IndexMap;
|
use indexmap::IndexMap;
|
||||||
use serde_json::Value;
|
|
||||||
|
|
||||||
use revive_dt_common::iterators::FilesWithExtensionIterator;
|
|
||||||
use revive_dt_compiler::{Compiler, SolidityCompiler};
|
use revive_dt_compiler::{Compiler, SolidityCompiler};
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_dt_format::case::CaseIdx;
|
use revive_dt_format::case::CaseIdx;
|
||||||
use revive_dt_format::input::{Calldata, EtherValue, Expected, ExpectedOutput, Method};
|
use revive_dt_format::input::{Calldata, Expected, ExpectedOutput, Method};
|
||||||
use revive_dt_format::metadata::{ContractInstance, ContractPathAndIdent};
|
use revive_dt_format::metadata::{ContractInstance, ContractPathAndIdentifier};
|
||||||
use revive_dt_format::{input::Input, metadata::Metadata, mode::SolcMode};
|
use revive_dt_format::{input::Input, metadata::Metadata, mode::SolcMode};
|
||||||
use revive_dt_node::Node;
|
use revive_dt_node::Node;
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
use revive_dt_report::reporter::{CompilationTask, Report, Span};
|
use revive_dt_report::reporter::{CompilationTask, Report, Span};
|
||||||
use revive_solc_json_interface::SolcStandardJsonOutput;
|
use revive_solc_json_interface::SolcStandardJsonOutput;
|
||||||
|
use serde_json::Value;
|
||||||
|
use std::fmt::Debug;
|
||||||
|
|
||||||
use crate::Platform;
|
use crate::Platform;
|
||||||
|
use crate::common::*;
|
||||||
|
|
||||||
pub struct State<'a, T: Platform> {
|
pub struct State<'a, T: Platform> {
|
||||||
/// The configuration that the framework was started with.
|
/// The configuration that the framework was started with.
|
||||||
@@ -57,12 +57,6 @@ pub struct State<'a, T: Platform> {
|
|||||||
/// files.
|
/// files.
|
||||||
deployed_contracts: HashMap<CaseIdx, HashMap<ContractInstance, (Address, JsonAbi)>>,
|
deployed_contracts: HashMap<CaseIdx, HashMap<ContractInstance, (Address, JsonAbi)>>,
|
||||||
|
|
||||||
/// This is a map of the deployed libraries.
|
|
||||||
///
|
|
||||||
/// This map is not per case, but rather, per metadata file. This means that we do not redeploy
|
|
||||||
/// the libraries with each case.
|
|
||||||
deployed_libraries: HashMap<ContractInstance, (Address, JsonAbi)>,
|
|
||||||
|
|
||||||
phantom: PhantomData<T>,
|
phantom: PhantomData<T>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -76,7 +70,6 @@ where
|
|||||||
span,
|
span,
|
||||||
contracts: Default::default(),
|
contracts: Default::default(),
|
||||||
deployed_contracts: Default::default(),
|
deployed_contracts: Default::default(),
|
||||||
deployed_libraries: Default::default(),
|
|
||||||
phantom: Default::default(),
|
phantom: Default::default(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -103,28 +96,10 @@ where
|
|||||||
let compiler = Compiler::<T::Compiler>::new()
|
let compiler = Compiler::<T::Compiler>::new()
|
||||||
.allow_path(metadata.directory()?)
|
.allow_path(metadata.directory()?)
|
||||||
.solc_optimizer(mode.solc_optimize());
|
.solc_optimizer(mode.solc_optimize());
|
||||||
let mut compiler = FilesWithExtensionIterator::new(metadata.directory()?)
|
|
||||||
|
let compiler = FilesWithExtensionIterator::new(metadata.directory()?)
|
||||||
.with_allowed_extension("sol")
|
.with_allowed_extension("sol")
|
||||||
.try_fold(compiler, |compiler, path| compiler.with_source(&path))?;
|
.try_fold(compiler, |compiler, path| compiler.with_source(&path))?;
|
||||||
for (library_instance, (library_address, _)) in self.deployed_libraries.iter() {
|
|
||||||
let library_ident = &metadata
|
|
||||||
.contracts
|
|
||||||
.as_ref()
|
|
||||||
.and_then(|contracts| contracts.get(library_instance))
|
|
||||||
.expect("Impossible for library to not be found in contracts")
|
|
||||||
.contract_ident;
|
|
||||||
|
|
||||||
// Note the following: we need to tell solc which files require the libraries to be
|
|
||||||
// linked into them. We do not have access to this information and therefore we choose
|
|
||||||
// an easier, yet more compute intensive route, of telling solc that all of the files
|
|
||||||
// need to link the library and it will only perform the linking for the files that do
|
|
||||||
// actually need the library.
|
|
||||||
compiler = FilesWithExtensionIterator::new(metadata.directory()?)
|
|
||||||
.with_allowed_extension("sol")
|
|
||||||
.fold(compiler, |compiler, path| {
|
|
||||||
compiler.with_library(&path, library_ident.as_str(), *library_address)
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
let mut task = CompilationTask {
|
let mut task = CompilationTask {
|
||||||
json_input: compiler.input(),
|
json_input: compiler.input(),
|
||||||
@@ -166,47 +141,18 @@ where
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn build_and_publish_libraries(
|
|
||||||
&mut self,
|
|
||||||
metadata: &Metadata,
|
|
||||||
mode: &SolcMode,
|
|
||||||
node: &T::Blockchain,
|
|
||||||
) -> anyhow::Result<()> {
|
|
||||||
self.build_contracts(mode, metadata)?;
|
|
||||||
|
|
||||||
for library_instance in metadata
|
|
||||||
.libraries
|
|
||||||
.iter()
|
|
||||||
.flatten()
|
|
||||||
.flat_map(|(_, map)| map.values())
|
|
||||||
{
|
|
||||||
self.get_or_deploy_contract_instance(
|
|
||||||
library_instance,
|
|
||||||
metadata,
|
|
||||||
None,
|
|
||||||
Input::default_caller(),
|
|
||||||
None,
|
|
||||||
None,
|
|
||||||
node,
|
|
||||||
)?;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn handle_input(
|
pub fn handle_input(
|
||||||
&mut self,
|
&mut self,
|
||||||
metadata: &Metadata,
|
metadata: &Metadata,
|
||||||
case_idx: CaseIdx,
|
case_idx: CaseIdx,
|
||||||
input: &Input,
|
input: &Input,
|
||||||
node: &T::Blockchain,
|
node: &T::Blockchain,
|
||||||
mode: &SolcMode,
|
|
||||||
) -> anyhow::Result<(TransactionReceipt, GethTrace, DiffMode)> {
|
) -> anyhow::Result<(TransactionReceipt, GethTrace, DiffMode)> {
|
||||||
let deployment_receipts =
|
let deployment_receipts =
|
||||||
self.handle_contract_deployment(metadata, case_idx, input, node)?;
|
self.handle_contract_deployment(metadata, case_idx, input, node)?;
|
||||||
let execution_receipt =
|
let execution_receipt =
|
||||||
self.handle_input_execution(case_idx, input, deployment_receipts, node)?;
|
self.handle_input_execution(case_idx, input, deployment_receipts, node)?;
|
||||||
self.handle_input_expectations(case_idx, input, &execution_receipt, node, mode)?;
|
self.handle_input_expectations(case_idx, input, &execution_receipt, node)?;
|
||||||
self.handle_input_diff(case_idx, execution_receipt, node)
|
self.handle_input_diff(case_idx, execution_receipt, node)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -227,7 +173,12 @@ where
|
|||||||
|
|
||||||
let mut instances_we_must_deploy = IndexMap::<ContractInstance, bool>::new();
|
let mut instances_we_must_deploy = IndexMap::<ContractInstance, bool>::new();
|
||||||
for instance in input.find_all_contract_instances().into_iter() {
|
for instance in input.find_all_contract_instances().into_iter() {
|
||||||
if !self.deployed_contracts(case_idx).contains_key(&instance) {
|
if !self
|
||||||
|
.deployed_contracts
|
||||||
|
.entry(case_idx)
|
||||||
|
.or_default()
|
||||||
|
.contains_key(&instance)
|
||||||
|
{
|
||||||
instances_we_must_deploy.entry(instance).or_insert(false);
|
instances_we_must_deploy.entry(instance).or_insert(false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -243,22 +194,124 @@ where
|
|||||||
|
|
||||||
let mut receipts = HashMap::new();
|
let mut receipts = HashMap::new();
|
||||||
for (instance, deploy_with_constructor_arguments) in instances_we_must_deploy.into_iter() {
|
for (instance, deploy_with_constructor_arguments) in instances_we_must_deploy.into_iter() {
|
||||||
let calldata = deploy_with_constructor_arguments.then_some(&input.calldata);
|
// What we have at this moment is just a contract instance which is kind of like a variable
|
||||||
let value = deploy_with_constructor_arguments
|
// name for an actual underlying contract. So, we need to resolve this instance to the info
|
||||||
.then_some(input.value)
|
// of the contract that it belongs to.
|
||||||
.flatten();
|
let Some(ContractPathAndIdentifier {
|
||||||
|
contract_source_path,
|
||||||
|
contract_ident,
|
||||||
|
}) = metadata.contract_sources()?.remove(&instance)
|
||||||
|
else {
|
||||||
|
tracing::error!("Contract source not found for instance");
|
||||||
|
anyhow::bail!("Contract source not found for instance {:?}", instance)
|
||||||
|
};
|
||||||
|
|
||||||
if let (_, _, Some(receipt)) = self.get_or_deploy_contract_instance(
|
let compiled_contract = self.contracts.iter().find_map(|output| {
|
||||||
&instance,
|
output
|
||||||
metadata,
|
.contracts
|
||||||
case_idx,
|
.as_ref()?
|
||||||
input.caller,
|
.get(&contract_source_path.display().to_string())
|
||||||
calldata,
|
.and_then(|source_file_contracts| {
|
||||||
value,
|
source_file_contracts.get(contract_ident.as_ref())
|
||||||
node,
|
})
|
||||||
)? {
|
});
|
||||||
receipts.insert(instance.clone(), receipt);
|
let Some(code) = compiled_contract
|
||||||
|
.and_then(|contract| contract.evm.as_ref().and_then(|evm| evm.bytecode.as_ref()))
|
||||||
|
else {
|
||||||
|
tracing::error!(
|
||||||
|
contract_source_path = contract_source_path.display().to_string(),
|
||||||
|
contract_ident = contract_ident.as_ref(),
|
||||||
|
"Failed to find bytecode for contract"
|
||||||
|
);
|
||||||
|
anyhow::bail!("Failed to find bytecode for contract {:?}", instance)
|
||||||
|
};
|
||||||
|
|
||||||
|
// TODO: When we want to do linking it would be best to do it at this stage here. We have
|
||||||
|
// the context from the metadata files and therefore know what needs to be linked and in
|
||||||
|
// what order it needs to happen.
|
||||||
|
|
||||||
|
let mut code = match alloy::hex::decode(&code.object) {
|
||||||
|
Ok(code) => code,
|
||||||
|
Err(error) => {
|
||||||
|
tracing::error!(
|
||||||
|
?error,
|
||||||
|
contract_source_path = contract_source_path.display().to_string(),
|
||||||
|
contract_ident = contract_ident.as_ref(),
|
||||||
|
"Failed to hex-decode byte code - This could possibly mean that the bytecode requires linking"
|
||||||
|
);
|
||||||
|
anyhow::bail!("Failed to hex-decode the byte code {}", error)
|
||||||
}
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
if deploy_with_constructor_arguments {
|
||||||
|
let encoded_input = input
|
||||||
|
.encoded_input(self.deployed_contracts.entry(case_idx).or_default(), node)?;
|
||||||
|
code.extend(encoded_input.to_vec());
|
||||||
|
}
|
||||||
|
|
||||||
|
let tx = {
|
||||||
|
let tx = TransactionRequest::default().from(input.caller);
|
||||||
|
let tx = match input.value {
|
||||||
|
Some(ref value) if deploy_with_constructor_arguments => {
|
||||||
|
tx.value(value.into_inner())
|
||||||
|
}
|
||||||
|
_ => tx,
|
||||||
|
};
|
||||||
|
TransactionBuilder::<Ethereum>::with_deploy_code(tx, code)
|
||||||
|
};
|
||||||
|
|
||||||
|
let receipt = match node.execute_transaction(tx) {
|
||||||
|
Ok(receipt) => receipt,
|
||||||
|
Err(error) => {
|
||||||
|
tracing::error!(
|
||||||
|
node = std::any::type_name::<T>(),
|
||||||
|
?error,
|
||||||
|
"Contract deployment transaction failed."
|
||||||
|
);
|
||||||
|
return Err(error);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let Some(address) = receipt.contract_address else {
|
||||||
|
tracing::error!("Contract deployment transaction didn't return an address");
|
||||||
|
anyhow::bail!("Contract deployment didn't return an address");
|
||||||
|
};
|
||||||
|
tracing::info!(
|
||||||
|
instance_name = ?instance,
|
||||||
|
instance_address = ?address,
|
||||||
|
"Deployed contract"
|
||||||
|
);
|
||||||
|
|
||||||
|
let Some(Value::String(metadata)) =
|
||||||
|
compiled_contract.and_then(|contract| contract.metadata.as_ref())
|
||||||
|
else {
|
||||||
|
tracing::error!("Contract does not have a metadata field");
|
||||||
|
anyhow::bail!("Contract does not have a metadata field");
|
||||||
|
};
|
||||||
|
|
||||||
|
let Ok(metadata) = serde_json::from_str::<Value>(metadata) else {
|
||||||
|
tracing::error!(%metadata, "Failed to parse solc metadata into a structured value");
|
||||||
|
anyhow::bail!("Failed to parse solc metadata into a structured value {metadata}");
|
||||||
|
};
|
||||||
|
|
||||||
|
let Some(abi) = metadata.get("output").and_then(|value| value.get("abi")) else {
|
||||||
|
tracing::error!(%metadata, "Failed to access the .output.abi field of the solc metadata");
|
||||||
|
anyhow::bail!(
|
||||||
|
"Failed to access the .output.abi field of the solc metadata {metadata}"
|
||||||
|
);
|
||||||
|
};
|
||||||
|
|
||||||
|
let Ok(abi) = serde_json::from_value::<JsonAbi>(abi.clone()) else {
|
||||||
|
tracing::error!(%metadata, "Failed to deserialize ABI into a structured format");
|
||||||
|
anyhow::bail!("Failed to deserialize ABI into a structured format {metadata}");
|
||||||
|
};
|
||||||
|
|
||||||
|
self.deployed_contracts
|
||||||
|
.entry(case_idx)
|
||||||
|
.or_default()
|
||||||
|
.insert(instance.clone(), (address, abi));
|
||||||
|
|
||||||
|
receipts.insert(instance.clone(), receipt);
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(receipts)
|
Ok(receipts)
|
||||||
@@ -279,7 +332,9 @@ where
|
|||||||
.remove(&input.instance)
|
.remove(&input.instance)
|
||||||
.context("Failed to find deployment receipt"),
|
.context("Failed to find deployment receipt"),
|
||||||
Method::Fallback | Method::FunctionName(_) => {
|
Method::Fallback | Method::FunctionName(_) => {
|
||||||
let tx = match input.legacy_transaction(self.deployed_contracts(case_idx), node) {
|
let tx = match input
|
||||||
|
.legacy_transaction(self.deployed_contracts.entry(case_idx).or_default(), node)
|
||||||
|
{
|
||||||
Ok(tx) => {
|
Ok(tx) => {
|
||||||
tracing::debug!("Legacy transaction data: {tx:#?}");
|
tracing::debug!("Legacy transaction data: {tx:#?}");
|
||||||
tx
|
tx
|
||||||
@@ -313,7 +368,6 @@ where
|
|||||||
input: &Input,
|
input: &Input,
|
||||||
execution_receipt: &TransactionReceipt,
|
execution_receipt: &TransactionReceipt,
|
||||||
node: &T::Blockchain,
|
node: &T::Blockchain,
|
||||||
mode: &SolcMode,
|
|
||||||
) -> anyhow::Result<()> {
|
) -> anyhow::Result<()> {
|
||||||
let span = tracing::info_span!("Handling input expectations");
|
let span = tracing::info_span!("Handling input expectations");
|
||||||
let _guard = span.enter();
|
let _guard = span.enter();
|
||||||
@@ -369,7 +423,6 @@ where
|
|||||||
node,
|
node,
|
||||||
expectation,
|
expectation,
|
||||||
&tracing_result,
|
&tracing_result,
|
||||||
mode,
|
|
||||||
)?;
|
)?;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -383,19 +436,15 @@ where
|
|||||||
node: &T::Blockchain,
|
node: &T::Blockchain,
|
||||||
expectation: &ExpectedOutput,
|
expectation: &ExpectedOutput,
|
||||||
tracing_result: &CallFrame,
|
tracing_result: &CallFrame,
|
||||||
mode: &SolcMode,
|
|
||||||
) -> anyhow::Result<()> {
|
) -> anyhow::Result<()> {
|
||||||
if let Some(ref version_requirement) = expectation.compiler_version {
|
// TODO: We want to respect the compiler version filter on the expected output but would
|
||||||
let Some(compiler_version) = mode.last_patch_version(&self.config.solc) else {
|
// require some changes to the interfaces of the compiler and such. So, we add it later.
|
||||||
anyhow::bail!("unsupported solc version: {:?}", &mode.solc_version);
|
// Additionally, what happens if the compiler filter doesn't match? Do we consider that the
|
||||||
};
|
// transaction should succeed? Do we just ignore the expectation?
|
||||||
if !version_requirement.matches(&compiler_version) {
|
|
||||||
return Ok(());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let deployed_contracts = self.deployed_contracts(case_idx);
|
let error_span =
|
||||||
let chain_state_provider = node;
|
tracing::error_span!("Exception failed", ?tracing_result, ?execution_receipt,);
|
||||||
|
let _guard = error_span.enter();
|
||||||
|
|
||||||
// Handling the receipt state assertion.
|
// Handling the receipt state assertion.
|
||||||
let expected = !expectation.exception;
|
let expected = !expectation.exception;
|
||||||
@@ -409,16 +458,17 @@ where
|
|||||||
|
|
||||||
// Handling the calldata assertion
|
// Handling the calldata assertion
|
||||||
if let Some(ref expected_calldata) = expectation.return_data {
|
if let Some(ref expected_calldata) = expectation.return_data {
|
||||||
let expected = expected_calldata;
|
let expected = expected_calldata
|
||||||
let actual = &tracing_result.output.as_ref().unwrap_or_default();
|
.calldata(self.deployed_contracts.entry(case_idx).or_default(), node)
|
||||||
if !expected.is_equivalent(actual, deployed_contracts, chain_state_provider)? {
|
.map(Bytes::from)?;
|
||||||
|
let actual = tracing_result.output.clone().unwrap_or_default();
|
||||||
|
if !expected.starts_with(&actual) {
|
||||||
tracing::error!(
|
tracing::error!(
|
||||||
?execution_receipt,
|
%expected,
|
||||||
?expected,
|
|
||||||
%actual,
|
%actual,
|
||||||
"Calldata assertion failed"
|
"Calldata assertion failed"
|
||||||
);
|
);
|
||||||
anyhow::bail!("Calldata assertion failed - Expected {expected:?} but got {actual}",);
|
anyhow::bail!("Calldata assertion failed - Expected {expected} but got {actual}",);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -455,24 +505,17 @@ where
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Handling the topics assertion.
|
// Handling the topics assertion.
|
||||||
for (expected, actual) in expected_event
|
for (expected_topic, actual_topic) in expected_event
|
||||||
.topics
|
.topics
|
||||||
.as_slice()
|
.as_slice()
|
||||||
.iter()
|
.iter()
|
||||||
.zip(actual_event.topics())
|
.zip(actual_event.topics())
|
||||||
{
|
{
|
||||||
let expected = Calldata::new_compound([expected]);
|
let expected = Calldata::Compound(vec![expected_topic.clone()])
|
||||||
if !expected.is_equivalent(
|
.calldata(self.deployed_contracts.entry(case_idx).or_default(), node)?;
|
||||||
&actual.0,
|
let actual = actual_topic.to_vec();
|
||||||
deployed_contracts,
|
if actual != expected {
|
||||||
chain_state_provider,
|
tracing::error!(?expected, ?actual, "Event topics assertion failed",);
|
||||||
)? {
|
|
||||||
tracing::error!(
|
|
||||||
?execution_receipt,
|
|
||||||
?expected,
|
|
||||||
?actual,
|
|
||||||
"Event topics assertion failed",
|
|
||||||
);
|
|
||||||
anyhow::bail!(
|
anyhow::bail!(
|
||||||
"Event topics assertion failed - Expected {expected:?} but got {actual:?}",
|
"Event topics assertion failed - Expected {expected:?} but got {actual:?}",
|
||||||
);
|
);
|
||||||
@@ -480,15 +523,13 @@ where
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Handling the values assertion.
|
// Handling the values assertion.
|
||||||
let expected = &expected_event.values;
|
let expected = &expected_event
|
||||||
|
.values
|
||||||
|
.calldata(self.deployed_contracts.entry(case_idx).or_default(), node)
|
||||||
|
.map(Bytes::from)?;
|
||||||
let actual = &actual_event.data().data;
|
let actual = &actual_event.data().data;
|
||||||
if !expected.is_equivalent(&actual.0, deployed_contracts, chain_state_provider)? {
|
if !expected.starts_with(actual) {
|
||||||
tracing::error!(
|
tracing::error!(?expected, ?actual, "Event value assertion failed",);
|
||||||
?execution_receipt,
|
|
||||||
?expected,
|
|
||||||
?actual,
|
|
||||||
"Event value assertion failed",
|
|
||||||
);
|
|
||||||
anyhow::bail!(
|
anyhow::bail!(
|
||||||
"Event value assertion failed - Expected {expected:?} but got {actual:?}",
|
"Event value assertion failed - Expected {expected:?} but got {actual:?}",
|
||||||
);
|
);
|
||||||
@@ -519,162 +560,6 @@ where
|
|||||||
|
|
||||||
Ok((execution_receipt, trace, diff))
|
Ok((execution_receipt, trace, diff))
|
||||||
}
|
}
|
||||||
|
|
||||||
fn deployed_contracts(
|
|
||||||
&mut self,
|
|
||||||
case_idx: impl Into<Option<CaseIdx>>,
|
|
||||||
) -> &mut HashMap<ContractInstance, (Address, JsonAbi)> {
|
|
||||||
match case_idx.into() {
|
|
||||||
Some(case_idx) => self
|
|
||||||
.deployed_contracts
|
|
||||||
.entry(case_idx)
|
|
||||||
.or_insert_with(|| self.deployed_libraries.clone()),
|
|
||||||
None => &mut self.deployed_libraries,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Gets the information of a deployed contract or library from the state. If it's found to not
|
|
||||||
/// be deployed then it will be deployed.
|
|
||||||
///
|
|
||||||
/// If a [`CaseIdx`] is not specified then this contact instance address will be stored in the
|
|
||||||
/// cross-case deployed contracts address mapping.
|
|
||||||
#[allow(clippy::too_many_arguments)]
|
|
||||||
pub fn get_or_deploy_contract_instance(
|
|
||||||
&mut self,
|
|
||||||
contract_instance: &ContractInstance,
|
|
||||||
metadata: &Metadata,
|
|
||||||
case_idx: impl Into<Option<CaseIdx>>,
|
|
||||||
deployer: Address,
|
|
||||||
calldata: Option<&Calldata>,
|
|
||||||
value: Option<EtherValue>,
|
|
||||||
node: &T::Blockchain,
|
|
||||||
) -> anyhow::Result<(Address, JsonAbi, Option<TransactionReceipt>)> {
|
|
||||||
let case_idx = case_idx.into();
|
|
||||||
|
|
||||||
if let Some((address, abi)) = self.deployed_libraries.get(contract_instance) {
|
|
||||||
return Ok((*address, abi.clone(), None));
|
|
||||||
}
|
|
||||||
if let Some(case_idx) = case_idx {
|
|
||||||
if let Some((address, abi)) = self
|
|
||||||
.deployed_contracts
|
|
||||||
.get(&case_idx)
|
|
||||||
.and_then(|contracts| contracts.get(contract_instance))
|
|
||||||
{
|
|
||||||
return Ok((*address, abi.clone(), None));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let Some(ContractPathAndIdent {
|
|
||||||
contract_source_path,
|
|
||||||
contract_ident,
|
|
||||||
}) = metadata.contract_sources()?.remove(contract_instance)
|
|
||||||
else {
|
|
||||||
tracing::error!("Contract source not found for instance");
|
|
||||||
anyhow::bail!(
|
|
||||||
"Contract source not found for instance {:?}",
|
|
||||||
contract_instance
|
|
||||||
)
|
|
||||||
};
|
|
||||||
|
|
||||||
let compiled_contract = self.contracts.iter().rev().find_map(|output| {
|
|
||||||
output
|
|
||||||
.contracts
|
|
||||||
.as_ref()?
|
|
||||||
.get(&contract_source_path.display().to_string())
|
|
||||||
.and_then(|source_file_contracts| {
|
|
||||||
source_file_contracts.get(contract_ident.as_ref())
|
|
||||||
})
|
|
||||||
});
|
|
||||||
let Some(code) = compiled_contract
|
|
||||||
.and_then(|contract| contract.evm.as_ref().and_then(|evm| evm.bytecode.as_ref()))
|
|
||||||
else {
|
|
||||||
tracing::error!(
|
|
||||||
contract_source_path = contract_source_path.display().to_string(),
|
|
||||||
contract_ident = contract_ident.as_ref(),
|
|
||||||
"Failed to find bytecode for contract"
|
|
||||||
);
|
|
||||||
anyhow::bail!(
|
|
||||||
"Failed to find bytecode for contract {:?}",
|
|
||||||
contract_instance
|
|
||||||
)
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut code = match alloy::hex::decode(&code.object) {
|
|
||||||
Ok(code) => code,
|
|
||||||
Err(error) => {
|
|
||||||
tracing::error!(
|
|
||||||
?error,
|
|
||||||
contract_source_path = contract_source_path.display().to_string(),
|
|
||||||
contract_ident = contract_ident.as_ref(),
|
|
||||||
"Failed to hex-decode byte code - This could possibly mean that the bytecode requires linking"
|
|
||||||
);
|
|
||||||
anyhow::bail!("Failed to hex-decode the byte code {}", error)
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let Some(Value::String(metadata)) =
|
|
||||||
compiled_contract.and_then(|contract| contract.metadata.as_ref())
|
|
||||||
else {
|
|
||||||
tracing::error!("Contract does not have a metadata field");
|
|
||||||
anyhow::bail!("Contract does not have a metadata field");
|
|
||||||
};
|
|
||||||
|
|
||||||
let Ok(metadata) = serde_json::from_str::<Value>(metadata) else {
|
|
||||||
tracing::error!(%metadata, "Failed to parse solc metadata into a structured value");
|
|
||||||
anyhow::bail!("Failed to parse solc metadata into a structured value {metadata}");
|
|
||||||
};
|
|
||||||
|
|
||||||
let Some(abi) = metadata.get("output").and_then(|value| value.get("abi")) else {
|
|
||||||
tracing::error!(%metadata, "Failed to access the .output.abi field of the solc metadata");
|
|
||||||
anyhow::bail!("Failed to access the .output.abi field of the solc metadata {metadata}");
|
|
||||||
};
|
|
||||||
|
|
||||||
let Ok(abi) = serde_json::from_value::<JsonAbi>(abi.clone()) else {
|
|
||||||
tracing::error!(%metadata, "Failed to deserialize ABI into a structured format");
|
|
||||||
anyhow::bail!("Failed to deserialize ABI into a structured format {metadata}");
|
|
||||||
};
|
|
||||||
|
|
||||||
if let Some(calldata) = calldata {
|
|
||||||
let calldata = calldata.calldata(self.deployed_contracts(case_idx), node)?;
|
|
||||||
code.extend(calldata);
|
|
||||||
}
|
|
||||||
|
|
||||||
let tx = {
|
|
||||||
let tx = TransactionRequest::default().from(deployer);
|
|
||||||
let tx = match value {
|
|
||||||
Some(ref value) => tx.value(value.into_inner()),
|
|
||||||
_ => tx,
|
|
||||||
};
|
|
||||||
TransactionBuilder::<Ethereum>::with_deploy_code(tx, code)
|
|
||||||
};
|
|
||||||
|
|
||||||
let receipt = match node.execute_transaction(tx) {
|
|
||||||
Ok(receipt) => receipt,
|
|
||||||
Err(error) => {
|
|
||||||
tracing::error!(
|
|
||||||
node = std::any::type_name::<T>(),
|
|
||||||
?error,
|
|
||||||
"Contract deployment transaction failed."
|
|
||||||
);
|
|
||||||
return Err(error);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
let Some(address) = receipt.contract_address else {
|
|
||||||
tracing::error!("Contract deployment transaction didn't return an address");
|
|
||||||
anyhow::bail!("Contract deployment didn't return an address");
|
|
||||||
};
|
|
||||||
tracing::info!(
|
|
||||||
instance_name = ?contract_instance,
|
|
||||||
instance_address = ?address,
|
|
||||||
"Deployed contract"
|
|
||||||
);
|
|
||||||
|
|
||||||
self.deployed_contracts(case_idx)
|
|
||||||
.insert(contract_instance.clone(), (address, abi.clone()));
|
|
||||||
|
|
||||||
Ok((address, abi, Some(receipt)))
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct Driver<'a, Leader: Platform, Follower: Platform> {
|
pub struct Driver<'a, Leader: Platform, Follower: Platform> {
|
||||||
@@ -781,42 +666,6 @@ where
|
|||||||
let mut leader_state = State::<L>::new(self.config, span);
|
let mut leader_state = State::<L>::new(self.config, span);
|
||||||
let mut follower_state = State::<F>::new(self.config, span);
|
let mut follower_state = State::<F>::new(self.config, span);
|
||||||
|
|
||||||
// Note: we are currently forced to do two compilation passes due to linking. In the
|
|
||||||
// first compilation pass we compile the libraries and publish them to the chain. In the
|
|
||||||
// second compilation pass we compile the contracts with the library addresses so that
|
|
||||||
// they're linked at compile-time.
|
|
||||||
let build_result = tracing::info_span!("Building and publishing libraries")
|
|
||||||
.in_scope(|| {
|
|
||||||
match leader_state.build_and_publish_libraries(self.metadata, &mode, self.leader_node) {
|
|
||||||
Ok(_) => {
|
|
||||||
tracing::debug!(target = ?Target::Leader, "Library building succeeded");
|
|
||||||
execution_result.add_successful_build(Target::Leader, mode.clone());
|
|
||||||
},
|
|
||||||
Err(error) => {
|
|
||||||
tracing::error!(target = ?Target::Leader, ?error, "Library building failed");
|
|
||||||
execution_result.add_failed_build(Target::Leader, mode.clone(), error);
|
|
||||||
return Err(());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
match follower_state.build_and_publish_libraries(self.metadata, &mode, self.follower_node) {
|
|
||||||
Ok(_) => {
|
|
||||||
tracing::debug!(target = ?Target::Follower, "Library building succeeded");
|
|
||||||
execution_result.add_successful_build(Target::Follower, mode.clone());
|
|
||||||
},
|
|
||||||
Err(error) => {
|
|
||||||
tracing::error!(target = ?Target::Follower, ?error, "Library building failed");
|
|
||||||
execution_result.add_failed_build(Target::Follower, mode.clone(), error);
|
|
||||||
return Err(());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
Ok(())
|
|
||||||
});
|
|
||||||
if build_result.is_err() {
|
|
||||||
// Note: We skip to the next solc mode as there's nothing that we can do at this
|
|
||||||
// point, the building has failed. We do NOT bail out of the execution as a whole.
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
|
|
||||||
// We build the contracts. If building the contracts for the metadata file fails then we
|
// We build the contracts. If building the contracts for the metadata file fails then we
|
||||||
// have no other option but to keep note of this error and move on to the next solc mode
|
// have no other option but to keep note of this error and move on to the next solc mode
|
||||||
// and NOT just bail out of the execution as a whole.
|
// and NOT just bail out of the execution as a whole.
|
||||||
@@ -853,6 +702,7 @@ where
|
|||||||
|
|
||||||
// For cases if one of the inputs fail then we move on to the next case and we do NOT
|
// For cases if one of the inputs fail then we move on to the next case and we do NOT
|
||||||
// bail out of the whole thing.
|
// bail out of the whole thing.
|
||||||
|
|
||||||
'case_loop: for (case_idx, case) in self.metadata.cases.iter().enumerate() {
|
'case_loop: for (case_idx, case) in self.metadata.cases.iter().enumerate() {
|
||||||
let tracing_span = tracing::info_span!(
|
let tracing_span = tracing::info_span!(
|
||||||
"Handling case",
|
"Handling case",
|
||||||
@@ -861,7 +711,7 @@ where
|
|||||||
);
|
);
|
||||||
let _guard = tracing_span.enter();
|
let _guard = tracing_span.enter();
|
||||||
|
|
||||||
let case_idx = CaseIdx::new(case_idx);
|
let case_idx = CaseIdx::new_from(case_idx);
|
||||||
|
|
||||||
// For inputs if one of the inputs fail we move on to the next case (we do not move
|
// For inputs if one of the inputs fail we move on to the next case (we do not move
|
||||||
// on to the next input as it doesn't make sense. It depends on the previous one).
|
// on to the next input as it doesn't make sense. It depends on the previous one).
|
||||||
@@ -873,13 +723,8 @@ where
|
|||||||
tracing::info_span!("Executing input", contract_name = ?input.instance)
|
tracing::info_span!("Executing input", contract_name = ?input.instance)
|
||||||
.in_scope(|| {
|
.in_scope(|| {
|
||||||
let (leader_receipt, _, leader_diff) = match leader_state
|
let (leader_receipt, _, leader_diff) = match leader_state
|
||||||
.handle_input(
|
.handle_input(self.metadata, case_idx, &input, self.leader_node)
|
||||||
self.metadata,
|
{
|
||||||
case_idx,
|
|
||||||
&input,
|
|
||||||
self.leader_node,
|
|
||||||
&mode,
|
|
||||||
) {
|
|
||||||
Ok(result) => result,
|
Ok(result) => result,
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
tracing::error!(
|
tracing::error!(
|
||||||
@@ -908,7 +753,6 @@ where
|
|||||||
case_idx,
|
case_idx,
|
||||||
&input,
|
&input,
|
||||||
self.follower_node,
|
self.follower_node,
|
||||||
&mode,
|
|
||||||
) {
|
) {
|
||||||
Ok(result) => result,
|
Ok(result) => result,
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
|
|||||||
@@ -5,17 +5,17 @@
|
|||||||
|
|
||||||
use revive_dt_compiler::{SolidityCompiler, revive_resolc, solc};
|
use revive_dt_compiler::{SolidityCompiler, revive_resolc, solc};
|
||||||
use revive_dt_config::TestingPlatform;
|
use revive_dt_config::TestingPlatform;
|
||||||
use revive_dt_format::traits::ResolverApi;
|
|
||||||
use revive_dt_node::{Node, geth, kitchensink::KitchensinkNode};
|
use revive_dt_node::{Node, geth, kitchensink::KitchensinkNode};
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
|
|
||||||
|
pub mod common;
|
||||||
pub mod driver;
|
pub mod driver;
|
||||||
|
|
||||||
/// One platform can be tested differentially against another.
|
/// One platform can be tested differentially against another.
|
||||||
///
|
///
|
||||||
/// For this we need a blockchain node implementation and a compiler.
|
/// For this we need a blockchain node implementation and a compiler.
|
||||||
pub trait Platform {
|
pub trait Platform {
|
||||||
type Blockchain: EthereumNode + Node + ResolverApi;
|
type Blockchain: EthereumNode + Node;
|
||||||
type Compiler: SolidityCompiler;
|
type Compiler: SolidityCompiler;
|
||||||
|
|
||||||
/// Returns the matching [TestingPlatform] of the [revive_dt_config::Arguments].
|
/// Returns the matching [TestingPlatform] of the [revive_dt_config::Arguments].
|
||||||
|
|||||||
+76
-10
@@ -1,5 +1,13 @@
|
|||||||
use std::{collections::HashMap, sync::LazyLock};
|
use std::{
|
||||||
|
collections::{HashMap, HashSet},
|
||||||
|
sync::LazyLock,
|
||||||
|
};
|
||||||
|
|
||||||
|
use alloy::{
|
||||||
|
network::TxSigner,
|
||||||
|
primitives::FixedBytes,
|
||||||
|
signers::{Signature, local::PrivateKeySigner},
|
||||||
|
};
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use rayon::{ThreadPoolBuilder, prelude::*};
|
use rayon::{ThreadPoolBuilder, prelude::*};
|
||||||
|
|
||||||
@@ -8,7 +16,11 @@ use revive_dt_core::{
|
|||||||
Geth, Kitchensink, Platform,
|
Geth, Kitchensink, Platform,
|
||||||
driver::{Driver, State},
|
driver::{Driver, State},
|
||||||
};
|
};
|
||||||
use revive_dt_format::{corpus::Corpus, metadata::MetadataFile};
|
use revive_dt_format::{
|
||||||
|
corpus::Corpus,
|
||||||
|
input::default_caller,
|
||||||
|
metadata::{AddressReplacementMap, MetadataFile},
|
||||||
|
};
|
||||||
use revive_dt_node::pool::NodePool;
|
use revive_dt_node::pool::NodePool;
|
||||||
use revive_dt_report::reporter::{Report, Span};
|
use revive_dt_report::reporter::{Report, Span};
|
||||||
use temp_dir::TempDir;
|
use temp_dir::TempDir;
|
||||||
@@ -20,12 +32,48 @@ static TEMP_DIR: LazyLock<TempDir> = LazyLock::new(|| TempDir::new().unwrap());
|
|||||||
fn main() -> anyhow::Result<()> {
|
fn main() -> anyhow::Result<()> {
|
||||||
let args = init_cli()?;
|
let args = init_cli()?;
|
||||||
|
|
||||||
for (corpus, tests) in collect_corpora(&args)? {
|
let mut corpora = collect_corpora(&args)?;
|
||||||
|
let mut replacement_private_keys = HashSet::<FixedBytes<32>>::new();
|
||||||
|
for case in corpora
|
||||||
|
.values_mut()
|
||||||
|
.flat_map(|metadata| metadata.iter_mut())
|
||||||
|
.flat_map(|metadata| metadata.content.cases.iter_mut())
|
||||||
|
{
|
||||||
|
let mut replacement_map = AddressReplacementMap::new();
|
||||||
|
for address in case.inputs.iter().filter_map(|input| {
|
||||||
|
if input.caller != default_caller() {
|
||||||
|
Some(input.caller)
|
||||||
|
} else {
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}) {
|
||||||
|
replacement_map.add(address);
|
||||||
|
}
|
||||||
|
case.handle_address_replacement(&mut replacement_map)?;
|
||||||
|
replacement_private_keys.extend(
|
||||||
|
replacement_map
|
||||||
|
.into_inner()
|
||||||
|
.into_values()
|
||||||
|
.map(|(sk, _)| sk)
|
||||||
|
.map(|sk| sk.to_bytes()),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
for (corpus, tests) in corpora {
|
||||||
let span = Span::new(corpus, args.clone())?;
|
let span = Span::new(corpus, args.clone())?;
|
||||||
|
|
||||||
match &args.compile_only {
|
match &args.compile_only {
|
||||||
Some(platform) => compile_corpus(&args, &tests, platform, span),
|
Some(platform) => compile_corpus(&args, &tests, platform, span),
|
||||||
None => execute_corpus(&args, &tests, span)?,
|
None => execute_corpus(
|
||||||
|
&args,
|
||||||
|
&tests,
|
||||||
|
replacement_private_keys
|
||||||
|
.clone()
|
||||||
|
.into_iter()
|
||||||
|
.map(|bytes| PrivateKeySigner::from_bytes(&bytes).expect("Can't fail"))
|
||||||
|
.collect::<Vec<_>>(),
|
||||||
|
span,
|
||||||
|
)?,
|
||||||
}
|
}
|
||||||
|
|
||||||
Report::save()?;
|
Report::save()?;
|
||||||
@@ -83,15 +131,24 @@ fn collect_corpora(args: &Arguments) -> anyhow::Result<HashMap<Corpus, Vec<Metad
|
|||||||
Ok(corpora)
|
Ok(corpora)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn run_driver<L, F>(args: &Arguments, tests: &[MetadataFile], span: Span) -> anyhow::Result<()>
|
fn run_driver<L, F>(
|
||||||
|
args: &Arguments,
|
||||||
|
tests: &[MetadataFile],
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>
|
||||||
|
+ Clone
|
||||||
|
+ Send
|
||||||
|
+ Sync
|
||||||
|
+ 'static,
|
||||||
|
span: Span,
|
||||||
|
) -> anyhow::Result<()>
|
||||||
where
|
where
|
||||||
L: Platform,
|
L: Platform,
|
||||||
F: Platform,
|
F: Platform,
|
||||||
L::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
L::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
||||||
F::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
F::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
||||||
{
|
{
|
||||||
let leader_nodes = NodePool::<L::Blockchain>::new(args)?;
|
let leader_nodes = NodePool::<L::Blockchain>::new(args, additional_signers.clone())?;
|
||||||
let follower_nodes = NodePool::<F::Blockchain>::new(args)?;
|
let follower_nodes = NodePool::<F::Blockchain>::new(args, additional_signers)?;
|
||||||
|
|
||||||
tests.par_iter().for_each(
|
tests.par_iter().for_each(
|
||||||
|MetadataFile {
|
|MetadataFile {
|
||||||
@@ -141,13 +198,22 @@ where
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn execute_corpus(args: &Arguments, tests: &[MetadataFile], span: Span) -> anyhow::Result<()> {
|
fn execute_corpus(
|
||||||
|
args: &Arguments,
|
||||||
|
tests: &[MetadataFile],
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>
|
||||||
|
+ Clone
|
||||||
|
+ Send
|
||||||
|
+ Sync
|
||||||
|
+ 'static,
|
||||||
|
span: Span,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
match (&args.leader, &args.follower) {
|
match (&args.leader, &args.follower) {
|
||||||
(TestingPlatform::Geth, TestingPlatform::Kitchensink) => {
|
(TestingPlatform::Geth, TestingPlatform::Kitchensink) => {
|
||||||
run_driver::<Geth, Kitchensink>(args, tests, span)?
|
run_driver::<Geth, Kitchensink>(args, tests, additional_signers, span)?
|
||||||
}
|
}
|
||||||
(TestingPlatform::Geth, TestingPlatform::Geth) => {
|
(TestingPlatform::Geth, TestingPlatform::Geth) => {
|
||||||
run_driver::<Geth, Geth>(args, tests, span)?
|
run_driver::<Geth, Geth>(args, tests, additional_signers, span)?
|
||||||
}
|
}
|
||||||
_ => unimplemented!(),
|
_ => unimplemented!(),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ repository.workspace = true
|
|||||||
rust-version.workspace = true
|
rust-version.workspace = true
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
revive-dt-common = { workspace = true }
|
revive-dt-node-interaction = { workspace = true }
|
||||||
|
|
||||||
alloy = { workspace = true }
|
alloy = { workspace = true }
|
||||||
alloy-primitives = { workspace = true }
|
alloy-primitives = { workspace = true }
|
||||||
|
|||||||
@@ -1,9 +1,9 @@
|
|||||||
use serde::Deserialize;
|
use serde::Deserialize;
|
||||||
|
|
||||||
use revive_dt_common::macros::define_wrapper_type;
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
|
define_wrapper_type,
|
||||||
input::{Expected, Input},
|
input::{Expected, Input},
|
||||||
|
metadata::AddressReplacementMap,
|
||||||
mode::Mode,
|
mode::Mode,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -41,10 +41,23 @@ impl Case {
|
|||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn handle_address_replacement(
|
||||||
|
&mut self,
|
||||||
|
old_to_new_mapping: &mut AddressReplacementMap,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
for input in self.inputs.iter_mut() {
|
||||||
|
input.handle_address_replacement(old_to_new_mapping)?;
|
||||||
|
}
|
||||||
|
if let Some(ref mut expected) = self.expected {
|
||||||
|
expected.handle_address_replacement(old_to_new_mapping)?;
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
define_wrapper_type!(
|
define_wrapper_type!(
|
||||||
/// A wrapper type for the index of test cases found in metadata file.
|
/// A wrapper type for the index of test cases found in metadata file.
|
||||||
#[derive(Clone, Copy, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)]
|
#[derive(Clone, Copy, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)]
|
||||||
pub struct CaseIdx(usize);
|
CaseIdx(usize);
|
||||||
);
|
);
|
||||||
|
|||||||
+305
-503
File diff suppressed because it is too large
Load Diff
@@ -3,6 +3,6 @@
|
|||||||
pub mod case;
|
pub mod case;
|
||||||
pub mod corpus;
|
pub mod corpus;
|
||||||
pub mod input;
|
pub mod input;
|
||||||
|
pub mod macros;
|
||||||
pub mod metadata;
|
pub mod metadata;
|
||||||
pub mod mode;
|
pub mod mode;
|
||||||
pub mod traits;
|
|
||||||
|
|||||||
@@ -12,9 +12,11 @@
|
|||||||
/// pub struct CaseId(usize);
|
/// pub struct CaseId(usize);
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// And would also implement a number of methods on this type making it easier to use.
|
/// And would also implement a number of methods on this type making it easier
|
||||||
|
/// to use.
|
||||||
///
|
///
|
||||||
/// These wrapper types become very useful as they make the code a lot easier to read.
|
/// These wrapper types become very useful as they make the code a lot easier
|
||||||
|
/// to read.
|
||||||
///
|
///
|
||||||
/// Take the following as an example:
|
/// Take the following as an example:
|
||||||
///
|
///
|
||||||
@@ -24,31 +26,33 @@
|
|||||||
/// }
|
/// }
|
||||||
/// ```
|
/// ```
|
||||||
///
|
///
|
||||||
/// In the above code it's hard to understand what the various types refer to or what to expect them
|
/// In the above code it's hard to understand what the various types refer to or
|
||||||
/// to contain.
|
/// what to expect them to contain.
|
||||||
///
|
///
|
||||||
/// With these wrapper types we're able to create code that's self-documenting in that the types
|
/// With these wrapper types we're able to create code that's self-documenting
|
||||||
/// tell us what the code is referring to. The above code is transformed into
|
/// in that the types tell us what the code is referring to. The above code is
|
||||||
|
/// transformed into
|
||||||
///
|
///
|
||||||
/// ```rust,ignore
|
/// ```rust,ignore
|
||||||
/// struct State {
|
/// struct State {
|
||||||
/// contracts: HashMap<CaseId, HashMap<ContractName, ContractByteCode>>
|
/// contracts: HashMap<CaseId, HashMap<ContractName, ContractByteCode>>
|
||||||
/// }
|
/// }
|
||||||
/// ```
|
/// ```
|
||||||
///
|
|
||||||
/// Note that we follow the same syntax for defining wrapper structs but we do not permit the use of
|
|
||||||
/// generics.
|
|
||||||
#[macro_export]
|
#[macro_export]
|
||||||
macro_rules! define_wrapper_type {
|
macro_rules! define_wrapper_type {
|
||||||
(
|
(
|
||||||
$(#[$meta: meta])*
|
$(#[$meta: meta])*
|
||||||
$vis:vis struct $ident: ident($ty: ty);
|
$ident: ident($ty: ty) $(;)?
|
||||||
) => {
|
) => {
|
||||||
$(#[$meta])*
|
$(#[$meta])*
|
||||||
$vis struct $ident($ty);
|
pub struct $ident($ty);
|
||||||
|
|
||||||
impl $ident {
|
impl $ident {
|
||||||
pub fn new(value: impl Into<$ty>) -> Self {
|
pub fn new(value: $ty) -> Self {
|
||||||
|
Self(value)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn new_from<T: Into<$ty>>(value: T) -> Self {
|
||||||
Self(value.into())
|
Self(value.into())
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -100,7 +104,3 @@ macro_rules! define_wrapper_type {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Technically not needed but this allows for the macro to be found in the `macros` module of the
|
|
||||||
/// crate in addition to being found in the root of the crate.
|
|
||||||
pub use define_wrapper_type;
|
|
||||||
+159
-26
@@ -1,5 +1,5 @@
|
|||||||
use std::{
|
use std::{
|
||||||
collections::BTreeMap,
|
collections::{BTreeMap, HashMap},
|
||||||
fmt::Display,
|
fmt::Display,
|
||||||
fs::{File, read_to_string},
|
fs::{File, read_to_string},
|
||||||
ops::Deref,
|
ops::Deref,
|
||||||
@@ -7,12 +7,15 @@ use std::{
|
|||||||
str::FromStr,
|
str::FromStr,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use alloy::signers::local::PrivateKeySigner;
|
||||||
|
use alloy_primitives::Address;
|
||||||
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use revive_dt_common::macros::define_wrapper_type;
|
|
||||||
|
|
||||||
use crate::{
|
use crate::{
|
||||||
case::Case,
|
case::Case,
|
||||||
|
define_wrapper_type,
|
||||||
|
input::resolve_argument,
|
||||||
mode::{Mode, SolcMode},
|
mode::{Mode, SolcMode},
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -47,9 +50,9 @@ impl Deref for MetadataFile {
|
|||||||
pub struct Metadata {
|
pub struct Metadata {
|
||||||
pub targets: Option<Vec<String>>,
|
pub targets: Option<Vec<String>>,
|
||||||
pub cases: Vec<Case>,
|
pub cases: Vec<Case>,
|
||||||
pub contracts: Option<BTreeMap<ContractInstance, ContractPathAndIdent>>,
|
pub contracts: Option<BTreeMap<ContractInstance, ContractPathAndIdentifier>>,
|
||||||
// TODO: Convert into wrapper types for clarity.
|
// TODO: Convert into wrapper types for clarity.
|
||||||
pub libraries: Option<BTreeMap<PathBuf, BTreeMap<ContractIdent, ContractInstance>>>,
|
pub libraries: Option<BTreeMap<String, BTreeMap<String, String>>>,
|
||||||
pub ignore: Option<bool>,
|
pub ignore: Option<bool>,
|
||||||
pub modes: Option<Vec<Mode>>,
|
pub modes: Option<Vec<Mode>>,
|
||||||
pub file_path: Option<PathBuf>,
|
pub file_path: Option<PathBuf>,
|
||||||
@@ -86,7 +89,7 @@ impl Metadata {
|
|||||||
/// Returns the contract sources with canonicalized paths for the files
|
/// Returns the contract sources with canonicalized paths for the files
|
||||||
pub fn contract_sources(
|
pub fn contract_sources(
|
||||||
&self,
|
&self,
|
||||||
) -> anyhow::Result<BTreeMap<ContractInstance, ContractPathAndIdent>> {
|
) -> anyhow::Result<BTreeMap<ContractInstance, ContractPathAndIdentifier>> {
|
||||||
let directory = self.directory()?;
|
let directory = self.directory()?;
|
||||||
let mut sources = BTreeMap::new();
|
let mut sources = BTreeMap::new();
|
||||||
let Some(contracts) = &self.contracts else {
|
let Some(contracts) = &self.contracts else {
|
||||||
@@ -95,7 +98,7 @@ impl Metadata {
|
|||||||
|
|
||||||
for (
|
for (
|
||||||
alias,
|
alias,
|
||||||
ContractPathAndIdent {
|
ContractPathAndIdentifier {
|
||||||
contract_source_path,
|
contract_source_path,
|
||||||
contract_ident,
|
contract_ident,
|
||||||
},
|
},
|
||||||
@@ -107,7 +110,7 @@ impl Metadata {
|
|||||||
|
|
||||||
sources.insert(
|
sources.insert(
|
||||||
alias,
|
alias,
|
||||||
ContractPathAndIdent {
|
ContractPathAndIdentifier {
|
||||||
contract_source_path: absolute_path,
|
contract_source_path: absolute_path,
|
||||||
contract_ident,
|
contract_ident,
|
||||||
},
|
},
|
||||||
@@ -193,10 +196,10 @@ impl Metadata {
|
|||||||
metadata.file_path = Some(path.to_path_buf());
|
metadata.file_path = Some(path.to_path_buf());
|
||||||
metadata.contracts = Some(
|
metadata.contracts = Some(
|
||||||
[(
|
[(
|
||||||
ContractInstance::new("test"),
|
ContractInstance::new_from("test"),
|
||||||
ContractPathAndIdent {
|
ContractPathAndIdentifier {
|
||||||
contract_source_path: path.to_path_buf(),
|
contract_source_path: path.to_path_buf(),
|
||||||
contract_ident: ContractIdent::new("Test"),
|
contract_ident: ContractIdent::new_from("Test"),
|
||||||
},
|
},
|
||||||
)]
|
)]
|
||||||
.into(),
|
.into(),
|
||||||
@@ -212,28 +215,35 @@ impl Metadata {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn handle_address_replacement(
|
||||||
|
&mut self,
|
||||||
|
old_to_new_mapping: &mut AddressReplacementMap,
|
||||||
|
) -> anyhow::Result<()> {
|
||||||
|
for case in self.cases.iter_mut() {
|
||||||
|
case.handle_address_replacement(old_to_new_mapping)?;
|
||||||
|
}
|
||||||
|
tracing::debug!(metadata = ?self, "Performed replacement on metadata");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
define_wrapper_type!(
|
define_wrapper_type!(
|
||||||
/// Represents a contract instance found a metadata file.
|
/// Represents a contract instance found a metadata file.
|
||||||
///
|
///
|
||||||
/// Typically, this is used as the key to the "contracts" field of metadata files.
|
/// Typically, this is used as the key to the "contracts" field of metadata files.
|
||||||
#[derive(
|
#[derive(Clone, Debug, Default, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize)]
|
||||||
Clone, Debug, Default, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize,
|
|
||||||
)]
|
|
||||||
#[serde(transparent)]
|
#[serde(transparent)]
|
||||||
pub struct ContractInstance(String);
|
ContractInstance(String);
|
||||||
);
|
);
|
||||||
|
|
||||||
define_wrapper_type!(
|
define_wrapper_type!(
|
||||||
/// Represents a contract identifier found a metadata file.
|
/// Represents a contract identifier found a metadata file.
|
||||||
///
|
///
|
||||||
/// A contract identifier is the name of the contract in the source code.
|
/// A contract identifier is the name of the contract in the source code.
|
||||||
#[derive(
|
#[derive(Clone, Debug, Default, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize)]
|
||||||
Clone, Debug, Default, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize,
|
|
||||||
)]
|
|
||||||
#[serde(transparent)]
|
#[serde(transparent)]
|
||||||
pub struct ContractIdent(String);
|
ContractIdent(String);
|
||||||
);
|
);
|
||||||
|
|
||||||
/// Represents an identifier used for contracts.
|
/// Represents an identifier used for contracts.
|
||||||
@@ -245,7 +255,7 @@ define_wrapper_type!(
|
|||||||
/// ```
|
/// ```
|
||||||
#[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize)]
|
||||||
#[serde(try_from = "String", into = "String")]
|
#[serde(try_from = "String", into = "String")]
|
||||||
pub struct ContractPathAndIdent {
|
pub struct ContractPathAndIdentifier {
|
||||||
/// The path of the contract source code relative to the directory containing the metadata file.
|
/// The path of the contract source code relative to the directory containing the metadata file.
|
||||||
pub contract_source_path: PathBuf,
|
pub contract_source_path: PathBuf,
|
||||||
|
|
||||||
@@ -253,7 +263,7 @@ pub struct ContractPathAndIdent {
|
|||||||
pub contract_ident: ContractIdent,
|
pub contract_ident: ContractIdent,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Display for ContractPathAndIdent {
|
impl Display for ContractPathAndIdentifier {
|
||||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||||
write!(
|
write!(
|
||||||
f,
|
f,
|
||||||
@@ -264,7 +274,7 @@ impl Display for ContractPathAndIdent {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl FromStr for ContractPathAndIdent {
|
impl FromStr for ContractPathAndIdentifier {
|
||||||
type Err = anyhow::Error;
|
type Err = anyhow::Error;
|
||||||
|
|
||||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
@@ -300,7 +310,7 @@ impl FromStr for ContractPathAndIdent {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl TryFrom<String> for ContractPathAndIdent {
|
impl TryFrom<String> for ContractPathAndIdentifier {
|
||||||
type Error = anyhow::Error;
|
type Error = anyhow::Error;
|
||||||
|
|
||||||
fn try_from(value: String) -> Result<Self, Self::Error> {
|
fn try_from(value: String) -> Result<Self, Self::Error> {
|
||||||
@@ -308,12 +318,135 @@ impl TryFrom<String> for ContractPathAndIdent {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl From<ContractPathAndIdent> for String {
|
impl From<ContractPathAndIdentifier> for String {
|
||||||
fn from(value: ContractPathAndIdent) -> Self {
|
fn from(value: ContractPathAndIdentifier) -> Self {
|
||||||
value.to_string()
|
value.to_string()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Debug, Default)]
|
||||||
|
pub struct AddressReplacementMap(HashMap<Address, (PrivateKeySigner, Address)>);
|
||||||
|
|
||||||
|
impl AddressReplacementMap {
|
||||||
|
pub fn new() -> Self {
|
||||||
|
Self(Default::default())
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn into_inner(self) -> HashMap<Address, (PrivateKeySigner, Address)> {
|
||||||
|
self.0
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn contains_key(&self, address: &Address) -> bool {
|
||||||
|
self.0.contains_key(address)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn add(&mut self, address: Address) -> Address {
|
||||||
|
self.0
|
||||||
|
.entry(address)
|
||||||
|
.or_insert_with(|| {
|
||||||
|
let private_key = Self::new_random_private_key_signer();
|
||||||
|
let account = private_key.address();
|
||||||
|
tracing::debug!(
|
||||||
|
old_address = %address,
|
||||||
|
new_address = %account,
|
||||||
|
"Added a new address replacement"
|
||||||
|
);
|
||||||
|
(private_key, account)
|
||||||
|
})
|
||||||
|
.1
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn resolve(&self, value: &str) -> Option<Address> {
|
||||||
|
// We attempt to resolve the given string without any additional context of the deployed
|
||||||
|
// contracts or the node API as we do not need them. If the resolution fails then we know
|
||||||
|
// that this isn't an address and we skip it.
|
||||||
|
let Ok(resolved) = resolve_argument(value, &Default::default(), &UnimplementedEthereumNode)
|
||||||
|
else {
|
||||||
|
return None;
|
||||||
|
};
|
||||||
|
let resolved_bytes = resolved.to_be_bytes_trimmed_vec();
|
||||||
|
let Ok(address) = Address::try_from(resolved_bytes.as_slice()) else {
|
||||||
|
return None;
|
||||||
|
};
|
||||||
|
self.0.get(&address).map(|(_, address)| *address)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn new_random_private_key_signer() -> PrivateKeySigner {
|
||||||
|
// TODO: Use a seedable RNG to allow for deterministic allocation of the private keys so
|
||||||
|
// that we get reproducible runs.
|
||||||
|
PrivateKeySigner::random()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl AsRef<HashMap<Address, (PrivateKeySigner, Address)>> for AddressReplacementMap {
|
||||||
|
fn as_ref(&self) -> &HashMap<Address, (PrivateKeySigner, Address)> {
|
||||||
|
&self.0
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
struct UnimplementedEthereumNode;
|
||||||
|
|
||||||
|
impl EthereumNode for UnimplementedEthereumNode {
|
||||||
|
fn execute_transaction(
|
||||||
|
&self,
|
||||||
|
_: alloy::rpc::types::TransactionRequest,
|
||||||
|
) -> anyhow::Result<alloy::rpc::types::TransactionReceipt> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn chain_id(&self) -> anyhow::Result<alloy_primitives::ChainId> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn block_gas_limit(&self, _: alloy::eips::BlockNumberOrTag) -> anyhow::Result<u128> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn block_coinbase(&self, _: alloy::eips::BlockNumberOrTag) -> anyhow::Result<Address> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn block_difficulty(
|
||||||
|
&self,
|
||||||
|
_: alloy::eips::BlockNumberOrTag,
|
||||||
|
) -> anyhow::Result<alloy_primitives::U256> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn block_hash(
|
||||||
|
&self,
|
||||||
|
_: alloy::eips::BlockNumberOrTag,
|
||||||
|
) -> anyhow::Result<alloy_primitives::BlockHash> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn block_timestamp(
|
||||||
|
&self,
|
||||||
|
_: alloy::eips::BlockNumberOrTag,
|
||||||
|
) -> anyhow::Result<alloy_primitives::BlockTimestamp> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn last_block_number(&self) -> anyhow::Result<alloy_primitives::BlockNumber> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn trace_transaction(
|
||||||
|
&self,
|
||||||
|
_: &alloy::rpc::types::TransactionReceipt,
|
||||||
|
_: alloy::rpc::types::trace::geth::GethDebugTracingOptions,
|
||||||
|
) -> anyhow::Result<alloy::rpc::types::trace::geth::GethTrace> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn state_diff(
|
||||||
|
&self,
|
||||||
|
_: &alloy::rpc::types::TransactionReceipt,
|
||||||
|
) -> anyhow::Result<alloy::rpc::types::trace::geth::DiffMode> {
|
||||||
|
anyhow::bail!("Unimplemented")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod test {
|
mod test {
|
||||||
use super::*;
|
use super::*;
|
||||||
@@ -324,7 +457,7 @@ mod test {
|
|||||||
let string = "ERC20/ERC20.sol:ERC20";
|
let string = "ERC20/ERC20.sol:ERC20";
|
||||||
|
|
||||||
// Act
|
// Act
|
||||||
let identifier = ContractPathAndIdent::from_str(string);
|
let identifier = ContractPathAndIdentifier::from_str(string);
|
||||||
|
|
||||||
// Assert
|
// Assert
|
||||||
let identifier = identifier.expect("Failed to parse");
|
let identifier = identifier.expect("Failed to parse");
|
||||||
|
|||||||
@@ -1,30 +0,0 @@
|
|||||||
use alloy::eips::BlockNumberOrTag;
|
|
||||||
use alloy::primitives::{Address, BlockHash, BlockNumber, BlockTimestamp, ChainId, U256};
|
|
||||||
use anyhow::Result;
|
|
||||||
|
|
||||||
/// A trait of the interface are required to implement to be used by the resolution logic that this
|
|
||||||
/// crate implements to go from string calldata and into the bytes calldata.
|
|
||||||
pub trait ResolverApi {
|
|
||||||
/// Returns the ID of the chain that the node is on.
|
|
||||||
fn chain_id(&self) -> Result<ChainId>;
|
|
||||||
|
|
||||||
// TODO: This is currently a u128 due to Kitchensink needing more than 64 bits for its gas limit
|
|
||||||
// when we implement the changes to the gas we need to adjust this to be a u64.
|
|
||||||
/// Returns the gas limit of the specified block.
|
|
||||||
fn block_gas_limit(&self, number: BlockNumberOrTag) -> Result<u128>;
|
|
||||||
|
|
||||||
/// Returns the coinbase of the specified block.
|
|
||||||
fn block_coinbase(&self, number: BlockNumberOrTag) -> Result<Address>;
|
|
||||||
|
|
||||||
/// Returns the difficulty of the specified block.
|
|
||||||
fn block_difficulty(&self, number: BlockNumberOrTag) -> Result<U256>;
|
|
||||||
|
|
||||||
/// Returns the hash of the specified block.
|
|
||||||
fn block_hash(&self, number: BlockNumberOrTag) -> Result<BlockHash>;
|
|
||||||
|
|
||||||
/// Returns the timestamp of the specified block,
|
|
||||||
fn block_timestamp(&self, number: BlockNumberOrTag) -> Result<BlockTimestamp>;
|
|
||||||
|
|
||||||
/// Returns the number of the last block.
|
|
||||||
fn last_block_number(&self) -> Result<BlockNumber>;
|
|
||||||
}
|
|
||||||
@@ -11,3 +11,7 @@ rust-version.workspace = true
|
|||||||
[dependencies]
|
[dependencies]
|
||||||
alloy = { workspace = true }
|
alloy = { workspace = true }
|
||||||
anyhow = { workspace = true }
|
anyhow = { workspace = true }
|
||||||
|
futures = { workspace = true }
|
||||||
|
tracing = { workspace = true }
|
||||||
|
once_cell = { workspace = true }
|
||||||
|
tokio = { workspace = true }
|
||||||
|
|||||||
+17
-14
@@ -23,7 +23,7 @@ use tracing::Instrument;
|
|||||||
/// executor to drive an async computation:
|
/// executor to drive an async computation:
|
||||||
///
|
///
|
||||||
/// ```rust
|
/// ```rust
|
||||||
/// use revive_dt_common::concepts::*;
|
/// use revive_dt_node_interaction::*;
|
||||||
///
|
///
|
||||||
/// fn blocking_function() {
|
/// fn blocking_function() {
|
||||||
/// let result = BlockingExecutor::execute(async move {
|
/// let result = BlockingExecutor::execute(async move {
|
||||||
@@ -134,17 +134,22 @@ impl BlockingExecutor {
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
let result = match result {
|
match result.map(|result| {
|
||||||
Ok(result) => result,
|
*result
|
||||||
Err(error) => {
|
|
||||||
tracing::error!(?error, "An error occurred when running the async task");
|
|
||||||
anyhow::bail!("An error occurred when running the async task: {error:?}")
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
Ok(*result
|
|
||||||
.downcast::<R>()
|
.downcast::<R>()
|
||||||
.expect("An error occurred when downcasting into R. This is a bug"))
|
.expect("Type mismatch in the downcast")
|
||||||
|
}) {
|
||||||
|
Ok(result) => Ok(result),
|
||||||
|
Err(error) => {
|
||||||
|
tracing::error!(
|
||||||
|
?error,
|
||||||
|
"Failed to downcast the returned result into the expected type"
|
||||||
|
);
|
||||||
|
anyhow::bail!(
|
||||||
|
"Failed to downcast the returned result into the expected type: {error:?}"
|
||||||
|
)
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
/// Represents the state of the async runtime. This runtime is designed to be a singleton runtime
|
/// Represents the state of the async runtime. This runtime is designed to be a singleton runtime
|
||||||
@@ -203,9 +208,7 @@ mod test {
|
|||||||
fn panics_in_futures_are_caught() {
|
fn panics_in_futures_are_caught() {
|
||||||
// Act
|
// Act
|
||||||
let result = BlockingExecutor::execute(async move {
|
let result = BlockingExecutor::execute(async move {
|
||||||
panic!(
|
panic!("This is a panic!");
|
||||||
"If this panic causes, well, a panic, then this is an issue. If it's caught then all good!"
|
|
||||||
);
|
|
||||||
0xFFu8
|
0xFFu8
|
||||||
});
|
});
|
||||||
|
|
||||||
@@ -1,9 +1,14 @@
|
|||||||
//! This crate implements all node interactions.
|
//! This crate implements all node interactions.
|
||||||
|
|
||||||
|
use alloy::eips::BlockNumberOrTag;
|
||||||
|
use alloy::primitives::{Address, BlockHash, BlockNumber, BlockTimestamp, ChainId, U256};
|
||||||
use alloy::rpc::types::trace::geth::{DiffMode, GethDebugTracingOptions, GethTrace};
|
use alloy::rpc::types::trace::geth::{DiffMode, GethDebugTracingOptions, GethTrace};
|
||||||
use alloy::rpc::types::{TransactionReceipt, TransactionRequest};
|
use alloy::rpc::types::{TransactionReceipt, TransactionRequest};
|
||||||
use anyhow::Result;
|
use anyhow::Result;
|
||||||
|
|
||||||
|
mod blocking_executor;
|
||||||
|
pub use blocking_executor::*;
|
||||||
|
|
||||||
/// An interface for all interactions with Ethereum compatible nodes.
|
/// An interface for all interactions with Ethereum compatible nodes.
|
||||||
pub trait EthereumNode {
|
pub trait EthereumNode {
|
||||||
/// Execute the [TransactionRequest] and return a [TransactionReceipt].
|
/// Execute the [TransactionRequest] and return a [TransactionReceipt].
|
||||||
@@ -18,4 +23,27 @@ pub trait EthereumNode {
|
|||||||
|
|
||||||
/// Returns the state diff of the transaction hash in the [TransactionReceipt].
|
/// Returns the state diff of the transaction hash in the [TransactionReceipt].
|
||||||
fn state_diff(&self, receipt: &TransactionReceipt) -> Result<DiffMode>;
|
fn state_diff(&self, receipt: &TransactionReceipt) -> Result<DiffMode>;
|
||||||
|
|
||||||
|
/// Returns the ID of the chain that the node is on.
|
||||||
|
fn chain_id(&self) -> Result<ChainId>;
|
||||||
|
|
||||||
|
// TODO: This is currently a u128 due to Kitchensink needing more than 64 bits for its gas limit
|
||||||
|
// when we implement the changes to the gas we need to adjust this to be a u64.
|
||||||
|
/// Returns the gas limit of the specified block.
|
||||||
|
fn block_gas_limit(&self, number: BlockNumberOrTag) -> Result<u128>;
|
||||||
|
|
||||||
|
/// Returns the coinbase of the specified block.
|
||||||
|
fn block_coinbase(&self, number: BlockNumberOrTag) -> Result<Address>;
|
||||||
|
|
||||||
|
/// Returns the difficulty of the specified block.
|
||||||
|
fn block_difficulty(&self, number: BlockNumberOrTag) -> Result<U256>;
|
||||||
|
|
||||||
|
/// Returns the hash of the specified block.
|
||||||
|
fn block_hash(&self, number: BlockNumberOrTag) -> Result<BlockHash>;
|
||||||
|
|
||||||
|
/// Returns the timestamp of the specified block,
|
||||||
|
fn block_timestamp(&self, number: BlockNumberOrTag) -> Result<BlockTimestamp>;
|
||||||
|
|
||||||
|
/// Returns the number of the last block.
|
||||||
|
fn last_block_number(&self) -> Result<BlockNumber>;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -14,10 +14,8 @@ alloy = { workspace = true }
|
|||||||
tracing = { workspace = true }
|
tracing = { workspace = true }
|
||||||
tokio = { workspace = true }
|
tokio = { workspace = true }
|
||||||
|
|
||||||
revive-dt-common = { workspace = true }
|
|
||||||
revive-dt-config = { workspace = true }
|
|
||||||
revive-dt-format = { workspace = true }
|
|
||||||
revive-dt-node-interaction = { workspace = true }
|
revive-dt-node-interaction = { workspace = true }
|
||||||
|
revive-dt-config = { workspace = true }
|
||||||
|
|
||||||
serde = { workspace = true }
|
serde = { workspace = true }
|
||||||
serde_json = { workspace = true }
|
serde_json = { workspace = true }
|
||||||
|
|||||||
@@ -1,5 +0,0 @@
|
|||||||
/// This constant defines how much Wei accounts are pre-seeded with in genesis.
|
|
||||||
///
|
|
||||||
/// Note: After changing this number, check that the tests for kitchensink work as we encountered
|
|
||||||
/// some issues with different values of the initial balance on Kitchensink.
|
|
||||||
pub const INITIAL_BALANCE: u128 = 10u128.pow(37);
|
|
||||||
+64
-58
@@ -12,8 +12,8 @@ use std::{
|
|||||||
use alloy::{
|
use alloy::{
|
||||||
eips::BlockNumberOrTag,
|
eips::BlockNumberOrTag,
|
||||||
genesis::{Genesis, GenesisAccount},
|
genesis::{Genesis, GenesisAccount},
|
||||||
network::{Ethereum, EthereumWallet, NetworkWallet},
|
network::{Ethereum, EthereumWallet, NetworkWallet, TxSigner},
|
||||||
primitives::{Address, BlockHash, BlockNumber, BlockTimestamp, FixedBytes, U256},
|
primitives::{Address, BlockHash, BlockNumber, BlockTimestamp, U256},
|
||||||
providers::{
|
providers::{
|
||||||
Provider, ProviderBuilder,
|
Provider, ProviderBuilder,
|
||||||
ext::DebugApi,
|
ext::DebugApi,
|
||||||
@@ -23,15 +23,13 @@ use alloy::{
|
|||||||
TransactionReceipt, TransactionRequest,
|
TransactionReceipt, TransactionRequest,
|
||||||
trace::geth::{DiffMode, GethDebugTracingOptions, PreStateConfig, PreStateFrame},
|
trace::geth::{DiffMode, GethDebugTracingOptions, PreStateConfig, PreStateFrame},
|
||||||
},
|
},
|
||||||
signers::local::PrivateKeySigner,
|
signers::Signature,
|
||||||
};
|
};
|
||||||
use revive_dt_common::concepts::BlockingExecutor;
|
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_dt_format::traits::ResolverApi;
|
use revive_dt_node_interaction::{BlockingExecutor, EthereumNode};
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
|
||||||
use tracing::Level;
|
use tracing::Level;
|
||||||
|
|
||||||
use crate::{Node, common::FallbackGasFiller, constants::INITIAL_BALANCE};
|
use crate::{Node, common::FallbackGasFiller};
|
||||||
|
|
||||||
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
||||||
|
|
||||||
@@ -76,8 +74,6 @@ impl Instance {
|
|||||||
const GETH_STDOUT_LOG_FILE_NAME: &str = "node_stdout.log";
|
const GETH_STDOUT_LOG_FILE_NAME: &str = "node_stdout.log";
|
||||||
const GETH_STDERR_LOG_FILE_NAME: &str = "node_stderr.log";
|
const GETH_STDERR_LOG_FILE_NAME: &str = "node_stderr.log";
|
||||||
|
|
||||||
const TRANSACTION_INDEXING_ERROR: &str = "transaction indexing is in progress";
|
|
||||||
|
|
||||||
/// Create the node directory and call `geth init` to configure the genesis.
|
/// Create the node directory and call `geth init` to configure the genesis.
|
||||||
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
||||||
fn init(&mut self, genesis: String) -> anyhow::Result<&mut Self> {
|
fn init(&mut self, genesis: String) -> anyhow::Result<&mut Self> {
|
||||||
@@ -88,12 +84,10 @@ impl Instance {
|
|||||||
for signer_address in
|
for signer_address in
|
||||||
<EthereumWallet as NetworkWallet<Ethereum>>::signer_addresses(&self.wallet)
|
<EthereumWallet as NetworkWallet<Ethereum>>::signer_addresses(&self.wallet)
|
||||||
{
|
{
|
||||||
// Note, the use of the entry API here means that we only modify the entries for any
|
genesis.alloc.entry(signer_address).or_insert(
|
||||||
// account that is not in the `alloc` field of the genesis state.
|
GenesisAccount::default()
|
||||||
genesis
|
.with_balance(10000000000000000000000u128.try_into().unwrap()),
|
||||||
.alloc
|
);
|
||||||
.entry(signer_address)
|
|
||||||
.or_insert(GenesisAccount::default().with_balance(U256::from(INITIAL_BALANCE)));
|
|
||||||
}
|
}
|
||||||
let genesis_path = self.base_directory.join(Self::GENESIS_JSON_FILE);
|
let genesis_path = self.base_directory.join(Self::GENESIS_JSON_FILE);
|
||||||
serde_json::to_writer(File::create(&genesis_path)?, &genesis)?;
|
serde_json::to_writer(File::create(&genesis_path)?, &genesis)?;
|
||||||
@@ -271,45 +265,57 @@ impl EthereumNode for Instance {
|
|||||||
// it eventually works, but we only do that if the error we get back is the "transaction
|
// it eventually works, but we only do that if the error we get back is the "transaction
|
||||||
// indexing is in progress" error or if the receipt is None.
|
// indexing is in progress" error or if the receipt is None.
|
||||||
//
|
//
|
||||||
// Getting the transaction indexed and taking a receipt can take a long time especially
|
// At the moment we do not allow for the 60 seconds to be modified and we take it as
|
||||||
// when a lot of transactions are being submitted to the node. Thus, while initially we
|
// being an implementation detail that's invisible to anything outside of this module.
|
||||||
// only allowed for 60 seconds of waiting with a 1 second delay in polling, we need to
|
//
|
||||||
// allow for a larger wait time. Therefore, in here we allow for 5 minutes of waiting
|
// We allow a total of 60 retries for getting the receipt with one second between each
|
||||||
// with exponential backoff each time we attempt to get the receipt and find that it's
|
// retry and the next which means that we allow for a total of 60 seconds of waiting
|
||||||
// not available.
|
// before we consider that we're unable to get the transaction receipt.
|
||||||
let mut retries = 0;
|
let mut retries = 0;
|
||||||
let mut total_wait_duration = Duration::from_secs(0);
|
|
||||||
let max_allowed_wait_duration = Duration::from_secs(5 * 60);
|
|
||||||
loop {
|
loop {
|
||||||
if total_wait_duration >= max_allowed_wait_duration {
|
|
||||||
tracing::error!(
|
|
||||||
?total_wait_duration,
|
|
||||||
?max_allowed_wait_duration,
|
|
||||||
retry_count = retries,
|
|
||||||
"Failed to get receipt after polling for it"
|
|
||||||
);
|
|
||||||
anyhow::bail!(
|
|
||||||
"Polled for receipt for {total_wait_duration:?} but failed to get it"
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
match provider.get_transaction_receipt(*transaction_hash).await {
|
match provider.get_transaction_receipt(*transaction_hash).await {
|
||||||
Ok(Some(receipt)) => break Ok(receipt),
|
Ok(Some(receipt)) => {
|
||||||
Ok(None) => {}
|
tracing::info!("Obtained the transaction receipt");
|
||||||
|
break Ok(receipt);
|
||||||
|
}
|
||||||
|
Ok(None) => {
|
||||||
|
if retries == 60 {
|
||||||
|
tracing::error!(
|
||||||
|
"Polled for transaction receipt for 60 seconds but failed to get it"
|
||||||
|
);
|
||||||
|
break Err(anyhow::anyhow!("Failed to get the transaction receipt"));
|
||||||
|
} else {
|
||||||
|
tracing::trace!(
|
||||||
|
retries,
|
||||||
|
"Sleeping for 1 second and trying to get the receipt again"
|
||||||
|
);
|
||||||
|
retries += 1;
|
||||||
|
tokio::time::sleep(std::time::Duration::from_secs(1)).await;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
}
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
let error_string = error.to_string();
|
let error_string = error.to_string();
|
||||||
if !error_string.contains(Self::TRANSACTION_INDEXING_ERROR) {
|
if error_string.contains("transaction indexing is in progress") {
|
||||||
|
if retries == 60 {
|
||||||
|
tracing::error!(
|
||||||
|
"Polled for transaction receipt for 60 seconds but failed to get it"
|
||||||
|
);
|
||||||
|
break Err(error.into());
|
||||||
|
} else {
|
||||||
|
tracing::trace!(
|
||||||
|
retries,
|
||||||
|
"Sleeping for 1 second and trying to get the receipt again"
|
||||||
|
);
|
||||||
|
retries += 1;
|
||||||
|
tokio::time::sleep(std::time::Duration::from_secs(1)).await;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
break Err(error.into());
|
break Err(error.into());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
}
|
||||||
|
|
||||||
let next_wait_duration = Duration::from_secs(2u64.pow(retries))
|
|
||||||
.min(max_allowed_wait_duration - total_wait_duration);
|
|
||||||
total_wait_duration += next_wait_duration;
|
|
||||||
retries += 1;
|
|
||||||
|
|
||||||
tokio::time::sleep(next_wait_duration).await;
|
|
||||||
}
|
}
|
||||||
})?
|
})?
|
||||||
}
|
}
|
||||||
@@ -345,9 +351,7 @@ impl EthereumNode for Instance {
|
|||||||
_ => anyhow::bail!("expected a diff mode trace"),
|
_ => anyhow::bail!("expected a diff mode trace"),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
impl ResolverApi for Instance {
|
|
||||||
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
||||||
fn chain_id(&self) -> anyhow::Result<alloy::primitives::ChainId> {
|
fn chain_id(&self) -> anyhow::Result<alloy::primitives::ChainId> {
|
||||||
let provider = self.provider();
|
let provider = self.provider();
|
||||||
@@ -431,17 +435,16 @@ impl ResolverApi for Instance {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Node for Instance {
|
impl Node for Instance {
|
||||||
fn new(config: &Arguments) -> Self {
|
fn new(
|
||||||
|
config: &Arguments,
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>,
|
||||||
|
) -> Self {
|
||||||
let geth_directory = config.directory().join(Self::BASE_DIRECTORY);
|
let geth_directory = config.directory().join(Self::BASE_DIRECTORY);
|
||||||
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
||||||
let base_directory = geth_directory.join(id.to_string());
|
let base_directory = geth_directory.join(id.to_string());
|
||||||
|
|
||||||
let mut wallet = config.wallet();
|
let mut wallet = config.wallet();
|
||||||
for signer in (1..=config.private_keys_to_add)
|
for signer in additional_signers {
|
||||||
.map(|id| U256::from(id))
|
|
||||||
.map(|id| id.to_be_bytes::<32>())
|
|
||||||
.map(|id| PrivateKeySigner::from_bytes(&FixedBytes(id)).unwrap())
|
|
||||||
{
|
|
||||||
wallet.register_signer(signer);
|
wallet.register_signer(signer);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -529,6 +532,7 @@ impl Drop for Instance {
|
|||||||
mod tests {
|
mod tests {
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
|
|
||||||
|
use alloy::signers::local::PrivateKeySigner;
|
||||||
use temp_dir::TempDir;
|
use temp_dir::TempDir;
|
||||||
|
|
||||||
use crate::{GENESIS_JSON, Node};
|
use crate::{GENESIS_JSON, Node};
|
||||||
@@ -545,7 +549,7 @@ mod tests {
|
|||||||
|
|
||||||
fn new_node() -> (Instance, TempDir) {
|
fn new_node() -> (Instance, TempDir) {
|
||||||
let (args, temp_dir) = test_config();
|
let (args, temp_dir) = test_config();
|
||||||
let mut node = Instance::new(&args);
|
let mut node = Instance::new(&args, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
node.init(GENESIS_JSON.to_owned())
|
node.init(GENESIS_JSON.to_owned())
|
||||||
.expect("Failed to initialize the node")
|
.expect("Failed to initialize the node")
|
||||||
.spawn_process()
|
.spawn_process()
|
||||||
@@ -555,21 +559,23 @@ mod tests {
|
|||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn init_works() {
|
fn init_works() {
|
||||||
Instance::new(&test_config().0)
|
Instance::new(&test_config().0, Vec::<PrivateKeySigner>::with_capacity(0))
|
||||||
.init(GENESIS_JSON.to_string())
|
.init(GENESIS_JSON.to_string())
|
||||||
.unwrap();
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn spawn_works() {
|
fn spawn_works() {
|
||||||
Instance::new(&test_config().0)
|
Instance::new(&test_config().0, Vec::<PrivateKeySigner>::with_capacity(0))
|
||||||
.spawn(GENESIS_JSON.to_string())
|
.spawn(GENESIS_JSON.to_string())
|
||||||
.unwrap();
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn version_works() {
|
fn version_works() {
|
||||||
let version = Instance::new(&test_config().0).version().unwrap();
|
let version = Instance::new(&test_config().0, Vec::<PrivateKeySigner>::with_capacity(0))
|
||||||
|
.version()
|
||||||
|
.unwrap();
|
||||||
assert!(
|
assert!(
|
||||||
version.starts_with("geth version"),
|
version.starts_with("geth version"),
|
||||||
"expected version string, got: '{version}'"
|
"expected version string, got: '{version}'"
|
||||||
|
|||||||
@@ -13,11 +13,9 @@ use alloy::{
|
|||||||
genesis::{Genesis, GenesisAccount},
|
genesis::{Genesis, GenesisAccount},
|
||||||
network::{
|
network::{
|
||||||
Ethereum, EthereumWallet, Network, NetworkWallet, TransactionBuilder,
|
Ethereum, EthereumWallet, Network, NetworkWallet, TransactionBuilder,
|
||||||
TransactionBuilderError, UnbuiltTransactionError,
|
TransactionBuilderError, TxSigner, UnbuiltTransactionError,
|
||||||
},
|
|
||||||
primitives::{
|
|
||||||
Address, B64, B256, BlockHash, BlockNumber, BlockTimestamp, Bloom, Bytes, FixedBytes, U256,
|
|
||||||
},
|
},
|
||||||
|
primitives::{Address, B64, B256, BlockHash, BlockNumber, BlockTimestamp, Bloom, Bytes, U256},
|
||||||
providers::{
|
providers::{
|
||||||
Provider, ProviderBuilder,
|
Provider, ProviderBuilder,
|
||||||
ext::DebugApi,
|
ext::DebugApi,
|
||||||
@@ -28,20 +26,18 @@ use alloy::{
|
|||||||
eth::{Block, Header, Transaction},
|
eth::{Block, Header, Transaction},
|
||||||
trace::geth::{DiffMode, GethDebugTracingOptions, PreStateConfig, PreStateFrame},
|
trace::geth::{DiffMode, GethDebugTracingOptions, PreStateConfig, PreStateFrame},
|
||||||
},
|
},
|
||||||
signers::local::PrivateKeySigner,
|
signers::Signature,
|
||||||
};
|
};
|
||||||
use revive_dt_format::traits::ResolverApi;
|
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use serde_json::{Value as JsonValue, json};
|
use serde_json::{Value as JsonValue, json};
|
||||||
use sp_core::crypto::Ss58Codec;
|
use sp_core::crypto::Ss58Codec;
|
||||||
use sp_runtime::AccountId32;
|
use sp_runtime::AccountId32;
|
||||||
use tracing::Level;
|
use tracing::Level;
|
||||||
|
|
||||||
use revive_dt_common::concepts::BlockingExecutor;
|
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::{BlockingExecutor, EthereumNode};
|
||||||
|
|
||||||
use crate::{Node, common::FallbackGasFiller, constants::INITIAL_BALANCE};
|
use crate::{Node, common::FallbackGasFiller};
|
||||||
|
|
||||||
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
||||||
|
|
||||||
@@ -133,12 +129,10 @@ impl KitchensinkNode {
|
|||||||
for signer_address in
|
for signer_address in
|
||||||
<EthereumWallet as NetworkWallet<Ethereum>>::signer_addresses(&self.wallet)
|
<EthereumWallet as NetworkWallet<Ethereum>>::signer_addresses(&self.wallet)
|
||||||
{
|
{
|
||||||
// Note, the use of the entry API here means that we only modify the entries for any
|
genesis.alloc.entry(signer_address).or_insert(
|
||||||
// account that is not in the `alloc` field of the genesis state.
|
GenesisAccount::default()
|
||||||
genesis
|
.with_balance(10000000000000000000000u128.try_into().unwrap()),
|
||||||
.alloc
|
);
|
||||||
.entry(signer_address)
|
|
||||||
.or_insert(GenesisAccount::default().with_balance(U256::from(INITIAL_BALANCE)));
|
|
||||||
}
|
}
|
||||||
self.extract_balance_from_genesis_file(&genesis)?
|
self.extract_balance_from_genesis_file(&genesis)?
|
||||||
};
|
};
|
||||||
@@ -426,9 +420,7 @@ impl EthereumNode for KitchensinkNode {
|
|||||||
_ => anyhow::bail!("expected a diff mode trace"),
|
_ => anyhow::bail!("expected a diff mode trace"),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
impl ResolverApi for KitchensinkNode {
|
|
||||||
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
||||||
fn chain_id(&self) -> anyhow::Result<alloy::primitives::ChainId> {
|
fn chain_id(&self) -> anyhow::Result<alloy::primitives::ChainId> {
|
||||||
let provider = self.provider();
|
let provider = self.provider();
|
||||||
@@ -512,18 +504,17 @@ impl ResolverApi for KitchensinkNode {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Node for KitchensinkNode {
|
impl Node for KitchensinkNode {
|
||||||
fn new(config: &Arguments) -> Self {
|
fn new(
|
||||||
|
config: &Arguments,
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>,
|
||||||
|
) -> Self {
|
||||||
let kitchensink_directory = config.directory().join(Self::BASE_DIRECTORY);
|
let kitchensink_directory = config.directory().join(Self::BASE_DIRECTORY);
|
||||||
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
||||||
let base_directory = kitchensink_directory.join(id.to_string());
|
let base_directory = kitchensink_directory.join(id.to_string());
|
||||||
let logs_directory = base_directory.join(Self::LOGS_DIRECTORY);
|
let logs_directory = base_directory.join(Self::LOGS_DIRECTORY);
|
||||||
|
|
||||||
let mut wallet = config.wallet();
|
let mut wallet = config.wallet();
|
||||||
for signer in (1..=config.private_keys_to_add)
|
for signer in additional_signers {
|
||||||
.map(|id| U256::from(id))
|
|
||||||
.map(|id| id.to_be_bytes::<32>())
|
|
||||||
.map(|id| PrivateKeySigner::from_bytes(&FixedBytes(id)).unwrap())
|
|
||||||
{
|
|
||||||
wallet.register_signer(signer);
|
wallet.register_signer(signer);
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1039,7 +1030,7 @@ impl BlockHeader for KitchenSinkHeader {
|
|||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use alloy::rpc::types::TransactionRequest;
|
use alloy::{rpc::types::TransactionRequest, signers::local::PrivateKeySigner};
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
use std::sync::{LazyLock, Mutex};
|
use std::sync::{LazyLock, Mutex};
|
||||||
@@ -1082,7 +1073,7 @@ mod tests {
|
|||||||
let _guard = NODE_START_MUTEX.lock().unwrap();
|
let _guard = NODE_START_MUTEX.lock().unwrap();
|
||||||
|
|
||||||
let (args, temp_dir) = test_config();
|
let (args, temp_dir) = test_config();
|
||||||
let mut node = KitchensinkNode::new(&args);
|
let mut node = KitchensinkNode::new(&args, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
node.init(GENESIS_JSON)
|
node.init(GENESIS_JSON)
|
||||||
.expect("Failed to initialize the node")
|
.expect("Failed to initialize the node")
|
||||||
.spawn_process()
|
.spawn_process()
|
||||||
@@ -1137,7 +1128,8 @@ mod tests {
|
|||||||
}
|
}
|
||||||
"#;
|
"#;
|
||||||
|
|
||||||
let mut dummy_node = KitchensinkNode::new(&test_config().0);
|
let mut dummy_node =
|
||||||
|
KitchensinkNode::new(&test_config().0, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
|
|
||||||
// Call `init()`
|
// Call `init()`
|
||||||
dummy_node.init(genesis_content).expect("init failed");
|
dummy_node.init(genesis_content).expect("init failed");
|
||||||
@@ -1181,7 +1173,8 @@ mod tests {
|
|||||||
}
|
}
|
||||||
"#;
|
"#;
|
||||||
|
|
||||||
let node = KitchensinkNode::new(&test_config().0);
|
let node =
|
||||||
|
KitchensinkNode::new(&test_config().0, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
|
|
||||||
let result = node
|
let result = node
|
||||||
.extract_balance_from_genesis_file(&serde_json::from_str(genesis_json).unwrap())
|
.extract_balance_from_genesis_file(&serde_json::from_str(genesis_json).unwrap())
|
||||||
@@ -1254,7 +1247,7 @@ mod tests {
|
|||||||
fn spawn_works() {
|
fn spawn_works() {
|
||||||
let (config, _temp_dir) = test_config();
|
let (config, _temp_dir) = test_config();
|
||||||
|
|
||||||
let mut node = KitchensinkNode::new(&config);
|
let mut node = KitchensinkNode::new(&config, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
node.spawn(GENESIS_JSON.to_string()).unwrap();
|
node.spawn(GENESIS_JSON.to_string()).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1262,7 +1255,7 @@ mod tests {
|
|||||||
fn version_works() {
|
fn version_works() {
|
||||||
let (config, _temp_dir) = test_config();
|
let (config, _temp_dir) = test_config();
|
||||||
|
|
||||||
let node = KitchensinkNode::new(&config);
|
let node = KitchensinkNode::new(&config, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
let version = node.version().unwrap();
|
let version = node.version().unwrap();
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
@@ -1275,7 +1268,7 @@ mod tests {
|
|||||||
fn eth_rpc_version_works() {
|
fn eth_rpc_version_works() {
|
||||||
let (config, _temp_dir) = test_config();
|
let (config, _temp_dir) = test_config();
|
||||||
|
|
||||||
let node = KitchensinkNode::new(&config);
|
let node = KitchensinkNode::new(&config, Vec::<PrivateKeySigner>::with_capacity(0));
|
||||||
let version = node.eth_rpc_version().unwrap();
|
let version = node.eth_rpc_version().unwrap();
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
|
|||||||
@@ -1,10 +1,10 @@
|
|||||||
//! This crate implements the testing nodes.
|
//! This crate implements the testing nodes.
|
||||||
|
|
||||||
|
use alloy::{network::TxSigner, signers::Signature};
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
|
|
||||||
pub mod common;
|
pub mod common;
|
||||||
pub mod constants;
|
|
||||||
pub mod geth;
|
pub mod geth;
|
||||||
pub mod kitchensink;
|
pub mod kitchensink;
|
||||||
pub mod pool;
|
pub mod pool;
|
||||||
@@ -15,7 +15,10 @@ pub const GENESIS_JSON: &str = include_str!("../../../genesis.json");
|
|||||||
/// An abstract interface for testing nodes.
|
/// An abstract interface for testing nodes.
|
||||||
pub trait Node: EthereumNode {
|
pub trait Node: EthereumNode {
|
||||||
/// Create a new uninitialized instance.
|
/// Create a new uninitialized instance.
|
||||||
fn new(config: &Arguments) -> Self;
|
fn new(
|
||||||
|
config: &Arguments,
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>,
|
||||||
|
) -> Self;
|
||||||
|
|
||||||
/// Spawns a node configured according to the genesis json.
|
/// Spawns a node configured according to the genesis json.
|
||||||
///
|
///
|
||||||
|
|||||||
+19
-4
@@ -6,6 +6,7 @@ use std::{
|
|||||||
thread,
|
thread,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
use alloy::{network::TxSigner, signers::Signature};
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
|
|
||||||
@@ -23,7 +24,14 @@ where
|
|||||||
T: Node + Send + 'static,
|
T: Node + Send + 'static,
|
||||||
{
|
{
|
||||||
/// Create a new Pool. This will start as many nodes as there are workers in `config`.
|
/// Create a new Pool. This will start as many nodes as there are workers in `config`.
|
||||||
pub fn new(config: &Arguments) -> anyhow::Result<Self> {
|
pub fn new(
|
||||||
|
config: &Arguments,
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>
|
||||||
|
+ Clone
|
||||||
|
+ Send
|
||||||
|
+ Sync
|
||||||
|
+ 'static,
|
||||||
|
) -> anyhow::Result<Self> {
|
||||||
let nodes = config.workers;
|
let nodes = config.workers;
|
||||||
let genesis = read_to_string(&config.genesis_file).context(format!(
|
let genesis = read_to_string(&config.genesis_file).context(format!(
|
||||||
"can not read genesis file: {}",
|
"can not read genesis file: {}",
|
||||||
@@ -34,7 +42,10 @@ where
|
|||||||
for _ in 0..nodes {
|
for _ in 0..nodes {
|
||||||
let config = config.clone();
|
let config = config.clone();
|
||||||
let genesis = genesis.clone();
|
let genesis = genesis.clone();
|
||||||
handles.push(thread::spawn(move || spawn_node::<T>(&config, genesis)));
|
let additional_signers = additional_signers.clone();
|
||||||
|
handles.push(thread::spawn(move || {
|
||||||
|
spawn_node::<T>(&config, additional_signers, genesis)
|
||||||
|
}));
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut nodes = Vec::with_capacity(nodes);
|
let mut nodes = Vec::with_capacity(nodes);
|
||||||
@@ -60,8 +71,12 @@ where
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn spawn_node<T: Node + Send>(args: &Arguments, genesis: String) -> anyhow::Result<T> {
|
fn spawn_node<T: Node + Send>(
|
||||||
let mut node = T::new(args);
|
args: &Arguments,
|
||||||
|
additional_signers: impl IntoIterator<Item: TxSigner<Signature> + Send + Sync + 'static>,
|
||||||
|
genesis: String,
|
||||||
|
) -> anyhow::Result<T> {
|
||||||
|
let mut node = T::new(args, additional_signers);
|
||||||
tracing::info!("starting node: {}", node.connection_string());
|
tracing::info!("starting node: {}", node.connection_string());
|
||||||
node.spawn(genesis)?;
|
node.spawn(genesis)?;
|
||||||
Ok(node)
|
Ok(node)
|
||||||
|
|||||||
+5
-1
@@ -33,5 +33,9 @@
|
|||||||
"mixhash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
"mixhash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"parentHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
"parentHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"timestamp": "0x00",
|
"timestamp": "0x00",
|
||||||
"alloc": {}
|
"alloc": {
|
||||||
|
"90F8bf6A479f320ead074411a4B0e7944Ea8c9C1": {
|
||||||
|
"balance": "10000000000000000000000"
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
Reference in New Issue
Block a user