mirror of
https://github.com/pezkuwichain/revive-differential-tests.git
synced 2026-04-26 06:17:58 +00:00
Compare commits
36 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| edbbbab934 | |||
| 757cbaf61e | |||
| 2c7e95bb4f | |||
| 2b6ee18e40 | |||
| ca1ce3e5b4 | |||
| 3d28179dde | |||
| 6303f3b917 | |||
| e434c163c7 | |||
| 404d2f7452 | |||
| 750d9fff27 | |||
| 94afdbecc7 | |||
| ce1b4862b4 | |||
| 799f07f04f | |||
| 2a1f81fb6d | |||
| 51191013c1 | |||
| c81279fc8f | |||
| 0392b6b629 | |||
| ac94c972de | |||
| 5cc814b0e0 | |||
| 0722791a28 | |||
| 52042dfff5 | |||
| e3c717f4d9 | |||
| e7ebe4fa2f | |||
| 5c957e5ac1 | |||
| a0248b58f3 | |||
| 6a9acea748 | |||
| c31c7d94f9 | |||
| 589a5dc988 | |||
| dcc43d4ea6 | |||
| cb13c3c2cd | |||
| 7cda3416f0 | |||
| d7bc4f1fab | |||
| b6db597a57 | |||
| d7136d9a3d | |||
| c6d55515be | |||
| a9970eb2bb |
Generated
+4
-4
@@ -346,7 +346,7 @@ dependencies = [
|
|||||||
"keccak-asm",
|
"keccak-asm",
|
||||||
"paste",
|
"paste",
|
||||||
"proptest",
|
"proptest",
|
||||||
"rand 0.9.1",
|
"rand 0.9.2",
|
||||||
"ruint",
|
"ruint",
|
||||||
"rustc-hash",
|
"rustc-hash",
|
||||||
"serde",
|
"serde",
|
||||||
@@ -3737,9 +3737,9 @@ dependencies = [
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rand"
|
name = "rand"
|
||||||
version = "0.9.1"
|
version = "0.9.2"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "9fbfd9d094a40bf3ae768db9361049ace4c0e04a4fd6b359518bd7b73a73dd97"
|
checksum = "6db2770f06117d490610c7488547d543617b21bfa07796d7a12f6f1bd53850d1"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"rand_chacha 0.9.0",
|
"rand_chacha 0.9.0",
|
||||||
"rand_core 0.9.3",
|
"rand_core 0.9.3",
|
||||||
@@ -4131,7 +4131,7 @@ dependencies = [
|
|||||||
"primitive-types 0.12.2",
|
"primitive-types 0.12.2",
|
||||||
"proptest",
|
"proptest",
|
||||||
"rand 0.8.5",
|
"rand 0.8.5",
|
||||||
"rand 0.9.1",
|
"rand 0.9.2",
|
||||||
"rlp",
|
"rlp",
|
||||||
"ruint-macro",
|
"ruint-macro",
|
||||||
"serde",
|
"serde",
|
||||||
|
|||||||
@@ -73,6 +73,12 @@ pub struct Arguments {
|
|||||||
)]
|
)]
|
||||||
pub account: String,
|
pub account: String,
|
||||||
|
|
||||||
|
/// This argument controls which private keys the nodes should have access to and be added to
|
||||||
|
/// its wallet signers. With a value of N, private keys (0, N] will be added to the signer set
|
||||||
|
/// of the node.
|
||||||
|
#[arg(long = "private-keys-count", default_value_t = 30)]
|
||||||
|
pub private_keys_to_add: usize,
|
||||||
|
|
||||||
/// The differential testing leader node implementation.
|
/// The differential testing leader node implementation.
|
||||||
#[arg(short, long = "leader", default_value = "geth")]
|
#[arg(short, long = "leader", default_value = "geth")]
|
||||||
pub leader: TestingPlatform,
|
pub leader: TestingPlatform,
|
||||||
@@ -123,18 +129,16 @@ impl Arguments {
|
|||||||
panic!("should have a workdir configured")
|
panic!("should have a workdir configured")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn signer(&self) -> PrivateKeySigner {
|
|
||||||
self.account
|
|
||||||
.parse::<PrivateKeySigner>()
|
|
||||||
.unwrap_or_else(|error| {
|
|
||||||
panic!("private key '{}' parsing error: {error}", self.account);
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Try to parse `self.account` into a [PrivateKeySigner],
|
/// Try to parse `self.account` into a [PrivateKeySigner],
|
||||||
/// panicing on error.
|
/// panicing on error.
|
||||||
pub fn wallet(&self) -> EthereumWallet {
|
pub fn wallet(&self) -> EthereumWallet {
|
||||||
EthereumWallet::new(self.signer())
|
let signer = self
|
||||||
|
.account
|
||||||
|
.parse::<PrivateKeySigner>()
|
||||||
|
.unwrap_or_else(|error| {
|
||||||
|
panic!("private key '{}' parsing error: {error}", self.account);
|
||||||
|
});
|
||||||
|
EthereumWallet::new(signer)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -5,10 +5,10 @@ use std::marker::PhantomData;
|
|||||||
|
|
||||||
use alloy::json_abi::JsonAbi;
|
use alloy::json_abi::JsonAbi;
|
||||||
use alloy::network::{Ethereum, TransactionBuilder};
|
use alloy::network::{Ethereum, TransactionBuilder};
|
||||||
use alloy::primitives::Bytes;
|
|
||||||
use alloy::rpc::types::TransactionReceipt;
|
use alloy::rpc::types::TransactionReceipt;
|
||||||
use alloy::rpc::types::trace::geth::{
|
use alloy::rpc::types::trace::geth::{
|
||||||
DefaultFrame, GethDebugTracingOptions, GethDefaultTracingOptions, GethTrace, PreStateConfig,
|
CallFrame, GethDebugBuiltInTracerType, GethDebugTracerType, GethDebugTracingOptions, GethTrace,
|
||||||
|
PreStateConfig,
|
||||||
};
|
};
|
||||||
use alloy::{
|
use alloy::{
|
||||||
primitives::Address,
|
primitives::Address,
|
||||||
@@ -25,6 +25,7 @@ use revive_dt_format::case::CaseIdx;
|
|||||||
use revive_dt_format::input::{Calldata, Expected, ExpectedOutput, Method};
|
use revive_dt_format::input::{Calldata, Expected, ExpectedOutput, Method};
|
||||||
use revive_dt_format::metadata::{ContractInstance, ContractPathAndIdentifier};
|
use revive_dt_format::metadata::{ContractInstance, ContractPathAndIdentifier};
|
||||||
use revive_dt_format::{input::Input, metadata::Metadata, mode::SolcMode};
|
use revive_dt_format::{input::Input, metadata::Metadata, mode::SolcMode};
|
||||||
|
use revive_dt_node::Node;
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
use revive_dt_report::reporter::{CompilationTask, Report, Span};
|
use revive_dt_report::reporter::{CompilationTask, Report, Span};
|
||||||
use revive_solc_json_interface::SolcStandardJsonOutput;
|
use revive_solc_json_interface::SolcStandardJsonOutput;
|
||||||
@@ -249,6 +250,12 @@ where
|
|||||||
|
|
||||||
let tx = {
|
let tx = {
|
||||||
let tx = TransactionRequest::default().from(input.caller);
|
let tx = TransactionRequest::default().from(input.caller);
|
||||||
|
let tx = match input.value {
|
||||||
|
Some(ref value) if deploy_with_constructor_arguments => {
|
||||||
|
tx.value(value.into_inner())
|
||||||
|
}
|
||||||
|
_ => tx,
|
||||||
|
};
|
||||||
TransactionBuilder::<Ethereum>::with_deploy_code(tx, code)
|
TransactionBuilder::<Ethereum>::with_deploy_code(tx, code)
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -365,23 +372,7 @@ where
|
|||||||
let _guard = span.enter();
|
let _guard = span.enter();
|
||||||
|
|
||||||
// Resolving the `input.expected` into a series of expectations that we can then assert on.
|
// Resolving the `input.expected` into a series of expectations that we can then assert on.
|
||||||
let expectations = match input {
|
let mut expectations = match input {
|
||||||
// This is a bit of a special case and we have to support it separately on it's own. If
|
|
||||||
// it's a call to the deployer method, then the tests will assert that it "returns" the
|
|
||||||
// address of the contract. Deployments do not return the address of the contract but
|
|
||||||
// the runtime code of the contracts. Therefore, this assertion would always fail. So,
|
|
||||||
// we replace it with an assertion of "check if it succeeded"
|
|
||||||
Input {
|
|
||||||
expected: Some(Expected::Calldata(Calldata::Compound(compound))),
|
|
||||||
method: Method::Deployer,
|
|
||||||
..
|
|
||||||
} if compound.len() == 1
|
|
||||||
&& compound
|
|
||||||
.first()
|
|
||||||
.is_some_and(|first| first.contains(".address")) =>
|
|
||||||
{
|
|
||||||
vec![ExpectedOutput::new().with_success()]
|
|
||||||
}
|
|
||||||
Input {
|
Input {
|
||||||
expected: Some(Expected::Calldata(calldata)),
|
expected: Some(Expected::Calldata(calldata)),
|
||||||
..
|
..
|
||||||
@@ -397,6 +388,17 @@ where
|
|||||||
Input { expected: None, .. } => vec![ExpectedOutput::new().with_success()],
|
Input { expected: None, .. } => vec![ExpectedOutput::new().with_success()],
|
||||||
};
|
};
|
||||||
|
|
||||||
|
// This is a bit of a special case and we have to support it separately on it's own. If it's
|
||||||
|
// a call to the deployer method, then the tests will assert that it "returns" the address
|
||||||
|
// of the contract. Deployments do not return the address of the contract but the runtime
|
||||||
|
// code of the contracts. Therefore, this assertion would always fail. So, we replace it
|
||||||
|
// with an assertion of "check if it succeeded"
|
||||||
|
if let Method::Deployer = &input.method {
|
||||||
|
for expectation in expectations.iter_mut() {
|
||||||
|
expectation.return_data = None;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Note: we need to do assertions and checks on the output of the last call and this isn't
|
// Note: we need to do assertions and checks on the output of the last call and this isn't
|
||||||
// available in the receipt. The only way to get this information is through tracing on the
|
// available in the receipt. The only way to get this information is through tracing on the
|
||||||
// node.
|
// node.
|
||||||
@@ -404,12 +406,14 @@ where
|
|||||||
.trace_transaction(
|
.trace_transaction(
|
||||||
execution_receipt,
|
execution_receipt,
|
||||||
GethDebugTracingOptions {
|
GethDebugTracingOptions {
|
||||||
config: GethDefaultTracingOptions::default().with_enable_return_data(true),
|
tracer: Some(GethDebugTracerType::BuiltInTracer(
|
||||||
|
GethDebugBuiltInTracerType::CallTracer,
|
||||||
|
)),
|
||||||
..Default::default()
|
..Default::default()
|
||||||
},
|
},
|
||||||
)?
|
)?
|
||||||
.try_into_default_frame()
|
.try_into_call_frame()
|
||||||
.expect("Impossible. We can't request default tracing and get some other type back");
|
.expect("Impossible - we requested a callframe trace so we must get it back");
|
||||||
|
|
||||||
for expectation in expectations.iter() {
|
for expectation in expectations.iter() {
|
||||||
self.handle_input_expectation_item(
|
self.handle_input_expectation_item(
|
||||||
@@ -430,23 +434,21 @@ where
|
|||||||
execution_receipt: &TransactionReceipt,
|
execution_receipt: &TransactionReceipt,
|
||||||
node: &T::Blockchain,
|
node: &T::Blockchain,
|
||||||
expectation: &ExpectedOutput,
|
expectation: &ExpectedOutput,
|
||||||
tracing_result: &DefaultFrame,
|
tracing_result: &CallFrame,
|
||||||
) -> anyhow::Result<()> {
|
) -> anyhow::Result<()> {
|
||||||
// TODO: We want to respect the compiler version filter on the expected output but would
|
// TODO: We want to respect the compiler version filter on the expected output but would
|
||||||
// require some changes to the interfaces of the compiler and such. So, we add it later.
|
// require some changes to the interfaces of the compiler and such. So, we add it later.
|
||||||
// Additionally, what happens if the compiler filter doesn't match? Do we consider that the
|
// Additionally, what happens if the compiler filter doesn't match? Do we consider that the
|
||||||
// transaction should succeed? Do we just ignore the expectation?
|
// transaction should succeed? Do we just ignore the expectation?
|
||||||
|
|
||||||
|
let deployed_contracts = self.deployed_contracts.entry(case_idx).or_default();
|
||||||
|
let chain_state_provider = node;
|
||||||
|
|
||||||
// Handling the receipt state assertion.
|
// Handling the receipt state assertion.
|
||||||
let expected = !expectation.exception;
|
let expected = !expectation.exception;
|
||||||
let actual = execution_receipt.status();
|
let actual = execution_receipt.status();
|
||||||
if actual != expected {
|
if actual != expected {
|
||||||
tracing::error!(
|
tracing::error!(expected, actual, "Transaction status assertion failed",);
|
||||||
?execution_receipt,
|
|
||||||
expected,
|
|
||||||
actual,
|
|
||||||
"Transaction status assertion failed",
|
|
||||||
);
|
|
||||||
anyhow::bail!(
|
anyhow::bail!(
|
||||||
"Transaction status assertion failed - Expected {expected} but got {actual}",
|
"Transaction status assertion failed - Expected {expected} but got {actual}",
|
||||||
);
|
);
|
||||||
@@ -454,13 +456,16 @@ where
|
|||||||
|
|
||||||
// Handling the calldata assertion
|
// Handling the calldata assertion
|
||||||
if let Some(ref expected_calldata) = expectation.return_data {
|
if let Some(ref expected_calldata) = expectation.return_data {
|
||||||
let expected = expected_calldata
|
let expected = expected_calldata;
|
||||||
.calldata(self.deployed_contracts.entry(case_idx).or_default(), node)
|
let actual = &tracing_result.output.as_ref().unwrap_or_default();
|
||||||
.map(Bytes::from)?;
|
if !expected.is_equivalent(actual, deployed_contracts, chain_state_provider)? {
|
||||||
let actual = tracing_result.return_value.clone();
|
tracing::error!(
|
||||||
if !expected.starts_with(&actual) {
|
?execution_receipt,
|
||||||
tracing::error!(?execution_receipt, %expected, %actual, "Calldata assertion failed");
|
?expected,
|
||||||
anyhow::bail!("Calldata assertion failed - Expected {expected} but got {actual}",);
|
%actual,
|
||||||
|
"Calldata assertion failed"
|
||||||
|
);
|
||||||
|
anyhow::bail!("Calldata assertion failed - Expected {expected:?} but got {actual}",);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -470,12 +475,7 @@ where
|
|||||||
let expected = expected_events.len();
|
let expected = expected_events.len();
|
||||||
let actual = execution_receipt.logs().len();
|
let actual = execution_receipt.logs().len();
|
||||||
if actual != expected {
|
if actual != expected {
|
||||||
tracing::error!(
|
tracing::error!(expected, actual, "Event count assertion failed",);
|
||||||
?execution_receipt,
|
|
||||||
expected,
|
|
||||||
actual,
|
|
||||||
"Event count assertion failed",
|
|
||||||
);
|
|
||||||
anyhow::bail!(
|
anyhow::bail!(
|
||||||
"Event count assertion failed - Expected {expected} but got {actual}",
|
"Event count assertion failed - Expected {expected} but got {actual}",
|
||||||
);
|
);
|
||||||
@@ -491,7 +491,6 @@ where
|
|||||||
let actual = actual_event.address();
|
let actual = actual_event.address();
|
||||||
if actual != expected {
|
if actual != expected {
|
||||||
tracing::error!(
|
tracing::error!(
|
||||||
?execution_receipt,
|
|
||||||
%expected,
|
%expected,
|
||||||
%actual,
|
%actual,
|
||||||
"Event emitter assertion failed",
|
"Event emitter assertion failed",
|
||||||
@@ -503,27 +502,34 @@ where
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Handling the topics assertion.
|
// Handling the topics assertion.
|
||||||
let expected = expected_event.topics.as_slice();
|
for (expected, actual) in expected_event
|
||||||
let actual = actual_event.topics();
|
.topics
|
||||||
if actual != expected {
|
.as_slice()
|
||||||
tracing::error!(
|
.iter()
|
||||||
?execution_receipt,
|
.zip(actual_event.topics())
|
||||||
?expected,
|
{
|
||||||
?actual,
|
let expected = Calldata::Compound(vec![expected.clone()]);
|
||||||
"Event topics assertion failed",
|
if !expected.is_equivalent(
|
||||||
);
|
&actual.0,
|
||||||
anyhow::bail!(
|
deployed_contracts,
|
||||||
"Event topics assertion failed - Expected {expected:?} but got {actual:?}",
|
chain_state_provider,
|
||||||
);
|
)? {
|
||||||
|
tracing::error!(
|
||||||
|
?execution_receipt,
|
||||||
|
?expected,
|
||||||
|
?actual,
|
||||||
|
"Event topics assertion failed",
|
||||||
|
);
|
||||||
|
anyhow::bail!(
|
||||||
|
"Event topics assertion failed - Expected {expected:?} but got {actual:?}",
|
||||||
|
);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handling the values assertion.
|
// Handling the values assertion.
|
||||||
let expected = &expected_event
|
let expected = &expected_event.values;
|
||||||
.values
|
|
||||||
.calldata(self.deployed_contracts.entry(case_idx).or_default(), node)
|
|
||||||
.map(Bytes::from)?;
|
|
||||||
let actual = &actual_event.data().data;
|
let actual = &actual_event.data().data;
|
||||||
if !expected.starts_with(actual) {
|
if !expected.is_equivalent(&actual.0, deployed_contracts, chain_state_provider)? {
|
||||||
tracing::error!(
|
tracing::error!(
|
||||||
?execution_receipt,
|
?execution_receipt,
|
||||||
?expected,
|
?expected,
|
||||||
@@ -643,6 +649,22 @@ where
|
|||||||
let tracing_span = tracing::info_span!("Handling metadata file");
|
let tracing_span = tracing::info_span!("Handling metadata file");
|
||||||
let _guard = tracing_span.enter();
|
let _guard = tracing_span.enter();
|
||||||
|
|
||||||
|
// We only execute this input if it's valid for the leader and the follower. Otherwise, we
|
||||||
|
// skip it with a warning.
|
||||||
|
if !self
|
||||||
|
.leader_node
|
||||||
|
.matches_target(self.metadata.targets.as_deref())
|
||||||
|
|| !self
|
||||||
|
.follower_node
|
||||||
|
.matches_target(self.metadata.targets.as_deref())
|
||||||
|
{
|
||||||
|
tracing::warn!(
|
||||||
|
targets = ?self.metadata.targets,
|
||||||
|
"Either the leader or follower node do not support the targets of the file"
|
||||||
|
);
|
||||||
|
return execution_result;
|
||||||
|
}
|
||||||
|
|
||||||
for mode in self.metadata.solc_modes() {
|
for mode in self.metadata.solc_modes() {
|
||||||
let tracing_span = tracing::info_span!("With solc mode", solc_mode = ?mode);
|
let tracing_span = tracing::info_span!("With solc mode", solc_mode = ?mode);
|
||||||
let _guard = tracing_span.enter();
|
let _guard = tracing_span.enter();
|
||||||
@@ -686,6 +708,7 @@ where
|
|||||||
|
|
||||||
// For cases if one of the inputs fail then we move on to the next case and we do NOT
|
// For cases if one of the inputs fail then we move on to the next case and we do NOT
|
||||||
// bail out of the whole thing.
|
// bail out of the whole thing.
|
||||||
|
|
||||||
'case_loop: for (case_idx, case) in self.metadata.cases.iter().enumerate() {
|
'case_loop: for (case_idx, case) in self.metadata.cases.iter().enumerate() {
|
||||||
let tracing_span = tracing::info_span!(
|
let tracing_span = tracing::info_span!(
|
||||||
"Handling case",
|
"Handling case",
|
||||||
|
|||||||
@@ -5,7 +5,7 @@
|
|||||||
|
|
||||||
use revive_dt_compiler::{SolidityCompiler, revive_resolc, solc};
|
use revive_dt_compiler::{SolidityCompiler, revive_resolc, solc};
|
||||||
use revive_dt_config::TestingPlatform;
|
use revive_dt_config::TestingPlatform;
|
||||||
use revive_dt_node::{geth, kitchensink::KitchensinkNode};
|
use revive_dt_node::{Node, geth, kitchensink::KitchensinkNode};
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
|
|
||||||
pub mod common;
|
pub mod common;
|
||||||
@@ -15,7 +15,7 @@ pub mod driver;
|
|||||||
///
|
///
|
||||||
/// For this we need a blockchain node implementation and a compiler.
|
/// For this we need a blockchain node implementation and a compiler.
|
||||||
pub trait Platform {
|
pub trait Platform {
|
||||||
type Blockchain: EthereumNode;
|
type Blockchain: EthereumNode + Node;
|
||||||
type Compiler: SolidityCompiler;
|
type Compiler: SolidityCompiler;
|
||||||
|
|
||||||
/// Returns the matching [TestingPlatform] of the [revive_dt_config::Arguments].
|
/// Returns the matching [TestingPlatform] of the [revive_dt_config::Arguments].
|
||||||
|
|||||||
+8
-31
@@ -1,6 +1,5 @@
|
|||||||
use std::{collections::HashMap, sync::LazyLock};
|
use std::{collections::HashMap, sync::LazyLock};
|
||||||
|
|
||||||
use alloy::primitives::Address;
|
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use rayon::{ThreadPoolBuilder, prelude::*};
|
use rayon::{ThreadPoolBuilder, prelude::*};
|
||||||
|
|
||||||
@@ -9,7 +8,7 @@ use revive_dt_core::{
|
|||||||
Geth, Kitchensink, Platform,
|
Geth, Kitchensink, Platform,
|
||||||
driver::{Driver, State},
|
driver::{Driver, State},
|
||||||
};
|
};
|
||||||
use revive_dt_format::{corpus::Corpus, input::default_caller, metadata::MetadataFile};
|
use revive_dt_format::{corpus::Corpus, metadata::MetadataFile};
|
||||||
use revive_dt_node::pool::NodePool;
|
use revive_dt_node::pool::NodePool;
|
||||||
use revive_dt_report::reporter::{Report, Span};
|
use revive_dt_report::reporter::{Report, Span};
|
||||||
use temp_dir::TempDir;
|
use temp_dir::TempDir;
|
||||||
@@ -21,24 +20,12 @@ static TEMP_DIR: LazyLock<TempDir> = LazyLock::new(|| TempDir::new().unwrap());
|
|||||||
fn main() -> anyhow::Result<()> {
|
fn main() -> anyhow::Result<()> {
|
||||||
let args = init_cli()?;
|
let args = init_cli()?;
|
||||||
|
|
||||||
let corpora = collect_corpora(&args)?;
|
|
||||||
let additional_callers = corpora
|
|
||||||
.values()
|
|
||||||
.flat_map(|value| value.iter().map(|metadata| &metadata.cases))
|
|
||||||
.flat_map(|case| case.iter().map(|case| &case.inputs))
|
|
||||||
.flatten()
|
|
||||||
.map(|input| input.caller)
|
|
||||||
.filter(|caller| caller != &default_caller())
|
|
||||||
.collect::<Vec<_>>();
|
|
||||||
|
|
||||||
tracing::debug!(?additional_callers, "Discovered callers");
|
|
||||||
|
|
||||||
for (corpus, tests) in collect_corpora(&args)? {
|
for (corpus, tests) in collect_corpora(&args)? {
|
||||||
let span = Span::new(corpus, args.clone())?;
|
let span = Span::new(corpus, args.clone())?;
|
||||||
|
|
||||||
match &args.compile_only {
|
match &args.compile_only {
|
||||||
Some(platform) => compile_corpus(&args, &tests, platform, span),
|
Some(platform) => compile_corpus(&args, &tests, platform, span),
|
||||||
None => execute_corpus(&args, &tests, &additional_callers, span)?,
|
None => execute_corpus(&args, &tests, span)?,
|
||||||
}
|
}
|
||||||
|
|
||||||
Report::save()?;
|
Report::save()?;
|
||||||
@@ -96,20 +83,15 @@ fn collect_corpora(args: &Arguments) -> anyhow::Result<HashMap<Corpus, Vec<Metad
|
|||||||
Ok(corpora)
|
Ok(corpora)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn run_driver<L, F>(
|
fn run_driver<L, F>(args: &Arguments, tests: &[MetadataFile], span: Span) -> anyhow::Result<()>
|
||||||
args: &Arguments,
|
|
||||||
tests: &[MetadataFile],
|
|
||||||
additional_callers: &[Address],
|
|
||||||
span: Span,
|
|
||||||
) -> anyhow::Result<()>
|
|
||||||
where
|
where
|
||||||
L: Platform,
|
L: Platform,
|
||||||
F: Platform,
|
F: Platform,
|
||||||
L::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
L::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
||||||
F::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
F::Blockchain: revive_dt_node::Node + Send + Sync + 'static,
|
||||||
{
|
{
|
||||||
let leader_nodes = NodePool::<L::Blockchain>::new(args, additional_callers)?;
|
let leader_nodes = NodePool::<L::Blockchain>::new(args)?;
|
||||||
let follower_nodes = NodePool::<F::Blockchain>::new(args, additional_callers)?;
|
let follower_nodes = NodePool::<F::Blockchain>::new(args)?;
|
||||||
|
|
||||||
tests.par_iter().for_each(
|
tests.par_iter().for_each(
|
||||||
|MetadataFile {
|
|MetadataFile {
|
||||||
@@ -159,18 +141,13 @@ where
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
fn execute_corpus(
|
fn execute_corpus(args: &Arguments, tests: &[MetadataFile], span: Span) -> anyhow::Result<()> {
|
||||||
args: &Arguments,
|
|
||||||
tests: &[MetadataFile],
|
|
||||||
additional_callers: &[Address],
|
|
||||||
span: Span,
|
|
||||||
) -> anyhow::Result<()> {
|
|
||||||
match (&args.leader, &args.follower) {
|
match (&args.leader, &args.follower) {
|
||||||
(TestingPlatform::Geth, TestingPlatform::Kitchensink) => {
|
(TestingPlatform::Geth, TestingPlatform::Kitchensink) => {
|
||||||
run_driver::<Geth, Kitchensink>(args, tests, additional_callers, span)?
|
run_driver::<Geth, Kitchensink>(args, tests, span)?
|
||||||
}
|
}
|
||||||
(TestingPlatform::Geth, TestingPlatform::Geth) => {
|
(TestingPlatform::Geth, TestingPlatform::Geth) => {
|
||||||
run_driver::<Geth, Geth>(args, tests, additional_callers, span)?
|
run_driver::<Geth, Geth>(args, tests, span)?
|
||||||
}
|
}
|
||||||
_ => unimplemented!(),
|
_ => unimplemented!(),
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -16,12 +16,6 @@ pub struct Case {
|
|||||||
pub expected: Option<Expected>,
|
pub expected: Option<Expected>,
|
||||||
}
|
}
|
||||||
|
|
||||||
define_wrapper_type!(
|
|
||||||
/// A wrapper type for the index of test cases found in metadata file.
|
|
||||||
#[derive(Clone, Copy, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)]
|
|
||||||
CaseIdx(usize);
|
|
||||||
);
|
|
||||||
|
|
||||||
impl Case {
|
impl Case {
|
||||||
pub fn inputs_iterator(&self) -> impl Iterator<Item = Input> {
|
pub fn inputs_iterator(&self) -> impl Iterator<Item = Input> {
|
||||||
let inputs_len = self.inputs.len();
|
let inputs_len = self.inputs.len();
|
||||||
@@ -31,7 +25,15 @@ impl Case {
|
|||||||
.enumerate()
|
.enumerate()
|
||||||
.map(move |(idx, mut input)| {
|
.map(move |(idx, mut input)| {
|
||||||
if idx + 1 == inputs_len {
|
if idx + 1 == inputs_len {
|
||||||
input.expected = self.expected.clone();
|
if input.expected.is_none() {
|
||||||
|
input.expected = self.expected.clone();
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: What does it mean for us to have an `expected` field on the case itself
|
||||||
|
// but the final input also has an expected field that doesn't match the one on
|
||||||
|
// the case? What are we supposed to do with that final expected field on the
|
||||||
|
// case?
|
||||||
|
|
||||||
input
|
input
|
||||||
} else {
|
} else {
|
||||||
input
|
input
|
||||||
@@ -39,3 +41,9 @@ impl Case {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define_wrapper_type!(
|
||||||
|
/// A wrapper type for the index of test cases found in metadata file.
|
||||||
|
#[derive(Clone, Copy, Debug, PartialEq, Eq, PartialOrd, Ord, Hash)]
|
||||||
|
CaseIdx(usize);
|
||||||
|
);
|
||||||
|
|||||||
+93
-18
@@ -7,13 +7,13 @@ use alloy::{
|
|||||||
primitives::{Address, Bytes, U256},
|
primitives::{Address, Bytes, U256},
|
||||||
rpc::types::TransactionRequest,
|
rpc::types::TransactionRequest,
|
||||||
};
|
};
|
||||||
use alloy_primitives::B256;
|
use alloy_primitives::{FixedBytes, utils::parse_units};
|
||||||
use semver::VersionReq;
|
use semver::VersionReq;
|
||||||
use serde::Deserialize;
|
use serde::{Deserialize, Serialize};
|
||||||
|
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
|
|
||||||
use crate::metadata::ContractInstance;
|
use crate::{define_wrapper_type, metadata::ContractInstance};
|
||||||
|
|
||||||
#[derive(Clone, Debug, Default, Deserialize, Eq, PartialEq)]
|
#[derive(Clone, Debug, Default, Deserialize, Eq, PartialEq)]
|
||||||
pub struct Input {
|
pub struct Input {
|
||||||
@@ -26,7 +26,7 @@ pub struct Input {
|
|||||||
#[serde(default)]
|
#[serde(default)]
|
||||||
pub calldata: Calldata,
|
pub calldata: Calldata,
|
||||||
pub expected: Option<Expected>,
|
pub expected: Option<Expected>,
|
||||||
pub value: Option<String>,
|
pub value: Option<EtherValue>,
|
||||||
pub storage: Option<HashMap<String, Calldata>>,
|
pub storage: Option<HashMap<String, Calldata>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -50,14 +50,14 @@ pub struct ExpectedOutput {
|
|||||||
#[derive(Clone, Debug, Default, Deserialize, Eq, PartialEq)]
|
#[derive(Clone, Debug, Default, Deserialize, Eq, PartialEq)]
|
||||||
pub struct Event {
|
pub struct Event {
|
||||||
pub address: Option<Address>,
|
pub address: Option<Address>,
|
||||||
pub topics: Vec<B256>,
|
pub topics: Vec<String>,
|
||||||
pub values: Calldata,
|
pub values: Calldata,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, Deserialize, Eq, PartialEq)]
|
#[derive(Clone, Debug, Deserialize, Eq, PartialEq)]
|
||||||
#[serde(untagged)]
|
#[serde(untagged)]
|
||||||
pub enum Calldata {
|
pub enum Calldata {
|
||||||
Single(String),
|
Single(Bytes),
|
||||||
Compound(Vec<String>),
|
Compound(Vec<String>),
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -82,6 +82,37 @@ pub enum Method {
|
|||||||
FunctionName(String),
|
FunctionName(String),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
define_wrapper_type!(
|
||||||
|
#[derive(Clone, Copy, Debug, Default, PartialEq, Eq, PartialOrd, Ord, Hash)]
|
||||||
|
EtherValue(U256);
|
||||||
|
);
|
||||||
|
|
||||||
|
impl Serialize for EtherValue {
|
||||||
|
fn serialize<S>(&self, serializer: S) -> Result<S::Ok, S::Error>
|
||||||
|
where
|
||||||
|
S: serde::Serializer,
|
||||||
|
{
|
||||||
|
format!("{} wei", self.0).serialize(serializer)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<'de> Deserialize<'de> for EtherValue {
|
||||||
|
fn deserialize<D>(deserializer: D) -> Result<Self, D::Error>
|
||||||
|
where
|
||||||
|
D: serde::Deserializer<'de>,
|
||||||
|
{
|
||||||
|
let string = String::deserialize(deserializer)?;
|
||||||
|
let mut splitted = string.split(' ');
|
||||||
|
let (Some(value), Some(unit)) = (splitted.next(), splitted.next()) else {
|
||||||
|
return Err(serde::de::Error::custom("Failed to parse the value"));
|
||||||
|
};
|
||||||
|
let parsed = parse_units(value, unit.replace("eth", "ether"))
|
||||||
|
.map_err(|_| serde::de::Error::custom("Failed to parse units"))?
|
||||||
|
.into();
|
||||||
|
Ok(Self(parsed))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl ExpectedOutput {
|
impl ExpectedOutput {
|
||||||
pub fn new() -> Self {
|
pub fn new() -> Self {
|
||||||
Default::default()
|
Default::default()
|
||||||
@@ -137,8 +168,8 @@ impl Calldata {
|
|||||||
chain_state_provider: &impl EthereumNode,
|
chain_state_provider: &impl EthereumNode,
|
||||||
) -> anyhow::Result<()> {
|
) -> anyhow::Result<()> {
|
||||||
match self {
|
match self {
|
||||||
Calldata::Single(string) => {
|
Calldata::Single(bytes) => {
|
||||||
alloy::hex::decode_to_slice(string, buffer)?;
|
buffer.extend_from_slice(bytes);
|
||||||
}
|
}
|
||||||
Calldata::Compound(items) => {
|
Calldata::Compound(items) => {
|
||||||
for (arg_idx, arg) in items.iter().enumerate() {
|
for (arg_idx, arg) in items.iter().enumerate() {
|
||||||
@@ -159,10 +190,48 @@ impl Calldata {
|
|||||||
|
|
||||||
pub fn size_requirement(&self) -> usize {
|
pub fn size_requirement(&self) -> usize {
|
||||||
match self {
|
match self {
|
||||||
Calldata::Single(single) => (single.len() - 2) / 2,
|
Calldata::Single(single) => single.len(),
|
||||||
Calldata::Compound(items) => items.len() * 32,
|
Calldata::Compound(items) => items.len() * 32,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Checks if this [`Calldata`] is equivalent to the passed calldata bytes.
|
||||||
|
pub fn is_equivalent(
|
||||||
|
&self,
|
||||||
|
other: &[u8],
|
||||||
|
deployed_contracts: &HashMap<ContractInstance, (Address, JsonAbi)>,
|
||||||
|
chain_state_provider: &impl EthereumNode,
|
||||||
|
) -> anyhow::Result<bool> {
|
||||||
|
match self {
|
||||||
|
Calldata::Single(calldata) => Ok(calldata == other),
|
||||||
|
Calldata::Compound(items) => {
|
||||||
|
// Chunking the "other" calldata into 32 byte chunks since each
|
||||||
|
// one of the items in the compound calldata represents 32 bytes
|
||||||
|
for (this, other) in items.iter().zip(other.chunks(32)) {
|
||||||
|
// The matterlabs format supports wildcards and therefore we
|
||||||
|
// also need to support them.
|
||||||
|
if this == "*" {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let other = if other.len() < 32 {
|
||||||
|
let mut vec = other.to_vec();
|
||||||
|
vec.resize(32, 0);
|
||||||
|
std::borrow::Cow::Owned(vec)
|
||||||
|
} else {
|
||||||
|
std::borrow::Cow::Borrowed(other)
|
||||||
|
};
|
||||||
|
|
||||||
|
let this = resolve_argument(this, deployed_contracts, chain_state_provider)?;
|
||||||
|
let other = U256::from_be_slice(&other);
|
||||||
|
if this != other {
|
||||||
|
return Ok(false);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
Ok(true)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Input {
|
impl Input {
|
||||||
@@ -250,7 +319,11 @@ impl Input {
|
|||||||
chain_state_provider: &impl EthereumNode,
|
chain_state_provider: &impl EthereumNode,
|
||||||
) -> anyhow::Result<TransactionRequest> {
|
) -> anyhow::Result<TransactionRequest> {
|
||||||
let input_data = self.encoded_input(deployed_contracts, chain_state_provider)?;
|
let input_data = self.encoded_input(deployed_contracts, chain_state_provider)?;
|
||||||
let transaction_request = TransactionRequest::default();
|
let transaction_request = TransactionRequest::default().from(self.caller).value(
|
||||||
|
self.value
|
||||||
|
.map(|value| value.into_inner())
|
||||||
|
.unwrap_or_default(),
|
||||||
|
);
|
||||||
match self.method {
|
match self.method {
|
||||||
Method::Deployer => Ok(transaction_request.with_deploy_code(input_data)),
|
Method::Deployer => Ok(transaction_request.with_deploy_code(input_data)),
|
||||||
_ => Ok(transaction_request
|
_ => Ok(transaction_request
|
||||||
@@ -269,12 +342,14 @@ impl Input {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn default_instance() -> ContractInstance {
|
fn default_instance() -> ContractInstance {
|
||||||
ContractInstance::new_from("Test")
|
ContractInstance::new_from("Test")
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn default_caller() -> Address {
|
pub const fn default_caller() -> Address {
|
||||||
"90F8bf6A479f320ead074411a4B0e7944Ea8c9C1".parse().unwrap()
|
Address(FixedBytes(alloy::hex!(
|
||||||
|
"90F8bf6A479f320ead074411a4B0e7944Ea8c9C1"
|
||||||
|
)))
|
||||||
}
|
}
|
||||||
|
|
||||||
/// This function takes in the string calldata argument provided in the JSON input and resolves it
|
/// This function takes in the string calldata argument provided in the JSON input and resolves it
|
||||||
@@ -462,7 +537,7 @@ mod tests {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_encoded_input_address() {
|
fn test_encoded_input_address_with_signature() {
|
||||||
let raw_abi = r#"[
|
let raw_abi = r#"[
|
||||||
{
|
{
|
||||||
"inputs": [{"name": "recipient", "type": "address"}],
|
"inputs": [{"name": "recipient", "type": "address"}],
|
||||||
@@ -483,8 +558,8 @@ mod tests {
|
|||||||
.0;
|
.0;
|
||||||
|
|
||||||
let input: Input = Input {
|
let input: Input = Input {
|
||||||
instance: ContractInstance::new_from("Contract"),
|
instance: "Contract".to_owned().into(),
|
||||||
method: Method::FunctionName("send".to_owned()),
|
method: Method::FunctionName("send(address)".to_owned()),
|
||||||
calldata: Calldata::Compound(vec![
|
calldata: Calldata::Compound(vec![
|
||||||
"0x1000000000000000000000000000000000000001".to_string(),
|
"0x1000000000000000000000000000000000000001".to_string(),
|
||||||
]),
|
]),
|
||||||
@@ -509,7 +584,7 @@ mod tests {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_encoded_input_address_with_signature() {
|
fn test_encoded_input_address() {
|
||||||
let raw_abi = r#"[
|
let raw_abi = r#"[
|
||||||
{
|
{
|
||||||
"inputs": [{"name": "recipient", "type": "address"}],
|
"inputs": [{"name": "recipient", "type": "address"}],
|
||||||
@@ -531,7 +606,7 @@ mod tests {
|
|||||||
|
|
||||||
let input: Input = Input {
|
let input: Input = Input {
|
||||||
instance: ContractInstance::new_from("Contract"),
|
instance: ContractInstance::new_from("Contract"),
|
||||||
method: Method::FunctionName("send(address)".to_owned()),
|
method: Method::FunctionName("send".to_owned()),
|
||||||
calldata: Calldata::Compound(vec![
|
calldata: Calldata::Compound(vec![
|
||||||
"0x1000000000000000000000000000000000000001".to_string(),
|
"0x1000000000000000000000000000000000000001".to_string(),
|
||||||
]),
|
]),
|
||||||
|
|||||||
@@ -44,6 +44,7 @@ impl Deref for MetadataFile {
|
|||||||
|
|
||||||
#[derive(Debug, Default, Deserialize, Clone, Eq, PartialEq)]
|
#[derive(Debug, Default, Deserialize, Clone, Eq, PartialEq)]
|
||||||
pub struct Metadata {
|
pub struct Metadata {
|
||||||
|
pub targets: Option<Vec<String>>,
|
||||||
pub cases: Vec<Case>,
|
pub cases: Vec<Case>,
|
||||||
pub contracts: Option<BTreeMap<ContractInstance, ContractPathAndIdentifier>>,
|
pub contracts: Option<BTreeMap<ContractInstance, ContractPathAndIdentifier>>,
|
||||||
// TODO: Convert into wrapper types for clarity.
|
// TODO: Convert into wrapper types for clarity.
|
||||||
|
|||||||
@@ -9,6 +9,7 @@ use tokio::{
|
|||||||
runtime::Builder,
|
runtime::Builder,
|
||||||
sync::{mpsc::UnboundedSender, oneshot},
|
sync::{mpsc::UnboundedSender, oneshot},
|
||||||
};
|
};
|
||||||
|
use tracing::Instrument;
|
||||||
|
|
||||||
/// A blocking async executor.
|
/// A blocking async executor.
|
||||||
///
|
///
|
||||||
@@ -63,6 +64,11 @@ impl BlockingExecutor {
|
|||||||
let (tx, mut rx) = tokio::sync::mpsc::unbounded_channel::<TaskMessage>();
|
let (tx, mut rx) = tokio::sync::mpsc::unbounded_channel::<TaskMessage>();
|
||||||
|
|
||||||
thread::spawn(move || {
|
thread::spawn(move || {
|
||||||
|
tracing::info!(
|
||||||
|
thread_id = ?std::thread::current().id(),
|
||||||
|
"Starting async runtime thread"
|
||||||
|
);
|
||||||
|
|
||||||
let runtime = Builder::new_current_thread()
|
let runtime = Builder::new_current_thread()
|
||||||
.enable_all()
|
.enable_all()
|
||||||
.build()
|
.build()
|
||||||
@@ -107,7 +113,9 @@ impl BlockingExecutor {
|
|||||||
// in the task message. In doing this conversion, we lose some of the type information since
|
// in the task message. In doing this conversion, we lose some of the type information since
|
||||||
// we're converting R => dyn Any. However, we will perform down-casting on the result to
|
// we're converting R => dyn Any. However, we will perform down-casting on the result to
|
||||||
// convert it back into R.
|
// convert it back into R.
|
||||||
let future = Box::pin(async move { Box::new(future.await) as Box<dyn Any + Send> });
|
let future = Box::pin(
|
||||||
|
async move { Box::new(future.await) as Box<dyn Any + Send> }.in_current_span(),
|
||||||
|
);
|
||||||
|
|
||||||
let task = TaskMessage::new(future, response_tx);
|
let task = TaskMessage::new(future, response_tx);
|
||||||
if let Err(error) = STATE.tx.send(task) {
|
if let Err(error) = STATE.tx.send(task) {
|
||||||
|
|||||||
@@ -1,14 +1,9 @@
|
|||||||
use ::core::pin::Pin;
|
|
||||||
|
|
||||||
use alloy::{
|
use alloy::{
|
||||||
consensus::SignableTransaction,
|
network::{Network, TransactionBuilder},
|
||||||
network::{Network, TransactionBuilder, TxSigner},
|
|
||||||
primitives::Address,
|
|
||||||
providers::{
|
providers::{
|
||||||
Provider, SendableTx,
|
Provider, SendableTx,
|
||||||
fillers::{GasFiller, TxFiller},
|
fillers::{GasFiller, TxFiller},
|
||||||
},
|
},
|
||||||
signers::{Signature, local::PrivateKeySigner},
|
|
||||||
transports::TransportResult,
|
transports::TransportResult,
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -81,28 +76,3 @@ where
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// This is a signer that is able to sign transactions for a specific address with another private
|
|
||||||
/// key.
|
|
||||||
pub struct AddressSigner {
|
|
||||||
pub private_key: PrivateKeySigner,
|
|
||||||
pub address: Address,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl TxSigner<Signature> for AddressSigner {
|
|
||||||
fn address(&self) -> Address {
|
|
||||||
self.address
|
|
||||||
}
|
|
||||||
|
|
||||||
fn sign_transaction<'a, 'b, 'c>(
|
|
||||||
&'a self,
|
|
||||||
tx: &'b mut dyn SignableTransaction<Signature>,
|
|
||||||
) -> Pin<Box<dyn Future<Output = Result<Signature, alloy::signers::Error>> + Send + 'c>>
|
|
||||||
where
|
|
||||||
'a: 'c,
|
|
||||||
'b: 'c,
|
|
||||||
Self: 'c,
|
|
||||||
{
|
|
||||||
<PrivateKeySigner as TxSigner<Signature>>::sign_transaction(&self.private_key, tx)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -0,0 +1,5 @@
|
|||||||
|
/// This constant defines how much Wei accounts are pre-seeded with in genesis.
|
||||||
|
///
|
||||||
|
/// We use [`u128::MAX`] here which means that accounts will be given 2^128 - 1 WEI which is
|
||||||
|
/// (2^128 - 1) / 10^18 ETH.
|
||||||
|
pub const INITIAL_BALANCE: u128 = u128::MAX;
|
||||||
+35
-32
@@ -11,9 +11,9 @@ use std::{
|
|||||||
|
|
||||||
use alloy::{
|
use alloy::{
|
||||||
eips::BlockNumberOrTag,
|
eips::BlockNumberOrTag,
|
||||||
genesis::Genesis,
|
genesis::{Genesis, GenesisAccount},
|
||||||
network::{Ethereum, EthereumWallet},
|
network::{Ethereum, EthereumWallet, NetworkWallet},
|
||||||
primitives::{Address, BlockHash, BlockNumber, BlockTimestamp, U256},
|
primitives::{Address, BlockHash, BlockNumber, BlockTimestamp, FixedBytes, U256},
|
||||||
providers::{
|
providers::{
|
||||||
Provider, ProviderBuilder,
|
Provider, ProviderBuilder,
|
||||||
ext::DebugApi,
|
ext::DebugApi,
|
||||||
@@ -29,10 +29,7 @@ use revive_dt_config::Arguments;
|
|||||||
use revive_dt_node_interaction::{BlockingExecutor, EthereumNode};
|
use revive_dt_node_interaction::{BlockingExecutor, EthereumNode};
|
||||||
use tracing::Level;
|
use tracing::Level;
|
||||||
|
|
||||||
use crate::{
|
use crate::{Node, common::FallbackGasFiller, constants::INITIAL_BALANCE};
|
||||||
Node,
|
|
||||||
common::{AddressSigner, FallbackGasFiller},
|
|
||||||
};
|
|
||||||
|
|
||||||
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
||||||
|
|
||||||
@@ -55,10 +52,7 @@ pub struct Instance {
|
|||||||
network_id: u64,
|
network_id: u64,
|
||||||
start_timeout: u64,
|
start_timeout: u64,
|
||||||
wallet: EthereumWallet,
|
wallet: EthereumWallet,
|
||||||
private_key: PrivateKeySigner,
|
|
||||||
nonce_manager: CachedNonceManager,
|
nonce_manager: CachedNonceManager,
|
||||||
additional_callers: Vec<Address>,
|
|
||||||
|
|
||||||
/// This vector stores [`File`] objects that we use for logging which we want to flush when the
|
/// This vector stores [`File`] objects that we use for logging which we want to flush when the
|
||||||
/// node object is dropped. We do not store them in a structured fashion at the moment (in
|
/// node object is dropped. We do not store them in a structured fashion at the moment (in
|
||||||
/// separate fields) as the logic that we need to apply to them is all the same regardless of
|
/// separate fields) as the logic that we need to apply to them is all the same regardless of
|
||||||
@@ -86,25 +80,18 @@ impl Instance {
|
|||||||
create_dir_all(&self.base_directory)?;
|
create_dir_all(&self.base_directory)?;
|
||||||
create_dir_all(&self.logs_directory)?;
|
create_dir_all(&self.logs_directory)?;
|
||||||
|
|
||||||
// Modifying the genesis file so that we get our private key to control the other accounts.
|
|
||||||
let mut genesis = serde_json::from_str::<Genesis>(&genesis)?;
|
let mut genesis = serde_json::from_str::<Genesis>(&genesis)?;
|
||||||
for additional_caller in self.additional_callers.iter() {
|
for signer_address in
|
||||||
let account = genesis.alloc.entry(*additional_caller).or_default();
|
<EthereumWallet as NetworkWallet<Ethereum>>::signer_addresses(&self.wallet)
|
||||||
account.private_key = Some(self.private_key.to_bytes());
|
{
|
||||||
*account = account
|
genesis
|
||||||
.clone()
|
.alloc
|
||||||
.with_balance("1000000000000000000".parse().expect("Can't fail"));
|
.entry(signer_address)
|
||||||
|
.or_insert(GenesisAccount::default().with_balance(U256::from(INITIAL_BALANCE)));
|
||||||
}
|
}
|
||||||
let genesis_path = self.base_directory.join(Self::GENESIS_JSON_FILE);
|
let genesis_path = self.base_directory.join(Self::GENESIS_JSON_FILE);
|
||||||
serde_json::to_writer(File::create(&genesis_path)?, &genesis)?;
|
serde_json::to_writer(File::create(&genesis_path)?, &genesis)?;
|
||||||
|
|
||||||
for additional_caller in self.additional_callers.iter() {
|
|
||||||
self.wallet.register_signer(AddressSigner {
|
|
||||||
private_key: self.private_key.clone(),
|
|
||||||
address: *additional_caller,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
let mut child = Command::new(&self.geth)
|
let mut child = Command::new(&self.geth)
|
||||||
.arg("init")
|
.arg("init")
|
||||||
.arg("--datadir")
|
.arg("--datadir")
|
||||||
@@ -448,11 +435,20 @@ impl EthereumNode for Instance {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Node for Instance {
|
impl Node for Instance {
|
||||||
fn new(config: &Arguments, additional_callers: &[Address]) -> Self {
|
fn new(config: &Arguments) -> Self {
|
||||||
let geth_directory = config.directory().join(Self::BASE_DIRECTORY);
|
let geth_directory = config.directory().join(Self::BASE_DIRECTORY);
|
||||||
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
||||||
let base_directory = geth_directory.join(id.to_string());
|
let base_directory = geth_directory.join(id.to_string());
|
||||||
|
|
||||||
|
let mut wallet = config.wallet();
|
||||||
|
for signer in (1..=config.private_keys_to_add)
|
||||||
|
.map(|id| U256::from(id))
|
||||||
|
.map(|id| id.to_be_bytes::<32>())
|
||||||
|
.map(|id| PrivateKeySigner::from_bytes(&FixedBytes(id)).unwrap())
|
||||||
|
{
|
||||||
|
wallet.register_signer(signer);
|
||||||
|
}
|
||||||
|
|
||||||
Self {
|
Self {
|
||||||
connection_string: base_directory.join(Self::IPC_FILE).display().to_string(),
|
connection_string: base_directory.join(Self::IPC_FILE).display().to_string(),
|
||||||
data_directory: base_directory.join(Self::DATA_DIRECTORY),
|
data_directory: base_directory.join(Self::DATA_DIRECTORY),
|
||||||
@@ -463,13 +459,11 @@ impl Node for Instance {
|
|||||||
handle: None,
|
handle: None,
|
||||||
network_id: config.network_id,
|
network_id: config.network_id,
|
||||||
start_timeout: config.geth_start_timeout,
|
start_timeout: config.geth_start_timeout,
|
||||||
wallet: config.wallet(),
|
wallet,
|
||||||
// We know that we only need to be storing 2 files so we can specify that when creating
|
// We know that we only need to be storing 2 files so we can specify that when creating
|
||||||
// the vector. It's the stdout and stderr of the geth node.
|
// the vector. It's the stdout and stderr of the geth node.
|
||||||
logs_file_to_flush: Vec::with_capacity(2),
|
logs_file_to_flush: Vec::with_capacity(2),
|
||||||
nonce_manager: Default::default(),
|
nonce_manager: Default::default(),
|
||||||
additional_callers: additional_callers.to_vec(),
|
|
||||||
private_key: config.signer(),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -518,6 +512,14 @@ impl Node for Instance {
|
|||||||
.stdout;
|
.stdout;
|
||||||
Ok(String::from_utf8_lossy(&output).into())
|
Ok(String::from_utf8_lossy(&output).into())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[tracing::instrument(skip_all, fields(geth_node_id = self.id))]
|
||||||
|
fn matches_target(&self, targets: Option<&[String]>) -> bool {
|
||||||
|
match targets {
|
||||||
|
None => true,
|
||||||
|
Some(targets) => targets.iter().any(|str| str.as_str() == "evm"),
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Drop for Instance {
|
impl Drop for Instance {
|
||||||
@@ -530,6 +532,7 @@ impl Drop for Instance {
|
|||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
|
|
||||||
use temp_dir::TempDir;
|
use temp_dir::TempDir;
|
||||||
|
|
||||||
use crate::{GENESIS_JSON, Node};
|
use crate::{GENESIS_JSON, Node};
|
||||||
@@ -546,7 +549,7 @@ mod tests {
|
|||||||
|
|
||||||
fn new_node() -> (Instance, TempDir) {
|
fn new_node() -> (Instance, TempDir) {
|
||||||
let (args, temp_dir) = test_config();
|
let (args, temp_dir) = test_config();
|
||||||
let mut node = Instance::new(&args, &[]);
|
let mut node = Instance::new(&args);
|
||||||
node.init(GENESIS_JSON.to_owned())
|
node.init(GENESIS_JSON.to_owned())
|
||||||
.expect("Failed to initialize the node")
|
.expect("Failed to initialize the node")
|
||||||
.spawn_process()
|
.spawn_process()
|
||||||
@@ -556,21 +559,21 @@ mod tests {
|
|||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn init_works() {
|
fn init_works() {
|
||||||
Instance::new(&test_config().0, &[])
|
Instance::new(&test_config().0)
|
||||||
.init(GENESIS_JSON.to_string())
|
.init(GENESIS_JSON.to_string())
|
||||||
.unwrap();
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn spawn_works() {
|
fn spawn_works() {
|
||||||
Instance::new(&test_config().0, &[])
|
Instance::new(&test_config().0)
|
||||||
.spawn(GENESIS_JSON.to_string())
|
.spawn(GENESIS_JSON.to_string())
|
||||||
.unwrap();
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn version_works() {
|
fn version_works() {
|
||||||
let version = Instance::new(&test_config().0, &[]).version().unwrap();
|
let version = Instance::new(&test_config().0).version().unwrap();
|
||||||
assert!(
|
assert!(
|
||||||
version.starts_with("geth version"),
|
version.starts_with("geth version"),
|
||||||
"expected version string, got: '{version}'"
|
"expected version string, got: '{version}'"
|
||||||
|
|||||||
@@ -10,12 +10,14 @@ use std::{
|
|||||||
use alloy::{
|
use alloy::{
|
||||||
consensus::{BlockHeader, TxEnvelope},
|
consensus::{BlockHeader, TxEnvelope},
|
||||||
eips::BlockNumberOrTag,
|
eips::BlockNumberOrTag,
|
||||||
hex,
|
genesis::{Genesis, GenesisAccount},
|
||||||
network::{
|
network::{
|
||||||
Ethereum, EthereumWallet, Network, TransactionBuilder, TransactionBuilderError,
|
Ethereum, EthereumWallet, Network, NetworkWallet, TransactionBuilder,
|
||||||
UnbuiltTransactionError,
|
TransactionBuilderError, UnbuiltTransactionError,
|
||||||
|
},
|
||||||
|
primitives::{
|
||||||
|
Address, B64, B256, BlockHash, BlockNumber, BlockTimestamp, Bloom, Bytes, FixedBytes, U256,
|
||||||
},
|
},
|
||||||
primitives::{Address, B64, B256, BlockHash, BlockNumber, BlockTimestamp, Bloom, Bytes, U256},
|
|
||||||
providers::{
|
providers::{
|
||||||
Provider, ProviderBuilder,
|
Provider, ProviderBuilder,
|
||||||
ext::DebugApi,
|
ext::DebugApi,
|
||||||
@@ -26,6 +28,7 @@ use alloy::{
|
|||||||
eth::{Block, Header, Transaction},
|
eth::{Block, Header, Transaction},
|
||||||
trace::geth::{DiffMode, GethDebugTracingOptions, PreStateConfig, PreStateFrame},
|
trace::geth::{DiffMode, GethDebugTracingOptions, PreStateConfig, PreStateFrame},
|
||||||
},
|
},
|
||||||
|
signers::local::PrivateKeySigner,
|
||||||
};
|
};
|
||||||
use serde::{Deserialize, Serialize};
|
use serde::{Deserialize, Serialize};
|
||||||
use serde_json::{Value as JsonValue, json};
|
use serde_json::{Value as JsonValue, json};
|
||||||
@@ -36,7 +39,7 @@ use tracing::Level;
|
|||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_dt_node_interaction::{BlockingExecutor, EthereumNode};
|
use revive_dt_node_interaction::{BlockingExecutor, EthereumNode};
|
||||||
|
|
||||||
use crate::{Node, common::FallbackGasFiller};
|
use crate::{Node, common::FallbackGasFiller, constants::INITIAL_BALANCE};
|
||||||
|
|
||||||
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
static NODE_COUNT: AtomicU32 = AtomicU32::new(0);
|
||||||
|
|
||||||
@@ -123,7 +126,18 @@ impl KitchensinkNode {
|
|||||||
None
|
None
|
||||||
})
|
})
|
||||||
.collect();
|
.collect();
|
||||||
let mut eth_balances = self.extract_balance_from_genesis_file(genesis)?;
|
let mut eth_balances = {
|
||||||
|
let mut genesis = serde_json::from_str::<Genesis>(genesis)?;
|
||||||
|
for signer_address in
|
||||||
|
<EthereumWallet as NetworkWallet<Ethereum>>::signer_addresses(&self.wallet)
|
||||||
|
{
|
||||||
|
genesis
|
||||||
|
.alloc
|
||||||
|
.entry(signer_address)
|
||||||
|
.or_insert(GenesisAccount::default().with_balance(U256::from(INITIAL_BALANCE)));
|
||||||
|
}
|
||||||
|
self.extract_balance_from_genesis_file(&genesis)?
|
||||||
|
};
|
||||||
merged_balances.append(&mut eth_balances);
|
merged_balances.append(&mut eth_balances);
|
||||||
|
|
||||||
chainspec_json["genesis"]["runtimeGenesis"]["patch"]["balances"]["balances"] =
|
chainspec_json["genesis"]["runtimeGenesis"]["patch"]["balances"]["balances"] =
|
||||||
@@ -237,42 +251,27 @@ impl KitchensinkNode {
|
|||||||
#[tracing::instrument(skip_all, fields(kitchensink_node_id = self.id))]
|
#[tracing::instrument(skip_all, fields(kitchensink_node_id = self.id))]
|
||||||
fn extract_balance_from_genesis_file(
|
fn extract_balance_from_genesis_file(
|
||||||
&self,
|
&self,
|
||||||
genesis_str: &str,
|
genesis: &Genesis,
|
||||||
) -> anyhow::Result<Vec<(String, u128)>> {
|
) -> anyhow::Result<Vec<(String, u128)>> {
|
||||||
let genesis_json: JsonValue = serde_json::from_str(genesis_str)?;
|
genesis
|
||||||
let alloc = genesis_json
|
.alloc
|
||||||
.get("alloc")
|
.iter()
|
||||||
.and_then(|a| a.as_object())
|
.try_fold(Vec::new(), |mut vec, (address, acc)| {
|
||||||
.ok_or_else(|| anyhow::anyhow!("Missing 'alloc' in genesis"))?;
|
let substrate_address = Self::eth_to_substrate_address(address);
|
||||||
|
let balance = acc.balance.try_into()?;
|
||||||
let mut balances = Vec::new();
|
vec.push((substrate_address, balance));
|
||||||
for (eth_addr, obj) in alloc.iter() {
|
Ok(vec)
|
||||||
let balance_str = obj.get("balance").and_then(|b| b.as_str()).unwrap_or("0");
|
})
|
||||||
let balance = if balance_str.starts_with("0x") {
|
|
||||||
u128::from_str_radix(balance_str.trim_start_matches("0x"), 16)?
|
|
||||||
} else {
|
|
||||||
balance_str.parse::<u128>()?
|
|
||||||
};
|
|
||||||
let substrate_addr = Self::eth_to_substrate_address(eth_addr)?;
|
|
||||||
balances.push((substrate_addr.clone(), balance));
|
|
||||||
}
|
|
||||||
Ok(balances)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn eth_to_substrate_address(eth_addr: &str) -> anyhow::Result<String> {
|
fn eth_to_substrate_address(address: &Address) -> String {
|
||||||
let eth_bytes = hex::decode(eth_addr.trim_start_matches("0x"))?;
|
let eth_bytes = address.0.0;
|
||||||
if eth_bytes.len() != 20 {
|
|
||||||
anyhow::bail!(
|
|
||||||
"Invalid Ethereum address length: expected 20 bytes, got {}",
|
|
||||||
eth_bytes.len()
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
let mut padded = [0xEEu8; 32];
|
let mut padded = [0xEEu8; 32];
|
||||||
padded[..20].copy_from_slice(ð_bytes);
|
padded[..20].copy_from_slice(ð_bytes);
|
||||||
|
|
||||||
let account_id = AccountId32::from(padded);
|
let account_id = AccountId32::from(padded);
|
||||||
Ok(account_id.to_ss58check())
|
account_id.to_ss58check()
|
||||||
}
|
}
|
||||||
|
|
||||||
fn wait_ready(logs_file_path: &Path, marker: &str, timeout: Duration) -> anyhow::Result<()> {
|
fn wait_ready(logs_file_path: &Path, marker: &str, timeout: Duration) -> anyhow::Result<()> {
|
||||||
@@ -507,18 +506,27 @@ impl EthereumNode for KitchensinkNode {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Node for KitchensinkNode {
|
impl Node for KitchensinkNode {
|
||||||
fn new(config: &Arguments, _additional_callers: &[Address]) -> Self {
|
fn new(config: &Arguments) -> Self {
|
||||||
let kitchensink_directory = config.directory().join(Self::BASE_DIRECTORY);
|
let kitchensink_directory = config.directory().join(Self::BASE_DIRECTORY);
|
||||||
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
let id = NODE_COUNT.fetch_add(1, Ordering::SeqCst);
|
||||||
let base_directory = kitchensink_directory.join(id.to_string());
|
let base_directory = kitchensink_directory.join(id.to_string());
|
||||||
let logs_directory = base_directory.join(Self::LOGS_DIRECTORY);
|
let logs_directory = base_directory.join(Self::LOGS_DIRECTORY);
|
||||||
|
|
||||||
|
let mut wallet = config.wallet();
|
||||||
|
for signer in (1..=config.private_keys_to_add)
|
||||||
|
.map(|id| U256::from(id))
|
||||||
|
.map(|id| id.to_be_bytes::<32>())
|
||||||
|
.map(|id| PrivateKeySigner::from_bytes(&FixedBytes(id)).unwrap())
|
||||||
|
{
|
||||||
|
wallet.register_signer(signer);
|
||||||
|
}
|
||||||
|
|
||||||
Self {
|
Self {
|
||||||
id,
|
id,
|
||||||
substrate_binary: config.kitchensink.clone(),
|
substrate_binary: config.kitchensink.clone(),
|
||||||
eth_proxy_binary: config.eth_proxy.clone(),
|
eth_proxy_binary: config.eth_proxy.clone(),
|
||||||
rpc_url: String::new(),
|
rpc_url: String::new(),
|
||||||
wallet: config.wallet(),
|
wallet,
|
||||||
base_directory,
|
base_directory,
|
||||||
logs_directory,
|
logs_directory,
|
||||||
process_substrate: None,
|
process_substrate: None,
|
||||||
@@ -579,6 +587,14 @@ impl Node for KitchensinkNode {
|
|||||||
.stdout;
|
.stdout;
|
||||||
Ok(String::from_utf8_lossy(&output).into())
|
Ok(String::from_utf8_lossy(&output).into())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[tracing::instrument(skip_all, fields(kitchensink_node_id = self.id))]
|
||||||
|
fn matches_target(&self, targets: Option<&[String]>) -> bool {
|
||||||
|
match targets {
|
||||||
|
None => true,
|
||||||
|
Some(targets) => targets.iter().any(|str| str.as_str() == "pvm"),
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Drop for KitchensinkNode {
|
impl Drop for KitchensinkNode {
|
||||||
@@ -634,6 +650,12 @@ impl TransactionBuilder<KitchenSinkNetwork> for <Ethereum as Network>::Transacti
|
|||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn take_nonce(&mut self) -> Option<u64> {
|
||||||
|
<<Ethereum as Network>::TransactionRequest as TransactionBuilder<Ethereum>>::take_nonce(
|
||||||
|
self,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
fn input(&self) -> Option<&alloy::primitives::Bytes> {
|
fn input(&self) -> Option<&alloy::primitives::Bytes> {
|
||||||
<<Ethereum as Network>::TransactionRequest as TransactionBuilder<Ethereum>>::input(self)
|
<<Ethereum as Network>::TransactionRequest as TransactionBuilder<Ethereum>>::input(self)
|
||||||
}
|
}
|
||||||
@@ -814,12 +836,6 @@ impl TransactionBuilder<KitchenSinkNetwork> for <Ethereum as Network>::Transacti
|
|||||||
> {
|
> {
|
||||||
Ok(wallet.sign_request(self).await?)
|
Ok(wallet.sign_request(self).await?)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn take_nonce(&mut self) -> Option<u64> {
|
|
||||||
<<Ethereum as Network>::TransactionRequest as TransactionBuilder<Ethereum>>::take_nonce(
|
|
||||||
self,
|
|
||||||
)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Clone, Debug, PartialEq, Eq, Hash, Serialize, Deserialize)]
|
#[derive(Clone, Debug, PartialEq, Eq, Hash, Serialize, Deserialize)]
|
||||||
@@ -1060,7 +1076,7 @@ mod tests {
|
|||||||
let _guard = NODE_START_MUTEX.lock().unwrap();
|
let _guard = NODE_START_MUTEX.lock().unwrap();
|
||||||
|
|
||||||
let (args, temp_dir) = test_config();
|
let (args, temp_dir) = test_config();
|
||||||
let mut node = KitchensinkNode::new(&args, &[]);
|
let mut node = KitchensinkNode::new(&args);
|
||||||
node.init(GENESIS_JSON)
|
node.init(GENESIS_JSON)
|
||||||
.expect("Failed to initialize the node")
|
.expect("Failed to initialize the node")
|
||||||
.spawn_process()
|
.spawn_process()
|
||||||
@@ -1115,7 +1131,7 @@ mod tests {
|
|||||||
}
|
}
|
||||||
"#;
|
"#;
|
||||||
|
|
||||||
let mut dummy_node = KitchensinkNode::new(&test_config().0, &[]);
|
let mut dummy_node = KitchensinkNode::new(&test_config().0);
|
||||||
|
|
||||||
// Call `init()`
|
// Call `init()`
|
||||||
dummy_node.init(genesis_content).expect("init failed");
|
dummy_node.init(genesis_content).expect("init failed");
|
||||||
@@ -1129,12 +1145,12 @@ mod tests {
|
|||||||
let contents = fs::read_to_string(&final_chainspec_path).expect("Failed to read chainspec");
|
let contents = fs::read_to_string(&final_chainspec_path).expect("Failed to read chainspec");
|
||||||
|
|
||||||
// Validate that the Substrate addresses derived from the Ethereum addresses are in the file
|
// Validate that the Substrate addresses derived from the Ethereum addresses are in the file
|
||||||
let first_eth_addr =
|
let first_eth_addr = KitchensinkNode::eth_to_substrate_address(
|
||||||
KitchensinkNode::eth_to_substrate_address("90F8bf6A479f320ead074411a4B0e7944Ea8c9C1")
|
&"90F8bf6A479f320ead074411a4B0e7944Ea8c9C1".parse().unwrap(),
|
||||||
.unwrap();
|
);
|
||||||
let second_eth_addr =
|
let second_eth_addr = KitchensinkNode::eth_to_substrate_address(
|
||||||
KitchensinkNode::eth_to_substrate_address("Ab8483F64d9C6d1EcF9b849Ae677dD3315835cb2")
|
&"Ab8483F64d9C6d1EcF9b849Ae677dD3315835cb2".parse().unwrap(),
|
||||||
.unwrap();
|
);
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
contents.contains(&first_eth_addr),
|
contents.contains(&first_eth_addr),
|
||||||
@@ -1159,10 +1175,10 @@ mod tests {
|
|||||||
}
|
}
|
||||||
"#;
|
"#;
|
||||||
|
|
||||||
let node = KitchensinkNode::new(&test_config().0, &[]);
|
let node = KitchensinkNode::new(&test_config().0);
|
||||||
|
|
||||||
let result = node
|
let result = node
|
||||||
.extract_balance_from_genesis_file(genesis_json)
|
.extract_balance_from_genesis_file(&serde_json::from_str(genesis_json).unwrap())
|
||||||
.unwrap();
|
.unwrap();
|
||||||
|
|
||||||
let result_map: std::collections::HashMap<_, _> = result.into_iter().collect();
|
let result_map: std::collections::HashMap<_, _> = result.into_iter().collect();
|
||||||
@@ -1192,7 +1208,7 @@ mod tests {
|
|||||||
];
|
];
|
||||||
|
|
||||||
for eth_addr in eth_addresses {
|
for eth_addr in eth_addresses {
|
||||||
let ss58 = KitchensinkNode::eth_to_substrate_address(eth_addr).unwrap();
|
let ss58 = KitchensinkNode::eth_to_substrate_address(ð_addr.parse().unwrap());
|
||||||
|
|
||||||
println!("Ethereum: {eth_addr} -> Substrate SS58: {ss58}");
|
println!("Ethereum: {eth_addr} -> Substrate SS58: {ss58}");
|
||||||
}
|
}
|
||||||
@@ -1220,7 +1236,7 @@ mod tests {
|
|||||||
];
|
];
|
||||||
|
|
||||||
for (eth_addr, expected_ss58) in cases {
|
for (eth_addr, expected_ss58) in cases {
|
||||||
let result = KitchensinkNode::eth_to_substrate_address(eth_addr).unwrap();
|
let result = KitchensinkNode::eth_to_substrate_address(ð_addr.parse().unwrap());
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
result, expected_ss58,
|
result, expected_ss58,
|
||||||
"Mismatch for Ethereum address {eth_addr}"
|
"Mismatch for Ethereum address {eth_addr}"
|
||||||
@@ -1232,7 +1248,7 @@ mod tests {
|
|||||||
fn spawn_works() {
|
fn spawn_works() {
|
||||||
let (config, _temp_dir) = test_config();
|
let (config, _temp_dir) = test_config();
|
||||||
|
|
||||||
let mut node = KitchensinkNode::new(&config, &[]);
|
let mut node = KitchensinkNode::new(&config);
|
||||||
node.spawn(GENESIS_JSON.to_string()).unwrap();
|
node.spawn(GENESIS_JSON.to_string()).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1240,7 +1256,7 @@ mod tests {
|
|||||||
fn version_works() {
|
fn version_works() {
|
||||||
let (config, _temp_dir) = test_config();
|
let (config, _temp_dir) = test_config();
|
||||||
|
|
||||||
let node = KitchensinkNode::new(&config, &[]);
|
let node = KitchensinkNode::new(&config);
|
||||||
let version = node.version().unwrap();
|
let version = node.version().unwrap();
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
@@ -1253,7 +1269,7 @@ mod tests {
|
|||||||
fn eth_rpc_version_works() {
|
fn eth_rpc_version_works() {
|
||||||
let (config, _temp_dir) = test_config();
|
let (config, _temp_dir) = test_config();
|
||||||
|
|
||||||
let node = KitchensinkNode::new(&config, &[]);
|
let node = KitchensinkNode::new(&config);
|
||||||
let version = node.eth_rpc_version().unwrap();
|
let version = node.eth_rpc_version().unwrap();
|
||||||
|
|
||||||
assert!(
|
assert!(
|
||||||
|
|||||||
@@ -1,10 +1,10 @@
|
|||||||
//! This crate implements the testing nodes.
|
//! This crate implements the testing nodes.
|
||||||
|
|
||||||
use alloy::primitives::Address;
|
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
use revive_dt_node_interaction::EthereumNode;
|
use revive_dt_node_interaction::EthereumNode;
|
||||||
|
|
||||||
pub mod common;
|
pub mod common;
|
||||||
|
pub mod constants;
|
||||||
pub mod geth;
|
pub mod geth;
|
||||||
pub mod kitchensink;
|
pub mod kitchensink;
|
||||||
pub mod pool;
|
pub mod pool;
|
||||||
@@ -15,7 +15,7 @@ pub const GENESIS_JSON: &str = include_str!("../../../genesis.json");
|
|||||||
/// An abstract interface for testing nodes.
|
/// An abstract interface for testing nodes.
|
||||||
pub trait Node: EthereumNode {
|
pub trait Node: EthereumNode {
|
||||||
/// Create a new uninitialized instance.
|
/// Create a new uninitialized instance.
|
||||||
fn new(config: &Arguments, additional_callers: &[Address]) -> Self;
|
fn new(config: &Arguments) -> Self;
|
||||||
|
|
||||||
/// Spawns a node configured according to the genesis json.
|
/// Spawns a node configured according to the genesis json.
|
||||||
///
|
///
|
||||||
@@ -32,4 +32,8 @@ pub trait Node: EthereumNode {
|
|||||||
|
|
||||||
/// Returns the node version.
|
/// Returns the node version.
|
||||||
fn version(&self) -> anyhow::Result<String>;
|
fn version(&self) -> anyhow::Result<String>;
|
||||||
|
|
||||||
|
/// Given a list of targets from the metadata file, this function determines if the metadata
|
||||||
|
/// file can be ran on this node or not.
|
||||||
|
fn matches_target(&self, targets: Option<&[String]>) -> bool;
|
||||||
}
|
}
|
||||||
|
|||||||
+4
-12
@@ -6,7 +6,6 @@ use std::{
|
|||||||
thread,
|
thread,
|
||||||
};
|
};
|
||||||
|
|
||||||
use alloy::primitives::Address;
|
|
||||||
use anyhow::Context;
|
use anyhow::Context;
|
||||||
use revive_dt_config::Arguments;
|
use revive_dt_config::Arguments;
|
||||||
|
|
||||||
@@ -24,7 +23,7 @@ where
|
|||||||
T: Node + Send + 'static,
|
T: Node + Send + 'static,
|
||||||
{
|
{
|
||||||
/// Create a new Pool. This will start as many nodes as there are workers in `config`.
|
/// Create a new Pool. This will start as many nodes as there are workers in `config`.
|
||||||
pub fn new(config: &Arguments, additional_callers: &[Address]) -> anyhow::Result<Self> {
|
pub fn new(config: &Arguments) -> anyhow::Result<Self> {
|
||||||
let nodes = config.workers;
|
let nodes = config.workers;
|
||||||
let genesis = read_to_string(&config.genesis_file).context(format!(
|
let genesis = read_to_string(&config.genesis_file).context(format!(
|
||||||
"can not read genesis file: {}",
|
"can not read genesis file: {}",
|
||||||
@@ -35,10 +34,7 @@ where
|
|||||||
for _ in 0..nodes {
|
for _ in 0..nodes {
|
||||||
let config = config.clone();
|
let config = config.clone();
|
||||||
let genesis = genesis.clone();
|
let genesis = genesis.clone();
|
||||||
let additional_callers = additional_callers.to_vec();
|
handles.push(thread::spawn(move || spawn_node::<T>(&config, genesis)));
|
||||||
handles.push(thread::spawn(move || {
|
|
||||||
spawn_node::<T>(&config, genesis, &additional_callers)
|
|
||||||
}));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
let mut nodes = Vec::with_capacity(nodes);
|
let mut nodes = Vec::with_capacity(nodes);
|
||||||
@@ -64,12 +60,8 @@ where
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn spawn_node<T: Node + Send>(
|
fn spawn_node<T: Node + Send>(args: &Arguments, genesis: String) -> anyhow::Result<T> {
|
||||||
args: &Arguments,
|
let mut node = T::new(args);
|
||||||
genesis: String,
|
|
||||||
additional_callers: &[Address],
|
|
||||||
) -> anyhow::Result<T> {
|
|
||||||
let mut node = T::new(args, additional_callers);
|
|
||||||
tracing::info!("starting node: {}", node.connection_string());
|
tracing::info!("starting node: {}", node.connection_string());
|
||||||
node.spawn(genesis)?;
|
node.spawn(genesis)?;
|
||||||
Ok(node)
|
Ok(node)
|
||||||
|
|||||||
+1
-1
@@ -35,7 +35,7 @@
|
|||||||
"timestamp": "0x00",
|
"timestamp": "0x00",
|
||||||
"alloc": {
|
"alloc": {
|
||||||
"90F8bf6A479f320ead074411a4B0e7944Ea8c9C1": {
|
"90F8bf6A479f320ead074411a4B0e7944Ea8c9C1": {
|
||||||
"balance": "1000000000000000000"
|
"balance": "10000000000000000000000"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
Reference in New Issue
Block a user