Compare commits

..

9 Commits

Author SHA1 Message Date
Omar Abdulla 0a07a93475 Resolve merge conflicts 2026-01-16 19:21:26 +03:00
Omar Abdulla c0d330d035 Update report processor to only include failures 2026-01-16 19:18:28 +03:00
Omar Abdulla 87dd5edeb1 Update the format of the expectations file 2026-01-15 17:45:50 +03:00
Omar Abdulla 996a64d500 Fix CI issue with the paths 2026-01-15 16:34:28 +03:00
Omar Abdulla 2ce575bf72 Fix the path of the workdir in CI 2026-01-15 16:14:03 +03:00
Omar Abdulla 659263b8c5 Fix CI 2026-01-15 01:13:57 +03:00
Omar Abdulla e2d626d45b Fix an issue with CI 2026-01-15 00:30:38 +03:00
Omar Abdulla 1231a9b6c5 Add expectations tests to the CI action 2026-01-15 00:21:27 +03:00
Omar Abdulla 73ee78a01e Add a report processing tool 2026-01-14 23:01:04 +03:00
14 changed files with 1141 additions and 1296 deletions
@@ -18,9 +18,10 @@ inputs:
required: false
default: "main"
type: string
resolc-path:
description: "The path of the resolc compiler."
required: true
resolc-version:
description: "The version of resolc to install and use in tests."
required: false
default: "0.5.0"
type: string
use-compilation-caches:
description: "Controls if the compilation caches will be used for the test run or not."
@@ -28,10 +29,6 @@ inputs:
default: true
type: boolean
# Test Execution Arguments
# TODO: We need a better way for people to pass arguments to retester. This way is not very good
# because we need to add support for each argument separately and support defaults and all of that
# perhaps having people pass in a JSON String of the arguments is the better long term solution
# for this.
platform:
description: "The identifier of the platform to run the tests on (e.g., geth-evm-solc, revive-dev-node-revm-solc)"
required: true
@@ -59,6 +56,16 @@ runs:
ref: ${{ inputs['revive-differential-tests-ref'] }}
path: revive-differential-tests
submodules: recursive
- name: Installing the Latest Resolc
shell: bash
if: ${{ runner.os == 'Linux' && runner.arch == 'X64' }}
run: |
VERSION="${{ inputs['resolc-version'] }}"
ASSET_URL="https://github.com/paritytech/revive/releases/download/v$VERSION/resolc-x86_64-unknown-linux-musl"
echo "Downloading resolc v$VERSION from $ASSET_URL"
curl -Lsf --show-error -o resolc "$ASSET_URL"
chmod +x resolc
./resolc --version
- name: Installing Retester
shell: bash
run: ${{ inputs['cargo-command'] }} install --locked --path revive-differential-tests/crates/core
@@ -113,12 +120,11 @@ runs:
--revive-dev-node.path ${{ inputs['polkadot-sdk-path'] }}/target/release/revive-dev-node \
--eth-rpc.path ${{ inputs['polkadot-sdk-path'] }}/target/release/eth-rpc \
--polkadot-omni-node.path ${{ inputs['polkadot-sdk-path'] }}/target/release/polkadot-omni-node \
--resolc.path ${{ inputs['resolc-path'] }} \
--resolc.heap-size 500000 \
--resolc.path ./resolc \
"${OMNI_ARGS[@]}" || true
- name: Generate the expectation file
shell: bash
run: report-processor generate-expectations-file --report-path ./workdir/report.json --output-path ./workdir/expectations.json --remove-prefix ./revive-differential-tests/resolc-compiler-tests --include-status failed
run: report-processor generate-expectations-file --report-path ./workdir/report.json --output-path ./workdir/expectations.json --remove-prefix ./revive-differential-tests/resolc-compiler-tests
- name: Upload the Report to the CI
uses: actions/upload-artifact@b7c566a772e6b6bfb58ed0dc250532a479d7789f
with:
Generated
+1073 -1062
View File
File diff suppressed because it is too large Load Diff
+3 -3
View File
@@ -74,9 +74,9 @@ indexmap = { version = "2.10.0", default-features = false }
itertools = { version = "0.14.0" }
# revive compiler
revive-solc-json-interface = { version = "0.5.0" }
revive-common = { version = "0.3.0" }
revive-differential = { version = "0.3.0" }
revive-solc-json-interface = { git = "https://github.com/paritytech/revive", rev = "3389865af7c3ff6f29a586d82157e8bc573c1a8e" }
revive-common = { git = "https://github.com/paritytech/revive", rev = "3389865af7c3ff6f29a586d82157e8bc573c1a8e" }
revive-differential = { git = "https://github.com/paritytech/revive", rev = "3389865af7c3ff6f29a586d82157e8bc573c1a8e" }
zombienet-sdk = { git = "https://github.com/paritytech/zombienet-sdk.git", rev = "891f6554354ce466abd496366dbf8b4f82141241" }
-6
View File
@@ -32,12 +32,8 @@ pub enum PlatformIdentifier {
/// The Lighthouse Go-ethereum reference full node EVM implementation with the solc compiler.
LighthouseGethEvmSolc,
/// The revive dev node with the PolkaVM backend with the resolc compiler.
#[strum(serialize = "revive-dev-node-polkavm-resolc", serialize = "pez-revive-dev-node-polkavm-resolc")]
#[serde(alias = "pez-revive-dev-node-polkavm-resolc")]
ReviveDevNodePolkavmResolc,
/// The revive dev node with the REVM backend with the solc compiler.
#[strum(serialize = "revive-dev-node-revm-solc", serialize = "pez-revive-dev-node-revm-solc")]
#[serde(alias = "pez-revive-dev-node-revm-solc")]
ReviveDevNodeRevmSolc,
/// A zombienet based Substrate/Polkadot node with the PolkaVM backend with the resolc compiler.
ZombienetPolkavmResolc,
@@ -102,8 +98,6 @@ pub enum NodeIdentifier {
/// The go-ethereum node implementation.
LighthouseGeth,
/// The revive dev node implementation.
#[strum(serialize = "revive-dev-node", serialize = "pez-revive-dev-node")]
#[serde(alias = "pez-revive-dev-node")]
ReviveDevNode,
/// A zombienet spawned nodes
Zombienet,
+25 -73
View File
@@ -12,13 +12,9 @@ use dashmap::DashMap;
use revive_dt_common::types::VersionOrRequirement;
use revive_dt_config::{ResolcConfiguration, SolcConfiguration, WorkingDirectoryConfiguration};
use revive_solc_json_interface::{
PolkaVMDefaultHeapMemorySize, PolkaVMDefaultStackMemorySize, SolcStandardJsonInput,
SolcStandardJsonInputLanguage, SolcStandardJsonInputSettings,
SolcStandardJsonInputSettingsLibraries, SolcStandardJsonInputSettingsMetadata,
SolcStandardJsonInputSettingsOptimizer, SolcStandardJsonInputSettingsPolkaVM,
SolcStandardJsonInputSettingsPolkaVMMemory, SolcStandardJsonInputSettingsSelection,
SolcStandardJsonOutput, standard_json::input::settings::optimizer::Optimizer,
standard_json::input::settings::optimizer::details::Details,
SolcStandardJsonInput, SolcStandardJsonInputLanguage, SolcStandardJsonInputSettings,
SolcStandardJsonInputSettingsOptimizer, SolcStandardJsonInputSettingsSelection,
SolcStandardJsonOutput,
};
use tracing::{Span, field::display};
@@ -29,7 +25,6 @@ use crate::{
use alloy::json_abi::JsonAbi;
use anyhow::{Context as _, Result};
use semver::Version;
use std::collections::BTreeSet;
use tokio::{io::AsyncWriteExt, process::Command as AsyncCommand};
/// A wrapper around the `resolc` binary, emitting PVM-compatible bytecode.
@@ -42,10 +37,6 @@ struct ResolcInner {
solc: Solc,
/// Path to the `resolc` executable
resolc_path: PathBuf,
/// The PVM heap size in bytes.
pvm_heap_size: u32,
/// The PVM stack size in bytes.
pvm_stack_size: u32,
}
impl Resolc {
@@ -72,35 +63,10 @@ impl Resolc {
Self(Arc::new(ResolcInner {
solc,
resolc_path: resolc_configuration.path.clone(),
pvm_heap_size: resolc_configuration
.heap_size
.unwrap_or(PolkaVMDefaultHeapMemorySize),
pvm_stack_size: resolc_configuration
.stack_size
.unwrap_or(PolkaVMDefaultStackMemorySize),
}))
})
.clone())
}
fn polkavm_settings(&self) -> SolcStandardJsonInputSettingsPolkaVM {
SolcStandardJsonInputSettingsPolkaVM::new(
Some(SolcStandardJsonInputSettingsPolkaVMMemory::new(
Some(self.0.pvm_heap_size),
Some(self.0.pvm_stack_size),
)),
false,
)
}
fn inject_polkavm_settings(&self, input: &SolcStandardJsonInput) -> Result<serde_json::Value> {
let mut input_value = serde_json::to_value(input)
.context("Failed to serialize Standard JSON input for resolc")?;
if let Some(settings) = input_value.get_mut("settings") {
settings["polkavm"] = serde_json::to_value(self.polkavm_settings()).unwrap();
}
Ok(input_value)
}
}
impl SolidityCompiler for Resolc {
@@ -155,8 +121,8 @@ impl SolidityCompiler for Resolc {
.collect(),
settings: SolcStandardJsonInputSettings {
evm_version,
libraries: SolcStandardJsonInputSettingsLibraries {
inner: libraries
libraries: Some(
libraries
.into_iter()
.map(|(source_code, libraries_map)| {
(
@@ -170,29 +136,23 @@ impl SolidityCompiler for Resolc {
)
})
.collect(),
},
remappings: BTreeSet::<String>::new(),
output_selection: SolcStandardJsonInputSettingsSelection::new_required(),
),
remappings: None,
output_selection: Some(SolcStandardJsonInputSettingsSelection::new_required()),
via_ir: Some(true),
optimizer: SolcStandardJsonInputSettingsOptimizer::new(
optimization
.unwrap_or(ModeOptimizerSetting::M0)
.optimizations_enabled(),
Optimizer::default_mode(),
Details::disabled(&Version::new(0, 0, 0)),
None,
&Version::new(0, 0, 0),
false,
),
polkavm: self.polkavm_settings(),
metadata: SolcStandardJsonInputSettingsMetadata::default(),
detect_missing_libraries: false,
metadata: None,
polkavm: None,
},
};
// Manually inject polkavm settings since it's marked skip_serializing in the upstream crate
let std_input_json = self.inject_polkavm_settings(&input)?;
Span::current().record(
"json_in",
display(serde_json::to_string(&std_input_json).unwrap()),
);
Span::current().record("json_in", display(serde_json::to_string(&input).unwrap()));
let path = &self.0.resolc_path;
let mut command = AsyncCommand::new(path);
@@ -221,9 +181,8 @@ impl SolidityCompiler for Resolc {
.with_context(|| format!("Failed to spawn resolc at {}", path.display()))?;
let stdin_pipe = child.stdin.as_mut().expect("stdin must be piped");
let serialized_input = serde_json::to_vec(&std_input_json)
let serialized_input = serde_json::to_vec(&input)
.context("Failed to serialize Standard JSON input for resolc")?;
stdin_pipe
.write_all(&serialized_input)
.await
@@ -269,7 +228,7 @@ impl SolidityCompiler for Resolc {
// Detecting if the compiler output contained errors and reporting them through logs and
// errors instead of returning the compiler output that might contain errors.
for error in parsed.errors.iter() {
for error in parsed.errors.iter().flatten() {
if error.severity == "error" {
tracing::error!(
?error,
@@ -281,12 +240,12 @@ impl SolidityCompiler for Resolc {
}
}
if parsed.contracts.is_empty() {
let Some(contracts) = parsed.contracts else {
anyhow::bail!("Unexpected error - resolc output doesn't have a contracts section");
}
};
let mut compiler_output = CompilerOutput::default();
for (source_path, contracts) in parsed.contracts.into_iter() {
for (source_path, contracts) in contracts.into_iter() {
let src_for_msg = source_path.clone();
let source_path = PathBuf::from(source_path)
.canonicalize()
@@ -294,22 +253,15 @@ impl SolidityCompiler for Resolc {
let map = compiler_output.contracts.entry(source_path).or_default();
for (contract_name, contract_information) in contracts.into_iter() {
let Some(bytecode) = contract_information
let bytecode = contract_information
.evm
.and_then(|evm| evm.bytecode.clone())
else {
tracing::debug!(
"Skipping abstract or interface contract {} - no bytecode",
contract_name
);
continue;
};
.context("Unexpected - Contract compiled with resolc has no bytecode")?;
let abi = {
let metadata = &contract_information.metadata;
if metadata.is_null() {
anyhow::bail!("No metadata found for the contract");
}
let metadata = contract_information
.metadata
.as_ref()
.context("No metadata found for the contract")?;
let solc_metadata_str = match metadata {
serde_json::Value::String(solc_metadata_str) => {
solc_metadata_str.as_str()
-12
View File
@@ -800,18 +800,6 @@ pub struct ResolcConfiguration {
/// provided in the user's $PATH.
#[clap(id = "resolc.path", long = "resolc.path", default_value = "resolc")]
pub path: PathBuf,
/// Specifies the PVM heap size in bytes.
///
/// If unspecified, the revive compiler default is used
#[clap(id = "resolc.heap-size", long = "resolc.heap-size")]
pub heap_size: Option<u32>,
/// Specifies the PVM stack size in bytes.
///
/// If unspecified, the revive compiler default is used
#[clap(id = "resolc.stack-size", long = "resolc.stack-size")]
pub stack_size: Option<u32>,
}
/// A set of configuration parameters for Polkadot Parachain.
+9 -9
View File
@@ -482,16 +482,15 @@ where
.context("Failed to find deployment receipt for constructor call"),
Method::Fallback | Method::FunctionName(_) => {
let resolver = self.platform_information.node.resolver().await?;
let mut tx = step
let tx = match step
.as_transaction(resolver.as_ref(), self.default_resolution_context())
.await?;
let gas_overrides = step
.gas_overrides
.get(&self.platform_information.platform.platform_identifier())
.copied()
.unwrap_or_default();
gas_overrides.apply_to::<Ethereum>(&mut tx);
.await
{
Ok(tx) => tx,
Err(err) => {
return Err(err);
}
};
self.platform_information.node.execute_transaction(tx).await
}
@@ -912,6 +911,7 @@ where
.get(contract_instance)
{
info!(
%address,
"Contract instance already deployed."
);
+7 -14
View File
@@ -223,24 +223,17 @@ impl<'a> TestDefinition<'a> {
/// Checks if the platforms all support the desired targets in the metadata file.
fn check_target_compatibility(&self) -> TestCheckFunctionResult {
// The case targets takes presence over the metadata targets.
let Some(targets) = self
.case
.targets
.as_ref()
.or(self.metadata.targets.as_ref())
else {
return Ok(());
};
let mut error_map = indexmap! {
"test_desired_targets" => json!(targets),
"test_desired_targets" => json!(self.metadata.targets.as_ref()),
};
let mut is_allowed = true;
for (_, platform_information) in self.platforms.iter() {
let is_allowed_for_platform =
targets.contains(&platform_information.platform.vm_identifier());
let is_allowed_for_platform = match self.metadata.targets.as_ref() {
None => true,
Some(required_vm_identifiers) => {
required_vm_identifiers.contains(&platform_information.platform.vm_identifier())
}
};
is_allowed &= is_allowed_for_platform;
error_map.insert(
platform_information.platform.platform_identifier().into(),
+2 -8
View File
@@ -1,22 +1,16 @@
use alloy::primitives::{Address, map::HashSet};
use alloy::primitives::Address;
use schemars::JsonSchema;
use serde::{Deserialize, Serialize};
use revive_dt_common::{
macros::define_wrapper_type,
types::{Mode, ParsedMode, VmIdentifier},
types::{Mode, ParsedMode},
};
use crate::steps::*;
#[derive(Debug, Default, Serialize, Deserialize, Clone, Eq, PartialEq, JsonSchema)]
pub struct Case {
/// An optional vector of targets that this Metadata file's cases can be executed on. As an
/// example, if we wish for the metadata file's cases to only be run on PolkaVM then we'd
/// specify a target of "PolkaVM" in here.
#[serde(skip_serializing_if = "Option::is_none")]
pub targets: Option<HashSet<VmIdentifier>>,
/// An optional name of the test case.
#[serde(skip_serializing_if = "Option::is_none")]
pub name: Option<String>,
+1 -2
View File
@@ -8,7 +8,6 @@ use std::{
str::FromStr,
};
use alloy::primitives::map::HashSet;
use schemars::JsonSchema;
use serde::{Deserialize, Serialize};
@@ -84,7 +83,7 @@ pub struct Metadata {
/// example, if we wish for the metadata file's cases to only be run on PolkaVM then we'd
/// specify a target of "PolkaVM" in here.
#[serde(skip_serializing_if = "Option::is_none")]
pub targets: Option<HashSet<VmIdentifier>>,
pub targets: Option<Vec<VmIdentifier>>,
/// A vector of the test cases and workloads contained within the metadata file. This is their
/// primary description.
-63
View File
@@ -1,7 +1,6 @@
use std::{collections::HashMap, fmt::Display, str::FromStr};
use alloy::hex::ToHexExt;
use alloy::network::Network;
use alloy::primitives::{FixedBytes, utils::parse_units};
use alloy::{
eips::BlockNumberOrTag,
@@ -12,7 +11,6 @@ use alloy::{
};
use anyhow::Context as _;
use futures::{FutureExt, StreamExt, TryFutureExt, TryStreamExt, stream};
use revive_dt_common::types::PlatformIdentifier;
use schemars::JsonSchema;
use semver::VersionReq;
use serde::{Deserialize, Serialize};
@@ -154,11 +152,6 @@ pub struct FunctionCallStep {
/// during the execution.
#[serde(skip_serializing_if = "Option::is_none")]
pub variable_assignments: Option<VariableAssignments>,
/// Allows for the test to set a specific value for the various gas parameter for each one of
/// the platforms we support. This is ignored for steps that perform contract deployments.
#[serde(default, skip_serializing_if = "HashMap::is_empty")]
pub gas_overrides: HashMap<PlatformIdentifier, GasOverrides>,
}
/// This represents a balance assertion step where the framework needs to query the balance of some
@@ -972,62 +965,6 @@ impl<'de> Deserialize<'de> for EtherValue {
}
}
#[derive(Clone, Copy, Debug, Default, Serialize, Deserialize, Eq, PartialEq, JsonSchema)]
pub struct GasOverrides {
#[serde(skip_serializing_if = "Option::is_none")]
pub gas_limit: Option<u64>,
#[serde(skip_serializing_if = "Option::is_none")]
pub gas_price: Option<u128>,
#[serde(skip_serializing_if = "Option::is_none")]
pub max_fee_per_gas: Option<u128>,
#[serde(skip_serializing_if = "Option::is_none")]
pub max_priority_fee_per_gas: Option<u128>,
}
impl GasOverrides {
pub fn new() -> Self {
Default::default()
}
pub fn with_gas_limit(mut self, value: impl Into<Option<u64>>) -> Self {
self.gas_limit = value.into();
self
}
pub fn with_gas_price(mut self, value: impl Into<Option<u128>>) -> Self {
self.gas_price = value.into();
self
}
pub fn with_max_fee_per_gas(mut self, value: impl Into<Option<u128>>) -> Self {
self.max_fee_per_gas = value.into();
self
}
pub fn with_max_priority_fee_per_gas(mut self, value: impl Into<Option<u128>>) -> Self {
self.max_priority_fee_per_gas = value.into();
self
}
pub fn apply_to<N: Network>(&self, transaction_request: &mut N::TransactionRequest) {
if let Some(gas_limit) = self.gas_limit {
transaction_request.set_gas_limit(gas_limit);
}
if let Some(gas_price) = self.gas_price {
transaction_request.set_gas_price(gas_price);
}
if let Some(max_fee_per_gas) = self.max_fee_per_gas {
transaction_request.set_max_fee_per_gas(max_fee_per_gas);
}
if let Some(max_priority_fee_per_gas) = self.max_priority_fee_per_gas {
transaction_request.set_max_priority_fee_per_gas(max_priority_fee_per_gas)
}
}
}
#[cfg(test)]
mod tests {
-1
View File
@@ -18,7 +18,6 @@ revive-dt-common = { workspace = true }
anyhow = { workspace = true }
clap = { workspace = true }
strum = { workspace = true }
serde = { workspace = true }
serde_json = { workspace = true }
+4 -32
View File
@@ -1,6 +1,6 @@
use std::{
borrow::Cow,
collections::{BTreeMap, BTreeSet, HashSet},
collections::{BTreeMap, BTreeSet},
fmt::Display,
fs::{File, OpenOptions},
ops::{Deref, DerefMut},
@@ -9,12 +9,11 @@ use std::{
};
use anyhow::{Context as _, Error, Result, bail};
use clap::{Parser, ValueEnum};
use clap::Parser;
use serde::{Deserialize, Serialize, de::DeserializeOwned};
use revive_dt_common::types::{Mode, ParsedTestSpecifier};
use revive_dt_report::{Report, TestCaseStatus};
use strum::EnumString;
fn main() -> Result<()> {
let cli = Cli::try_parse().context("Failed to parse the CLI arguments")?;
@@ -24,14 +23,11 @@ fn main() -> Result<()> {
report_path,
output_path: output_file,
remove_prefix,
include_status,
} => {
let remove_prefix = remove_prefix
.into_iter()
.map(|path| path.canonicalize().context("Failed to canonicalize path"))
.collect::<Result<Vec<_>>>()?;
let include_status =
include_status.map(|value| value.into_iter().collect::<HashSet<_>>());
let expectations = report_path
.execution_information
@@ -77,12 +73,7 @@ fn main() -> Result<()> {
Status::from(status),
)
})
.filter(|(_, status)| {
include_status
.as_ref()
.map(|allowed_status| allowed_status.contains(status))
.unwrap_or(true)
})
.filter(|(_, status)| *status == Status::Failed)
.collect::<Expectations>();
let output_file = OpenOptions::new()
@@ -152,11 +143,6 @@ pub enum Cli {
/// Prefix paths to remove from the paths in the final expectations file.
#[clap(long)]
remove_prefix: Vec<PathBuf>,
/// Controls which test case statuses are included in the generated expectations file. If
/// nothing is specified then it will include all of the test case status.
#[clap(long)]
include_status: Option<Vec<Status>>,
},
/// Compares two expectation files to ensure that they match each other.
@@ -171,21 +157,7 @@ pub enum Cli {
},
}
#[derive(
Clone,
Copy,
Debug,
PartialEq,
Eq,
PartialOrd,
Ord,
Hash,
Serialize,
Deserialize,
ValueEnum,
EnumString,
)]
#[strum(serialize_all = "kebab-case")]
#[derive(Clone, Copy, Debug, PartialEq, Eq, PartialOrd, Ord, Hash, Serialize, Deserialize)]
pub enum Status {
Succeeded,
Failed,