mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-25 21:07:56 +00:00
Reinitialize bridge relay subcommand (#1331)
* reinitialize bridge subcommand * PolkadotToKusama in reinit-bridge
This commit is contained in:
committed by
Bastian Köcher
parent
60f82cf459
commit
e656e0089e
@@ -31,15 +31,16 @@ bp-polkadot = { path = "../../primitives/chain-polkadot" }
|
||||
bp-rialto = { path = "../../primitives/chain-rialto" }
|
||||
bp-rialto-parachain = { path = "../../primitives/chain-rialto-parachain" }
|
||||
bp-rococo = { path = "../../primitives/chain-rococo" }
|
||||
bp-token-swap = { path = "../../primitives/token-swap" }
|
||||
bp-wococo = { path = "../../primitives/chain-wococo" }
|
||||
bp-runtime = { path = "../../primitives/runtime" }
|
||||
bp-token-swap = { path = "../../primitives/token-swap" }
|
||||
bp-westend = { path = "../../primitives/chain-westend" }
|
||||
bp-wococo = { path = "../../primitives/chain-wococo" }
|
||||
bridge-runtime-common = { path = "../../bin/runtime-common" }
|
||||
finality-relay = { path = "../finality" }
|
||||
messages-relay = { path = "../messages" }
|
||||
millau-runtime = { path = "../../bin/millau/runtime" }
|
||||
pallet-bridge-dispatch = { path = "../../modules/dispatch" }
|
||||
pallet-bridge-grandpa = { path = "../../modules/grandpa" }
|
||||
pallet-bridge-messages = { path = "../../modules/messages" }
|
||||
pallet-bridge-token-swap = { path = "../../modules/token-swap" }
|
||||
relay-kusama-client = { path = "../client-kusama" }
|
||||
@@ -73,6 +74,7 @@ polkadot-runtime-common = { git = "https://github.com/paritytech/polkadot", bran
|
||||
polkadot-runtime-parachains = { git = "https://github.com/paritytech/polkadot", branch = "master" }
|
||||
|
||||
[dev-dependencies]
|
||||
bp-test-utils = { path = "../../primitives/test-utils" }
|
||||
hex-literal = "0.3"
|
||||
pallet-bridge-grandpa = { path = "../../modules/grandpa" }
|
||||
sp-keyring = { git = "https://github.com/paritytech/substrate", branch = "master" }
|
||||
|
||||
@@ -35,6 +35,7 @@ pub(crate) mod send_message;
|
||||
mod derive_account;
|
||||
mod init_bridge;
|
||||
mod register_parachain;
|
||||
mod reinit_bridge;
|
||||
mod relay_headers;
|
||||
mod relay_headers_and_messages;
|
||||
mod relay_messages;
|
||||
@@ -71,6 +72,11 @@ pub enum Command {
|
||||
///
|
||||
/// Sends initialization transaction to bootstrap the bridge with current finalized block data.
|
||||
InitBridge(init_bridge::InitBridge),
|
||||
/// Reinitialize on-chain bridge pallet with current header data.
|
||||
///
|
||||
/// Sends all missing mandatory headers to bootstrap the bridge with current finalized block
|
||||
/// data.
|
||||
ReinitBridge(reinit_bridge::ReinitBridge),
|
||||
/// Send custom message over the bridge.
|
||||
///
|
||||
/// Allows interacting with the bridge by sending messages over `Messages` component.
|
||||
@@ -126,6 +132,7 @@ impl Command {
|
||||
Self::RelayMessages(arg) => arg.run().await?,
|
||||
Self::RelayHeadersAndMessages(arg) => arg.run().await?,
|
||||
Self::InitBridge(arg) => arg.run().await?,
|
||||
Self::ReinitBridge(arg) => arg.run().await?,
|
||||
Self::SendMessage(arg) => arg.run().await?,
|
||||
Self::EncodeCall(arg) => arg.run().await?,
|
||||
Self::EncodeMessage(arg) => arg.run().await?,
|
||||
|
||||
@@ -0,0 +1,553 @@
|
||||
// Copyright 2019-2021 Parity Technologies (UK) Ltd.
|
||||
// This file is part of Parity Bridges Common.
|
||||
|
||||
// Parity Bridges Common is free software: you can redistribute it and/or modify
|
||||
// it under the terms of the GNU General Public License as published by
|
||||
// the Free Software Foundation, either version 3 of the License, or
|
||||
// (at your option) any later version.
|
||||
|
||||
// Parity Bridges Common is distributed in the hope that it will be useful,
|
||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
// GNU General Public License for more details.
|
||||
|
||||
// You should have received a copy of the GNU General Public License
|
||||
// along with Parity Bridges Common. If not, see <http://www.gnu.org/licenses/>.
|
||||
|
||||
use crate::{
|
||||
chains::{
|
||||
kusama_headers_to_polkadot::KusamaFinalityToPolkadot,
|
||||
polkadot_headers_to_kusama::PolkadotFinalityToKusama,
|
||||
},
|
||||
cli::{
|
||||
swap_tokens::wait_until_transaction_is_finalized, SourceConnectionParams,
|
||||
TargetConnectionParams, TargetSigningParams,
|
||||
},
|
||||
};
|
||||
use bp_header_chain::justification::GrandpaJustification;
|
||||
use bp_runtime::Chain;
|
||||
use codec::Encode;
|
||||
use finality_relay::{SourceClient, SourceHeader};
|
||||
use frame_support::weights::Weight;
|
||||
use num_traits::One;
|
||||
use pallet_bridge_grandpa::weights::WeightInfo;
|
||||
use relay_substrate_client::{
|
||||
AccountIdOf, BlockNumberOf, Chain as _, Client, Error as SubstrateError, HeaderOf, SignParam,
|
||||
SyncHeader, TransactionEra, TransactionSignScheme, UnsignedTransaction,
|
||||
};
|
||||
use sp_core::{Bytes, Pair};
|
||||
use std::convert::{TryFrom, TryInto};
|
||||
use structopt::StructOpt;
|
||||
use strum::{EnumString, EnumVariantNames, VariantNames};
|
||||
use substrate_relay_helper::{
|
||||
finality_pipeline::SubstrateFinalitySyncPipeline, finality_source::SubstrateFinalitySource,
|
||||
finality_target::SubstrateFinalityTarget, messages_source::read_client_state,
|
||||
TransactionParams,
|
||||
};
|
||||
|
||||
/// Reinitialize bridge pallet.
|
||||
#[derive(Debug, PartialEq, StructOpt)]
|
||||
pub struct ReinitBridge {
|
||||
/// A bridge instance to reinitialize.
|
||||
#[structopt(possible_values = ReinitBridgeName::VARIANTS, case_insensitive = true)]
|
||||
bridge: ReinitBridgeName,
|
||||
#[structopt(flatten)]
|
||||
source: SourceConnectionParams,
|
||||
#[structopt(flatten)]
|
||||
target: TargetConnectionParams,
|
||||
#[structopt(flatten)]
|
||||
target_sign: TargetSigningParams,
|
||||
}
|
||||
|
||||
#[derive(Debug, EnumString, EnumVariantNames, PartialEq)]
|
||||
#[strum(serialize_all = "kebab_case")]
|
||||
/// Bridge to initialize.
|
||||
pub enum ReinitBridgeName {
|
||||
KusamaToPolkadot,
|
||||
PolkadotToKusama,
|
||||
}
|
||||
|
||||
macro_rules! select_bridge {
|
||||
($bridge: expr, $generic: tt) => {
|
||||
match $bridge {
|
||||
ReinitBridgeName::KusamaToPolkadot => {
|
||||
use relay_polkadot_client::runtime;
|
||||
|
||||
type Finality = KusamaFinalityToPolkadot;
|
||||
type Call = runtime::Call;
|
||||
|
||||
fn submit_finality_proof_call(
|
||||
header_and_proof: HeaderAndProof<Finality>,
|
||||
) -> runtime::Call {
|
||||
runtime::Call::BridgeKusamaGrandpa(
|
||||
runtime::BridgeKusamaGrandpaCall::submit_finality_proof(
|
||||
Box::new(header_and_proof.0.into_inner()),
|
||||
header_and_proof.1,
|
||||
),
|
||||
)
|
||||
}
|
||||
|
||||
fn set_pallet_operation_mode_call(operational: bool) -> runtime::Call {
|
||||
runtime::Call::BridgeKusamaGrandpa(
|
||||
runtime::BridgeKusamaGrandpaCall::set_operational(operational),
|
||||
)
|
||||
}
|
||||
|
||||
fn batch_all_call(calls: Vec<Call>) -> runtime::Call {
|
||||
runtime::Call::Utility(runtime::UtilityCall::batch_all(calls))
|
||||
}
|
||||
|
||||
$generic
|
||||
},
|
||||
ReinitBridgeName::PolkadotToKusama => {
|
||||
use relay_kusama_client::runtime;
|
||||
|
||||
type Finality = PolkadotFinalityToKusama;
|
||||
type Call = runtime::Call;
|
||||
|
||||
fn submit_finality_proof_call(
|
||||
header_and_proof: HeaderAndProof<Finality>,
|
||||
) -> runtime::Call {
|
||||
runtime::Call::BridgePolkadotGrandpa(
|
||||
runtime::BridgePolkadotGrandpaCall::submit_finality_proof(
|
||||
Box::new(header_and_proof.0.into_inner()),
|
||||
header_and_proof.1,
|
||||
),
|
||||
)
|
||||
}
|
||||
|
||||
fn set_pallet_operation_mode_call(operational: bool) -> runtime::Call {
|
||||
runtime::Call::BridgePolkadotGrandpa(
|
||||
runtime::BridgePolkadotGrandpaCall::set_operational(operational),
|
||||
)
|
||||
}
|
||||
|
||||
fn batch_all_call(calls: Vec<Call>) -> runtime::Call {
|
||||
runtime::Call::Utility(runtime::UtilityCall::batch_all(calls))
|
||||
}
|
||||
|
||||
$generic
|
||||
},
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
impl ReinitBridge {
|
||||
/// Run the command.
|
||||
pub async fn run(self) -> anyhow::Result<()> {
|
||||
select_bridge!(self.bridge, {
|
||||
type Source = <Finality as SubstrateFinalitySyncPipeline>::SourceChain;
|
||||
type Target = <Finality as SubstrateFinalitySyncPipeline>::TargetChain;
|
||||
|
||||
let source_client = self.source.to_client::<Source>().await?;
|
||||
let target_client = self.target.to_client::<Target>().await?;
|
||||
let target_sign = self.target_sign.to_keypair::<Target>()?;
|
||||
let transaction_params = TransactionParams {
|
||||
signer: target_sign,
|
||||
mortality: self.target_sign.target_transactions_mortality,
|
||||
};
|
||||
|
||||
let finality_source =
|
||||
SubstrateFinalitySource::<Finality>::new(source_client.clone(), None);
|
||||
let finality_target = SubstrateFinalityTarget::<Finality>::new(
|
||||
target_client.clone(),
|
||||
transaction_params.clone(),
|
||||
);
|
||||
|
||||
// this subcommand assumes that the pallet at the target chain is halted
|
||||
ensure_pallet_operating_mode(&finality_target, false).await?;
|
||||
|
||||
// we can't call `finality_target.best_finalized_source_block_id()`, because pallet is
|
||||
// halted and the call will fail => just use what it uses internally
|
||||
let current_number =
|
||||
best_source_block_number_at_target::<Finality>(&target_client).await?;
|
||||
let target_number = finality_source.best_finalized_block_number().await?;
|
||||
log::info!(
|
||||
target: "bridge",
|
||||
"Best finalized {} header: at {}: {}, at {}: {}",
|
||||
Source::NAME,
|
||||
Source::NAME,
|
||||
target_number,
|
||||
Target::NAME,
|
||||
current_number,
|
||||
);
|
||||
|
||||
// prepare list of mandatory headers from the range `(current_number; target_number]`
|
||||
let headers_to_submit = find_mandatory_headers_in_range(
|
||||
&finality_source,
|
||||
(current_number + 1, target_number),
|
||||
)
|
||||
.await?;
|
||||
let latest_andatory_header_number = headers_to_submit.last().map(|(h, _)| h.number());
|
||||
log::info!(
|
||||
target: "bridge",
|
||||
"Missing {} mandatory {} headers at {}",
|
||||
headers_to_submit.len(),
|
||||
Source::NAME,
|
||||
Target::NAME,
|
||||
);
|
||||
|
||||
// split all mandatory headers into batches
|
||||
let headers_batches =
|
||||
make_mandatory_headers_batches::<Finality, _>(headers_to_submit, |(_, proof)| {
|
||||
// we don't have an access to the Kusama/Polkadot chain runtimes here, so we'll
|
||||
// be using Millau weights. It isn't super-critical, unless real weights are
|
||||
// magnitude higher or so
|
||||
pallet_bridge_grandpa::weights::MillauWeight::<millau_runtime::Runtime>::submit_finality_proof(
|
||||
proof.commit.precommits.len().try_into().unwrap_or(u32::MAX),
|
||||
proof.votes_ancestries.len().try_into().unwrap_or(u32::MAX),
|
||||
)
|
||||
});
|
||||
log::info!(
|
||||
target: "bridge",
|
||||
"We're going to submit {} transactions to {} node",
|
||||
headers_batches.len(),
|
||||
Target::NAME,
|
||||
);
|
||||
|
||||
// each batch is submitted as a separate transaction
|
||||
let signer_account_id: AccountIdOf<Target> = transaction_params.signer.public().into();
|
||||
let genesis_hash = *target_client.genesis_hash();
|
||||
let (spec_version, transaction_version) =
|
||||
target_client.simple_runtime_version().await?;
|
||||
let last_batch_index = headers_batches.len() - 1;
|
||||
for (i, headers_batch) in headers_batches.into_iter().enumerate() {
|
||||
let is_last_batch = i == last_batch_index;
|
||||
let expected_number =
|
||||
headers_batch.last().expect("all batches are non-empty").0.number();
|
||||
let transaction_params = transaction_params.clone();
|
||||
log::info!(
|
||||
target: "bridge",
|
||||
"Going to submit transaction that updates best {} header at {} to {}",
|
||||
Source::NAME,
|
||||
Target::NAME,
|
||||
expected_number,
|
||||
);
|
||||
|
||||
// prepare `batch_all` call
|
||||
let mut batch_calls = Vec::with_capacity(headers_batch.len() + 2);
|
||||
// the first call is always resumes pallet operation
|
||||
batch_calls.push(set_pallet_operation_mode_call(true));
|
||||
// followed by submit-finality-proofs calls
|
||||
for header_and_proof in headers_batch {
|
||||
batch_calls.push(submit_finality_proof_call(header_and_proof));
|
||||
}
|
||||
// if it isn't the last batch, we shall halt pallet again
|
||||
if !is_last_batch {
|
||||
batch_calls.push(set_pallet_operation_mode_call(false));
|
||||
}
|
||||
let submit_batch_call = batch_all_call(batch_calls);
|
||||
|
||||
let batch_transaction_events = target_client
|
||||
.submit_and_watch_signed_extrinsic(
|
||||
signer_account_id.clone(),
|
||||
move |best_block_id, transaction_nonce| {
|
||||
Ok(Bytes(
|
||||
Target::sign_transaction(SignParam {
|
||||
spec_version,
|
||||
transaction_version,
|
||||
genesis_hash,
|
||||
signer: transaction_params.signer.clone(),
|
||||
era: TransactionEra::new(
|
||||
best_block_id,
|
||||
transaction_params.mortality,
|
||||
),
|
||||
unsigned: UnsignedTransaction::new(
|
||||
submit_batch_call.into(),
|
||||
transaction_nonce,
|
||||
),
|
||||
})?
|
||||
.encode(),
|
||||
))
|
||||
},
|
||||
)
|
||||
.await?;
|
||||
wait_until_transaction_is_finalized::<Target>(batch_transaction_events).await?;
|
||||
|
||||
// verify that the best finalized header at target has been updated
|
||||
let current_number =
|
||||
best_source_block_number_at_target::<Finality>(&target_client).await?;
|
||||
if current_number != expected_number {
|
||||
return Err(anyhow::format_err!(
|
||||
"Transaction has failed to update best {} header at {} to {}. It is {}",
|
||||
Source::NAME,
|
||||
Target::NAME,
|
||||
expected_number,
|
||||
current_number,
|
||||
))
|
||||
}
|
||||
|
||||
// verify that the pallet is still halted (or operational if it is the last batch)
|
||||
ensure_pallet_operating_mode(&finality_target, is_last_batch).await?;
|
||||
}
|
||||
|
||||
if let Some(latest_andatory_header_number) = latest_andatory_header_number {
|
||||
log::info!(
|
||||
target: "bridge",
|
||||
"Successfully updated best {} header at {} to {}. Pallet is now operational",
|
||||
Source::NAME,
|
||||
Target::NAME,
|
||||
latest_andatory_header_number,
|
||||
);
|
||||
}
|
||||
|
||||
Ok(())
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
/// Mandatory header and its finality proof.
|
||||
type HeaderAndProof<P> = (
|
||||
SyncHeader<HeaderOf<<P as SubstrateFinalitySyncPipeline>::SourceChain>>,
|
||||
GrandpaJustification<HeaderOf<<P as SubstrateFinalitySyncPipeline>::SourceChain>>,
|
||||
);
|
||||
/// Vector of mandatory headers and their finality proofs.
|
||||
type HeadersAndProofs<P> = Vec<HeaderAndProof<P>>;
|
||||
|
||||
/// Returns best finalized source header number known to the bridge GRANDPA pallet at the target
|
||||
/// chain.
|
||||
///
|
||||
/// This function works even if bridge GRANDPA pallet at the target chain is halted.
|
||||
async fn best_source_block_number_at_target<P: SubstrateFinalitySyncPipeline>(
|
||||
target_client: &Client<P::TargetChain>,
|
||||
) -> anyhow::Result<BlockNumberOf<P::SourceChain>> {
|
||||
Ok(read_client_state::<P::TargetChain, P::SourceChain>(
|
||||
&target_client,
|
||||
None,
|
||||
P::SourceChain::BEST_FINALIZED_HEADER_ID_METHOD,
|
||||
)
|
||||
.await?
|
||||
.best_finalized_peer_at_best_self
|
||||
.0)
|
||||
}
|
||||
|
||||
/// Verify that the bridge GRANDPA pallet at the target chain is either halted, or operational.
|
||||
async fn ensure_pallet_operating_mode<P: SubstrateFinalitySyncPipeline>(
|
||||
finality_target: &SubstrateFinalityTarget<P>,
|
||||
operational: bool,
|
||||
) -> anyhow::Result<()> {
|
||||
match (operational, finality_target.ensure_pallet_active().await) {
|
||||
(true, Ok(())) => Ok(()),
|
||||
(false, Err(SubstrateError::BridgePalletIsHalted)) => Ok(()),
|
||||
_ =>
|
||||
return Err(anyhow::format_err!(
|
||||
"Bridge GRANDPA pallet at {} is expected to be {}, but it isn't",
|
||||
P::TargetChain::NAME,
|
||||
if operational { "operational" } else { "halted" },
|
||||
)),
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns list of all mandatory headers in given range.
|
||||
async fn find_mandatory_headers_in_range<P: SubstrateFinalitySyncPipeline>(
|
||||
finality_source: &SubstrateFinalitySource<P>,
|
||||
range: (BlockNumberOf<P::SourceChain>, BlockNumberOf<P::SourceChain>),
|
||||
) -> anyhow::Result<HeadersAndProofs<P>> {
|
||||
let mut mandatory_headers = Vec::new();
|
||||
let mut current = range.0;
|
||||
while current <= range.1 {
|
||||
let (header, proof) = finality_source.header_and_finality_proof(current).await?.into();
|
||||
if header.is_mandatory() {
|
||||
match proof {
|
||||
Some(proof) => mandatory_headers.push((header, proof)),
|
||||
None =>
|
||||
return Err(anyhow::format_err!(
|
||||
"Missing GRANDPA justification for {} header {}",
|
||||
P::SourceChain::NAME,
|
||||
current,
|
||||
)),
|
||||
}
|
||||
}
|
||||
|
||||
current += One::one();
|
||||
}
|
||||
|
||||
Ok(mandatory_headers)
|
||||
}
|
||||
|
||||
/// Given list of mandatory headers, prepare batches of headers, so that every batch may fit into
|
||||
/// single transaction.
|
||||
fn make_mandatory_headers_batches<
|
||||
P: SubstrateFinalitySyncPipeline,
|
||||
F: Fn(&HeaderAndProof<P>) -> Weight,
|
||||
>(
|
||||
mut headers_to_submit: HeadersAndProofs<P>,
|
||||
submit_header_weight: F,
|
||||
) -> Vec<HeadersAndProofs<P>> {
|
||||
// now that we have all mandatory headers, let's prepare transactions
|
||||
// (let's keep all our transactions below 2/3 of max tx size/weight to have some reserve
|
||||
// for utility overhead + for halting transaction)
|
||||
let maximal_tx_size = P::TargetChain::max_extrinsic_size() * 2 / 3;
|
||||
let maximal_tx_weight = P::TargetChain::max_extrinsic_weight() * 2 / 3;
|
||||
let mut current_batch_size: u32 = 0;
|
||||
let mut current_batch_weight: Weight = 0;
|
||||
let mut batches = Vec::new();
|
||||
let mut i = 0;
|
||||
while i < headers_to_submit.len() {
|
||||
let header_and_proof_size =
|
||||
headers_to_submit[i].0.encode().len() + headers_to_submit[i].1.encode().len();
|
||||
let header_and_proof_weight = submit_header_weight(&headers_to_submit[i]);
|
||||
|
||||
let new_batch_size = current_batch_size
|
||||
.saturating_add(u32::try_from(header_and_proof_size).unwrap_or(u32::MAX));
|
||||
let new_batch_weight = current_batch_weight.saturating_add(header_and_proof_weight);
|
||||
|
||||
let is_exceeding_tx_size = new_batch_size > maximal_tx_size;
|
||||
let is_exceeding_tx_weight = new_batch_weight > maximal_tx_weight;
|
||||
let is_new_batch_required = is_exceeding_tx_size || is_exceeding_tx_weight;
|
||||
|
||||
if is_new_batch_required {
|
||||
// if `i` is 0 and we're here, it is a weird situation: even single header submission is
|
||||
// larger than we've planned for a bunch of headers. Let's be optimistic and hope that
|
||||
// the tx will still succeed.
|
||||
let spit_off_index = std::cmp::max(i, 1);
|
||||
let remaining_headers_to_submit = headers_to_submit.split_off(spit_off_index);
|
||||
batches.push(headers_to_submit);
|
||||
|
||||
// we'll reiterate the same header again => so set `current_*` to zero
|
||||
current_batch_size = 0;
|
||||
current_batch_weight = 0;
|
||||
headers_to_submit = remaining_headers_to_submit;
|
||||
i = 0;
|
||||
} else {
|
||||
current_batch_size = new_batch_size;
|
||||
current_batch_weight = new_batch_weight;
|
||||
i = i + 1;
|
||||
}
|
||||
}
|
||||
if !headers_to_submit.is_empty() {
|
||||
batches.push(headers_to_submit);
|
||||
}
|
||||
batches
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
use crate::cli::{RuntimeVersionType, SourceRuntimeVersionParams, TargetRuntimeVersionParams};
|
||||
use bp_test_utils::{make_default_justification, test_header};
|
||||
use relay_polkadot_client::Polkadot;
|
||||
use sp_runtime::{traits::Header as _, DigestItem};
|
||||
|
||||
fn make_header_and_justification(
|
||||
i: u32,
|
||||
size: u32,
|
||||
) -> (SyncHeader<bp_kusama::Header>, GrandpaJustification<bp_kusama::Header>) {
|
||||
let size = size as usize;
|
||||
let mut header: bp_kusama::Header = test_header(i);
|
||||
let justification = make_default_justification(&header);
|
||||
let actual_size = header.encode().len() + justification.encode().len();
|
||||
// additional digest means some additional bytes, so let's decrease `additional_digest_size`
|
||||
// a bit
|
||||
let additional_digest_size = size.saturating_sub(actual_size).saturating_sub(100);
|
||||
header.digest_mut().push(DigestItem::Other(vec![0u8; additional_digest_size]));
|
||||
let justification = make_default_justification(&header);
|
||||
println!("{} {}", size, header.encode().len() + justification.encode().len());
|
||||
(header.into(), justification)
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn should_parse_cli_options() {
|
||||
// when
|
||||
let res = ReinitBridge::from_iter(vec![
|
||||
"reinit-bridge",
|
||||
"kusama-to-polkadot",
|
||||
"--source-host",
|
||||
"127.0.0.1",
|
||||
"--source-port",
|
||||
"42",
|
||||
"--target-host",
|
||||
"127.0.0.1",
|
||||
"--target-port",
|
||||
"43",
|
||||
"--target-signer",
|
||||
"//Alice",
|
||||
]);
|
||||
|
||||
// then
|
||||
assert_eq!(
|
||||
res,
|
||||
ReinitBridge {
|
||||
bridge: ReinitBridgeName::KusamaToPolkadot,
|
||||
source: SourceConnectionParams {
|
||||
source_host: "127.0.0.1".into(),
|
||||
source_port: 42,
|
||||
source_secure: false,
|
||||
source_runtime_version: SourceRuntimeVersionParams {
|
||||
source_version_mode: RuntimeVersionType::Bundle,
|
||||
source_spec_version: None,
|
||||
source_transaction_version: None,
|
||||
}
|
||||
},
|
||||
target: TargetConnectionParams {
|
||||
target_host: "127.0.0.1".into(),
|
||||
target_port: 43,
|
||||
target_secure: false,
|
||||
target_runtime_version: TargetRuntimeVersionParams {
|
||||
target_version_mode: RuntimeVersionType::Bundle,
|
||||
target_spec_version: None,
|
||||
target_transaction_version: None,
|
||||
}
|
||||
},
|
||||
target_sign: TargetSigningParams {
|
||||
target_signer: Some("//Alice".into()),
|
||||
target_signer_password: None,
|
||||
target_signer_file: None,
|
||||
target_signer_password_file: None,
|
||||
target_transactions_mortality: None,
|
||||
},
|
||||
}
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn make_mandatory_headers_batches_and_empty_headers() {
|
||||
let batches = make_mandatory_headers_batches::<KusamaFinalityToPolkadot, _>(vec![], |_| 0);
|
||||
assert!(batches.is_empty());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn make_mandatory_headers_batches_with_single_batch() {
|
||||
let headers_to_submit =
|
||||
vec![make_header_and_justification(10, Polkadot::max_extrinsic_size() / 3)];
|
||||
let batches =
|
||||
make_mandatory_headers_batches::<KusamaFinalityToPolkadot, _>(headers_to_submit, |_| 0);
|
||||
assert_eq!(batches.into_iter().map(|x| x.len()).collect::<Vec<_>>(), vec![1],);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn make_mandatory_headers_batches_group_by_size() {
|
||||
let headers_to_submit = vec![
|
||||
make_header_and_justification(10, Polkadot::max_extrinsic_size() / 3),
|
||||
make_header_and_justification(20, Polkadot::max_extrinsic_size() / 3),
|
||||
make_header_and_justification(30, Polkadot::max_extrinsic_size() * 2 / 3),
|
||||
make_header_and_justification(40, Polkadot::max_extrinsic_size()),
|
||||
];
|
||||
let batches =
|
||||
make_mandatory_headers_batches::<KusamaFinalityToPolkadot, _>(headers_to_submit, |_| 0);
|
||||
assert_eq!(batches.into_iter().map(|x| x.len()).collect::<Vec<_>>(), vec![2, 1, 1],);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn make_mandatory_headers_batches_group_by_weight() {
|
||||
let headers_to_submit = vec![
|
||||
make_header_and_justification(10, 0),
|
||||
make_header_and_justification(20, 0),
|
||||
make_header_and_justification(30, 0),
|
||||
make_header_and_justification(40, 0),
|
||||
];
|
||||
let batches = make_mandatory_headers_batches::<KusamaFinalityToPolkadot, _>(
|
||||
headers_to_submit,
|
||||
|(header, _)| {
|
||||
if header.number() == 10 || header.number() == 20 {
|
||||
Polkadot::max_extrinsic_weight() / 3
|
||||
} else if header.number() == 30 {
|
||||
Polkadot::max_extrinsic_weight() * 2 / 3
|
||||
} else {
|
||||
Polkadot::max_extrinsic_weight()
|
||||
}
|
||||
},
|
||||
);
|
||||
assert_eq!(batches.into_iter().map(|x| x.len()).collect::<Vec<_>>(), vec![2, 1, 1],);
|
||||
}
|
||||
}
|
||||
@@ -70,6 +70,9 @@ pub enum Call {
|
||||
/// Balances pallet.
|
||||
#[codec(index = 4)]
|
||||
Balances(BalancesCall),
|
||||
/// Utility pallet.
|
||||
#[codec(index = 24)]
|
||||
Utility(UtilityCall),
|
||||
/// Polkadot bridge pallet.
|
||||
#[codec(index = 110)]
|
||||
BridgePolkadotGrandpa(BridgePolkadotGrandpaCall),
|
||||
@@ -102,6 +105,8 @@ pub enum BridgePolkadotGrandpaCall {
|
||||
),
|
||||
#[codec(index = 1)]
|
||||
initialize(bp_header_chain::InitializationData<<PolkadotLike as Chain>::Header>),
|
||||
#[codec(index = 3)]
|
||||
set_operational(bool),
|
||||
}
|
||||
|
||||
#[derive(Encode, Decode, Debug, PartialEq, Eq, Clone, TypeInfo)]
|
||||
@@ -136,6 +141,13 @@ pub enum BridgePolkadotMessagesCall {
|
||||
),
|
||||
}
|
||||
|
||||
#[derive(Encode, Decode, Debug, PartialEq, Eq, Clone, TypeInfo)]
|
||||
#[allow(non_camel_case_types)]
|
||||
pub enum UtilityCall {
|
||||
#[codec(index = 2)]
|
||||
batch_all(Vec<Call>),
|
||||
}
|
||||
|
||||
#[derive(Encode, Decode, Debug, PartialEq, Eq, Clone, TypeInfo)]
|
||||
pub enum BridgePolkadotMessagesParameter {
|
||||
#[codec(index = 0)]
|
||||
|
||||
@@ -70,6 +70,9 @@ pub enum Call {
|
||||
/// Balances pallet.
|
||||
#[codec(index = 5)]
|
||||
Balances(BalancesCall),
|
||||
/// Utility pallet.
|
||||
#[codec(index = 26)]
|
||||
Utility(UtilityCall),
|
||||
/// Kusama bridge pallet.
|
||||
#[codec(index = 110)]
|
||||
BridgeKusamaGrandpa(BridgeKusamaGrandpaCall),
|
||||
@@ -102,6 +105,8 @@ pub enum BridgeKusamaGrandpaCall {
|
||||
),
|
||||
#[codec(index = 1)]
|
||||
initialize(bp_header_chain::InitializationData<<PolkadotLike as Chain>::Header>),
|
||||
#[codec(index = 3)]
|
||||
set_operational(bool),
|
||||
}
|
||||
|
||||
#[derive(Encode, Decode, Debug, PartialEq, Eq, Clone, TypeInfo)]
|
||||
@@ -136,6 +141,13 @@ pub enum BridgeKusamaMessagesCall {
|
||||
),
|
||||
}
|
||||
|
||||
#[derive(Encode, Decode, Debug, PartialEq, Eq, Clone, TypeInfo)]
|
||||
#[allow(non_camel_case_types)]
|
||||
pub enum UtilityCall {
|
||||
#[codec(index = 2)]
|
||||
batch_all(Vec<Call>),
|
||||
}
|
||||
|
||||
#[derive(Encode, Decode, Debug, PartialEq, Eq, Clone, TypeInfo)]
|
||||
pub enum BridgeKusamaMessagesParameter {
|
||||
#[codec(index = 0)]
|
||||
|
||||
@@ -65,7 +65,7 @@ pub trait SubstrateFinalitySyncPipeline: 'static + Clone + Debug + Send + Sync {
|
||||
|
||||
/// Adapter that allows all `SubstrateFinalitySyncPipeline` to act as `FinalitySyncPipeline`.
|
||||
#[derive(Clone, Debug)]
|
||||
pub(crate) struct FinalitySyncPipelineAdapter<P: SubstrateFinalitySyncPipeline> {
|
||||
pub struct FinalitySyncPipelineAdapter<P: SubstrateFinalitySyncPipeline> {
|
||||
_phantom: PhantomData<P>,
|
||||
}
|
||||
|
||||
|
||||
@@ -52,7 +52,7 @@ impl<P: SubstrateFinalitySyncPipeline> SubstrateFinalityTarget<P> {
|
||||
}
|
||||
|
||||
/// Ensure that the GRANDPA pallet at target chain is active.
|
||||
async fn ensure_pallet_active(&self) -> Result<(), Error> {
|
||||
pub async fn ensure_pallet_active(&self) -> Result<(), Error> {
|
||||
let is_halted = self
|
||||
.client
|
||||
.storage_value(is_halted_key(P::SourceChain::WITH_CHAIN_GRANDPA_PALLET_NAME), None)
|
||||
|
||||
Reference in New Issue
Block a user