mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-28 23:47:56 +00:00
e1c033ebe1
(imported from https://github.com/paritytech/cumulus/pull/2157) ## Changes This MR refactores the XCMP, Parachains System and DMP pallets to use the [MessageQueue](https://github.com/paritytech/substrate/pull/12485) for delayed execution of incoming messages. The DMP pallet is entirely replaced by the MQ and thereby removed. This allows for PoV-bounded execution and resolves a number of issues that stem from the current work-around. All System Parachains adopt this change. The most important changes are in `primitives/core/src/lib.rs`, `parachains/common/src/process_xcm_message.rs`, `pallets/parachain-system/src/lib.rs`, `pallets/xcmp-queue/src/lib.rs` and the runtime configs. ### DMP Queue Pallet The pallet got removed and its logic refactored into parachain-system. Overweight message management can be done directly through the MQ pallet. Final undeployment migrations are provided by `cumulus_pallet_dmp_queue::UndeployDmpQueue` and `DeleteDmpQueue` that can be configured with an aux config trait like: ```rust parameter_types! { pub const DmpQueuePalletName: &'static str = \"DmpQueue\" < CHANGE ME; pub const RelayOrigin: AggregateMessageOrigin = AggregateMessageOrigin::Parent; } impl cumulus_pallet_dmp_queue::MigrationConfig for Runtime { type PalletName = DmpQueuePalletName; type DmpHandler = frame_support::traits::EnqueueWithOrigin<MessageQueue, RelayOrigin>; type DbWeight = <Runtime as frame_system::Config>::DbWeight; } // And adding them to your Migrations tuple: pub type Migrations = ( ... cumulus_pallet_dmp_queue::UndeployDmpQueue<Runtime>, cumulus_pallet_dmp_queue::DeleteDmpQueue<Runtime>, ); ``` ### XCMP Queue pallet Removed all dispatch queue functionality. Incoming XCMP messages are now either: Immediately handled if they are Signals, enqueued into the MQ pallet otherwise. New config items for the XCMP queue pallet: ```rust /// The actual queue implementation that retains the messages for later processing. type XcmpQueue: EnqueueMessage<ParaId>; /// How a XCM over HRMP from a sibling parachain should be processed. type XcmpProcessor: ProcessMessage<Origin = ParaId>; /// The maximal number of suspended XCMP channels at the same time. #[pallet::constant] type MaxInboundSuspended: Get<u32>; ``` How to configure those: ```rust // Use the MessageQueue pallet to store messages for later processing. The `TransformOrigin` is needed since // the MQ pallet itself operators on `AggregateMessageOrigin` but we want to enqueue `ParaId`s. type XcmpQueue = TransformOrigin<MessageQueue, AggregateMessageOrigin, ParaId, ParaIdToSibling>; // Process XCMP messages from siblings. This is type-safe to only accept `ParaId`s. They will be dispatched // with origin `Junction::Sibling(…)`. type XcmpProcessor = ProcessFromSibling< ProcessXcmMessage< AggregateMessageOrigin, xcm_executor::XcmExecutor<xcm_config::XcmConfig>, RuntimeCall, >, >; // Not really important what to choose here. Just something larger than the maximal number of channels. type MaxInboundSuspended = sp_core::ConstU32<1_000>; ``` The `InboundXcmpStatus` storage item was replaced by `InboundXcmpSuspended` since it now only tracks inbound queue suspension and no message indices anymore. Now only sends the most recent channel `Signals`, as all prio ones are out-dated anyway. ### Parachain System pallet For `DMP` messages instead of forwarding them to the `DMP` pallet, it now pushes them to the configured `DmpQueue`. The message processing which was triggered in `set_validation_data` is now being done by the MQ pallet `on_initialize`. XCMP messages are still handed off to the `XcmpMessageHandler` (XCMP-Queue pallet) - no change here. New config items for the parachain system pallet: ```rust /// Queues inbound downward messages for delayed processing. /// /// Analogous to the `XcmpQueue` of the XCMP queue pallet. type DmpQueue: EnqueueMessage<AggregateMessageOrigin>; ``` How to configure: ```rust /// Use the MQ pallet to store DMP messages for delayed processing. type DmpQueue = MessageQueue; ``` ## Message Flow The flow of messages on the parachain side. Messages come in from the left via the `Validation Data` and finally end up at the `Xcm Executor` on the right.  ## Further changes - Bumped the default suspension, drop and resume thresholds in `QueueConfigData::default()`. - `XcmpQueue::{suspend_xcm_execution, resume_xcm_execution}` errors when they would be a noop. - Properly validate the `QueueConfigData` before setting it. - Marked weight files as auto-generated so they wont auto-expand in the MR files view. - Move the `hypothetical` asserts to `frame_support` under the name `experimental_hypothetically` Questions: - [ ] What about the ugly `#[cfg(feature = \"runtime-benchmarks\")]` in the runtimes? Not sure how to best fix. Just having them like this makes tests fail that rely on the real message processor when the feature is enabled. - [ ] Need a good weight for `MessageQueueServiceWeight`. The scheduler already takes 80% so I put it to 10% but that is quite low. TODO: - [x] Remove c&p code after https://github.com/paritytech/polkadot/pull/6271 - [x] Use `HandleMessage` once it is public in Substrate - [x] fix `runtime-benchmarks` feature https://github.com/paritytech/polkadot/pull/6966 - [x] Benchmarks - [x] Tests - [ ] Migrate `InboundXcmpStatus` to `InboundXcmpSuspended` - [x] Possibly cleanup Migrations (DMP+XCMP) - [x] optional: create `TransformProcessMessageOrigin` in Substrate and replace `ProcessFromSibling` - [ ] Rerun weights on ref HW --------- Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> Co-authored-by: Liam Aharon <liam.aharon@hotmail.com> Co-authored-by: joe petrowski <25483142+joepetrowski@users.noreply.github.com> Co-authored-by: Kian Paimani <5588131+kianenigma@users.noreply.github.com> Co-authored-by: command-bot <>
1203 lines
32 KiB
Rust
Executable File
1203 lines
32 KiB
Rust
Executable File
// Copyright (C) Parity Technologies (UK) Ltd.
|
|
// This file is part of Cumulus.
|
|
|
|
// Cumulus is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// Cumulus is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with Cumulus. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
#![cfg(test)]
|
|
|
|
use super::*;
|
|
use crate::mock::*;
|
|
|
|
use cumulus_primitives_core::{AbridgedHrmpChannel, InboundDownwardMessage, InboundHrmpMessage};
|
|
use frame_support::{assert_ok, parameter_types, weights::Weight};
|
|
use frame_system::RawOrigin;
|
|
use hex_literal::hex;
|
|
use rand::Rng;
|
|
use relay_chain::HrmpChannelId;
|
|
use sp_core::H256;
|
|
use sp_std::num::NonZeroU32;
|
|
|
|
#[test]
|
|
#[should_panic]
|
|
fn block_tests_run_on_drop() {
|
|
BlockTests::new().add(123, || panic!("if this test passes, block tests run properly"));
|
|
}
|
|
|
|
#[test]
|
|
fn test_xcmp_source_keeps_messages() {
|
|
let recipient = ParaId::from(400);
|
|
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(3).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(2)
|
|
.with_relay_sproof_builder(move |_, block_number, sproof| {
|
|
sproof.host_config.hrmp_max_message_num_per_candidate = 10;
|
|
let channel = sproof.upsert_outbound_channel(recipient);
|
|
channel.max_total_size = 10;
|
|
channel.max_message_size = 10;
|
|
|
|
// Only fit messages starting from 3rd block.
|
|
channel.max_capacity = if block_number < 3 { 0 } else { 1 };
|
|
})
|
|
.add(1, || {})
|
|
.add_with_post_test(
|
|
2,
|
|
move || {
|
|
send_message(recipient, b"22".to_vec());
|
|
},
|
|
move || {
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert!(v.is_empty());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
3,
|
|
move || {},
|
|
move || {
|
|
// Not discarded.
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert_eq!(v, vec![OutboundHrmpMessage { recipient, data: b"22".to_vec() }]);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn unincluded_segment_works() {
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(10).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(1)
|
|
.add_with_post_test(
|
|
123,
|
|
|| {},
|
|
|| {
|
|
let segment = <UnincludedSegment<Test>>::get();
|
|
assert_eq!(segment.len(), 1);
|
|
assert!(<AggregatedUnincludedSegment<Test>>::get().is_some());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
124,
|
|
|| {},
|
|
|| {
|
|
let segment = <UnincludedSegment<Test>>::get();
|
|
assert_eq!(segment.len(), 2);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
125,
|
|
|| {},
|
|
|| {
|
|
let segment = <UnincludedSegment<Test>>::get();
|
|
// Block 123 was popped from the segment, the len is still 2.
|
|
assert_eq!(segment.len(), 2);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
#[should_panic = "no space left for the block in the unincluded segment"]
|
|
fn unincluded_segment_is_limited() {
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(1).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(2)
|
|
.add_with_post_test(
|
|
123,
|
|
|| {},
|
|
|| {
|
|
let segment = <UnincludedSegment<Test>>::get();
|
|
assert_eq!(segment.len(), 1);
|
|
assert!(<AggregatedUnincludedSegment<Test>>::get().is_some());
|
|
},
|
|
)
|
|
.add(124, || {}); // The previous block wasn't included yet, should panic in `create_inherent`.
|
|
}
|
|
|
|
#[test]
|
|
fn unincluded_code_upgrade_handles_signal() {
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(2).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(1)
|
|
.with_relay_sproof_builder(|_, block_number, builder| {
|
|
if block_number > 123 && block_number <= 125 {
|
|
builder.upgrade_go_ahead = Some(relay_chain::UpgradeGoAhead::GoAhead);
|
|
}
|
|
})
|
|
.add(123, || {
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
})
|
|
.add_with_post_test(
|
|
124,
|
|
|| {},
|
|
|| {
|
|
assert!(
|
|
!<PendingValidationCode<Test>>::exists(),
|
|
"validation function must have been unset"
|
|
);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
125,
|
|
|| {
|
|
// The signal is present in relay state proof and ignored.
|
|
// Block that processed the signal is still not included.
|
|
},
|
|
|| {
|
|
let segment = <UnincludedSegment<Test>>::get();
|
|
assert_eq!(segment.len(), 2);
|
|
let aggregated_segment =
|
|
<AggregatedUnincludedSegment<Test>>::get().expect("segment is non-empty");
|
|
assert_eq!(
|
|
aggregated_segment.consumed_go_ahead_signal(),
|
|
Some(relay_chain::UpgradeGoAhead::GoAhead)
|
|
);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
126,
|
|
|| {},
|
|
|| {
|
|
let aggregated_segment =
|
|
<AggregatedUnincludedSegment<Test>>::get().expect("segment is non-empty");
|
|
// Block that processed the signal is included.
|
|
assert!(aggregated_segment.consumed_go_ahead_signal().is_none());
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn unincluded_code_upgrade_scheduled_after_go_ahead() {
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(2).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(1)
|
|
.with_relay_sproof_builder(|_, block_number, builder| {
|
|
if block_number > 123 && block_number <= 125 {
|
|
builder.upgrade_go_ahead = Some(relay_chain::UpgradeGoAhead::GoAhead);
|
|
}
|
|
})
|
|
.add(123, || {
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
})
|
|
.add_with_post_test(
|
|
124,
|
|
|| {},
|
|
|| {
|
|
assert!(
|
|
!<PendingValidationCode<Test>>::exists(),
|
|
"validation function must have been unset"
|
|
);
|
|
// The previous go-ahead signal was processed, schedule another upgrade.
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
125,
|
|
|| {
|
|
// The signal is present in relay state proof and ignored.
|
|
// Block that processed the signal is still not included.
|
|
},
|
|
|| {
|
|
let segment = <UnincludedSegment<Test>>::get();
|
|
assert_eq!(segment.len(), 2);
|
|
let aggregated_segment =
|
|
<AggregatedUnincludedSegment<Test>>::get().expect("segment is non-empty");
|
|
assert_eq!(
|
|
aggregated_segment.consumed_go_ahead_signal(),
|
|
Some(relay_chain::UpgradeGoAhead::GoAhead)
|
|
);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
126,
|
|
|| {},
|
|
|| {
|
|
assert!(<PendingValidationCode<Test>>::exists(), "upgrade is pending");
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn inherent_processed_messages_are_ignored() {
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(2).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(1)
|
|
.with_relay_block_number(|block_number| 3.max(*block_number as RelayChainBlockNumber))
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| match relay_block_num {
|
|
3 => {
|
|
sproof.dmq_mqc_head =
|
|
Some(MessageQueueChain::default().extend_downward(&mk_dmp(3)).head());
|
|
sproof.upsert_inbound_channel(ParaId::from(200)).mqc_head = Some(
|
|
MessageQueueChain::default()
|
|
.extend_hrmp(&mk_hrmp(2))
|
|
.extend_hrmp(&mk_hrmp(3))
|
|
.head(),
|
|
);
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| match relay_block_num {
|
|
3 => {
|
|
data.downward_messages.push(mk_dmp(3));
|
|
data.horizontal_messages.insert(ParaId::from(200), vec![mk_hrmp(2), mk_hrmp(3)]);
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.add(1, || {
|
|
// Don't drop processed messages for this test.
|
|
HANDLED_DMP_MESSAGES.with(|m| {
|
|
let m = m.borrow();
|
|
// NOTE: if this fails, then run the test without benchmark features.
|
|
assert_eq!(&*m, &[mk_dmp(3).msg]);
|
|
});
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let m = m.borrow_mut();
|
|
assert_eq!(
|
|
&*m,
|
|
&[(ParaId::from(200), 2, b"2".to_vec()), (ParaId::from(200), 3, b"3".to_vec()),]
|
|
);
|
|
});
|
|
})
|
|
.add(2, || {})
|
|
.add(3, || {
|
|
HANDLED_DMP_MESSAGES.with(|m| {
|
|
let m = m.borrow();
|
|
assert_eq!(&*m, &[mk_dmp(3).msg]);
|
|
});
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let m = m.borrow_mut();
|
|
assert_eq!(
|
|
&*m,
|
|
&[(ParaId::from(200), 2, b"2".to_vec()), (ParaId::from(200), 3, b"3".to_vec()),]
|
|
);
|
|
});
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn hrmp_outbound_respects_used_bandwidth() {
|
|
let recipient = ParaId::from(400);
|
|
|
|
CONSENSUS_HOOK.with(|c| {
|
|
*c.borrow_mut() = Box::new(|_| (Weight::zero(), NonZeroU32::new(3).unwrap().into()))
|
|
});
|
|
|
|
BlockTests::new()
|
|
.with_inclusion_delay(2)
|
|
.with_relay_sproof_builder(move |_, block_number, sproof| {
|
|
sproof.host_config.hrmp_max_message_num_per_candidate = 10;
|
|
let channel = sproof.upsert_outbound_channel(recipient);
|
|
channel.max_capacity = 2;
|
|
channel.max_total_size = 4;
|
|
|
|
channel.max_message_size = 10;
|
|
|
|
// states:
|
|
// [relay_chain][unincluded_segment] + [message_queue]
|
|
// 2: []["2"] + ["2222"]
|
|
// 3: []["2", "3"] + ["2222"]
|
|
// 4: []["2", "3"] + ["2222", "444", "4"]
|
|
// 5: ["2"]["3"] + ["2222", "444", "4"]
|
|
// 6: ["2", "3"][] + ["2222", "444", "4"]
|
|
// 7: ["3"]["444"] + ["2222", "4"]
|
|
// 8: []["444", "4"] + ["2222"]
|
|
//
|
|
// 2 tests max bytes - there is message space but no byte space.
|
|
// 4 tests max capacity - there is byte space but no message space
|
|
|
|
match block_number {
|
|
5 => {
|
|
// 2 included.
|
|
// one message added
|
|
channel.msg_count = 1;
|
|
channel.total_size = 1;
|
|
},
|
|
6 => {
|
|
// 3 included.
|
|
// one message added
|
|
channel.msg_count = 2;
|
|
channel.total_size = 2;
|
|
},
|
|
7 => {
|
|
// 4 included.
|
|
// one message drained.
|
|
channel.msg_count = 1;
|
|
channel.total_size = 1;
|
|
},
|
|
8 => {
|
|
// 5 included. no messages added, one drained.
|
|
channel.msg_count = 0;
|
|
channel.total_size = 0;
|
|
},
|
|
_ => {
|
|
channel.msg_count = 0;
|
|
channel.total_size = 0;
|
|
},
|
|
}
|
|
})
|
|
.add(1, || {})
|
|
.add_with_post_test(
|
|
2,
|
|
move || {
|
|
send_message(recipient, b"2".to_vec());
|
|
send_message(recipient, b"2222".to_vec());
|
|
},
|
|
move || {
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert_eq!(v, vec![OutboundHrmpMessage { recipient, data: b"2".to_vec() }]);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
3,
|
|
move || {
|
|
send_message(recipient, b"3".to_vec());
|
|
},
|
|
move || {
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert_eq!(v, vec![OutboundHrmpMessage { recipient, data: b"3".to_vec() }]);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
4,
|
|
move || {
|
|
send_message(recipient, b"444".to_vec());
|
|
send_message(recipient, b"4".to_vec());
|
|
},
|
|
move || {
|
|
// Queue has byte capacity but not message capacity.
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert!(v.is_empty());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
5,
|
|
|| {},
|
|
move || {
|
|
// 1 is included here, channel not drained yet. nothing fits.
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert!(v.is_empty());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
6,
|
|
|| {},
|
|
move || {
|
|
// 2 is included here. channel is totally full.
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert!(v.is_empty());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
7,
|
|
|| {},
|
|
move || {
|
|
// 3 is included here. One message was drained out. The 3-byte message
|
|
// finally fits
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
// This line relies on test implementation of [`XcmpMessageSource`].
|
|
assert_eq!(v, vec![OutboundHrmpMessage { recipient, data: b"444".to_vec() }]);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
8,
|
|
|| {},
|
|
move || {
|
|
// 4 is included here. Relay-chain side of the queue is empty,
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
// This line relies on test implementation of [`XcmpMessageSource`].
|
|
assert_eq!(v, vec![OutboundHrmpMessage { recipient, data: b"4".to_vec() }]);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn runtime_upgrade_events() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, block_number, builder| {
|
|
if block_number > 123 {
|
|
builder.upgrade_go_ahead = Some(relay_chain::UpgradeGoAhead::GoAhead);
|
|
}
|
|
})
|
|
.add_with_post_test(
|
|
123,
|
|
|| {
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
},
|
|
|| {
|
|
let events = System::events();
|
|
assert_eq!(
|
|
events[0].event,
|
|
RuntimeEvent::ParachainSystem(crate::Event::ValidationFunctionStored)
|
|
);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
1234,
|
|
|| {},
|
|
|| {
|
|
let events = System::events();
|
|
|
|
assert_eq!(events[0].event, RuntimeEvent::System(frame_system::Event::CodeUpdated));
|
|
|
|
assert_eq!(
|
|
events[1].event,
|
|
RuntimeEvent::ParachainSystem(crate::Event::ValidationFunctionApplied {
|
|
relay_chain_block_num: 1234
|
|
})
|
|
);
|
|
|
|
assert!(System::digest()
|
|
.logs()
|
|
.iter()
|
|
.any(|d| *d == sp_runtime::generic::DigestItem::RuntimeEnvironmentUpdated));
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn non_overlapping() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, _, builder| {
|
|
builder.host_config.validation_upgrade_delay = 1000;
|
|
})
|
|
.add(123, || {
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
})
|
|
.add(234, || {
|
|
assert_eq!(
|
|
System::set_code(RawOrigin::Root.into(), Default::default()),
|
|
Err(Error::<Test>::OverlappingUpgrades.into()),
|
|
)
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn manipulates_storage() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, block_number, builder| {
|
|
if block_number > 123 {
|
|
builder.upgrade_go_ahead = Some(relay_chain::UpgradeGoAhead::GoAhead);
|
|
}
|
|
})
|
|
.add(123, || {
|
|
assert!(
|
|
!<PendingValidationCode<Test>>::exists(),
|
|
"validation function must not exist yet"
|
|
);
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
assert!(<PendingValidationCode<Test>>::exists(), "validation function must now exist");
|
|
})
|
|
.add_with_post_test(
|
|
1234,
|
|
|| {},
|
|
|| {
|
|
assert!(
|
|
!<PendingValidationCode<Test>>::exists(),
|
|
"validation function must have been unset"
|
|
);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn aborted_upgrade() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, block_number, builder| {
|
|
if block_number > 123 {
|
|
builder.upgrade_go_ahead = Some(relay_chain::UpgradeGoAhead::Abort);
|
|
}
|
|
})
|
|
.add(123, || {
|
|
assert_ok!(System::set_code(RawOrigin::Root.into(), Default::default()));
|
|
})
|
|
.add_with_post_test(
|
|
1234,
|
|
|| {},
|
|
|| {
|
|
assert!(
|
|
!<PendingValidationCode<Test>>::exists(),
|
|
"validation function must have been unset"
|
|
);
|
|
let events = System::events();
|
|
assert_eq!(
|
|
events[0].event,
|
|
RuntimeEvent::ParachainSystem(crate::Event::ValidationFunctionDiscarded)
|
|
);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn checks_size() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, _, builder| {
|
|
builder.host_config.max_code_size = 8;
|
|
})
|
|
.add(123, || {
|
|
assert_eq!(
|
|
System::set_code(RawOrigin::Root.into(), vec![0; 64]),
|
|
Err(Error::<Test>::TooBig.into()),
|
|
);
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn send_upward_message_num_per_candidate() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, _, sproof| {
|
|
sproof.host_config.max_upward_message_num_per_candidate = 1;
|
|
sproof.relay_dispatch_queue_remaining_capacity = None;
|
|
})
|
|
.add_with_post_test(
|
|
1,
|
|
|| {
|
|
ParachainSystem::send_upward_message(b"Mr F was here".to_vec()).unwrap();
|
|
ParachainSystem::send_upward_message(b"message 2".to_vec()).unwrap();
|
|
},
|
|
|| {
|
|
let v = UpwardMessages::<Test>::get();
|
|
assert_eq!(v, vec![b"Mr F was here".to_vec()]);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
2,
|
|
|| {
|
|
assert_eq!(UnincludedSegment::<Test>::get().len(), 0);
|
|
/* do nothing within block */
|
|
},
|
|
|| {
|
|
let v = UpwardMessages::<Test>::get();
|
|
assert_eq!(v, vec![b"message 2".to_vec()]);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn send_upward_message_relay_bottleneck() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| {
|
|
sproof.host_config.max_upward_message_num_per_candidate = 2;
|
|
sproof.host_config.max_upward_queue_count = 5;
|
|
|
|
match relay_block_num {
|
|
1 => sproof.relay_dispatch_queue_remaining_capacity = Some((0, 2048)),
|
|
2 => sproof.relay_dispatch_queue_remaining_capacity = Some((1, 2048)),
|
|
_ => unreachable!(),
|
|
}
|
|
})
|
|
.add_with_post_test(
|
|
1,
|
|
|| {
|
|
ParachainSystem::send_upward_message(vec![0u8; 8]).unwrap();
|
|
},
|
|
|| {
|
|
// The message won't be sent because there is already one message in queue.
|
|
let v = UpwardMessages::<Test>::get();
|
|
assert!(v.is_empty());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
2,
|
|
|| { /* do nothing within block */ },
|
|
|| {
|
|
let v = UpwardMessages::<Test>::get();
|
|
assert_eq!(v, vec![vec![0u8; 8]]);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn send_hrmp_message_buffer_channel_close() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| {
|
|
//
|
|
// Base case setup
|
|
//
|
|
sproof.para_id = ParaId::from(200);
|
|
sproof.hrmp_egress_channel_index = Some(vec![ParaId::from(300), ParaId::from(400)]);
|
|
sproof.hrmp_channels.insert(
|
|
HrmpChannelId { sender: ParaId::from(200), recipient: ParaId::from(300) },
|
|
AbridgedHrmpChannel {
|
|
max_capacity: 1,
|
|
msg_count: 1, // <- 1/1 means the channel is full
|
|
max_total_size: 1024,
|
|
max_message_size: 8,
|
|
total_size: 0,
|
|
mqc_head: Default::default(),
|
|
},
|
|
);
|
|
sproof.hrmp_channels.insert(
|
|
HrmpChannelId { sender: ParaId::from(200), recipient: ParaId::from(400) },
|
|
AbridgedHrmpChannel {
|
|
max_capacity: 1,
|
|
msg_count: 1,
|
|
max_total_size: 1024,
|
|
max_message_size: 8,
|
|
total_size: 0,
|
|
mqc_head: Default::default(),
|
|
},
|
|
);
|
|
|
|
//
|
|
// Adjustment according to block
|
|
//
|
|
match relay_block_num {
|
|
1 => {},
|
|
2 => {},
|
|
3 => {
|
|
// The channel 200->400 ceases to exist at the relay chain block 3
|
|
sproof
|
|
.hrmp_egress_channel_index
|
|
.as_mut()
|
|
.unwrap()
|
|
.retain(|n| n != &ParaId::from(400));
|
|
sproof.hrmp_channels.remove(&HrmpChannelId {
|
|
sender: ParaId::from(200),
|
|
recipient: ParaId::from(400),
|
|
});
|
|
|
|
// We also free up space for a message in the 200->300 channel.
|
|
sproof
|
|
.hrmp_channels
|
|
.get_mut(&HrmpChannelId {
|
|
sender: ParaId::from(200),
|
|
recipient: ParaId::from(300),
|
|
})
|
|
.unwrap()
|
|
.msg_count = 0;
|
|
},
|
|
_ => unreachable!(),
|
|
}
|
|
})
|
|
.add_with_post_test(
|
|
1,
|
|
|| {
|
|
send_message(ParaId::from(300), b"1".to_vec());
|
|
send_message(ParaId::from(400), b"2".to_vec());
|
|
},
|
|
|| {},
|
|
)
|
|
.add_with_post_test(
|
|
2,
|
|
|| {},
|
|
|| {
|
|
// both channels are at capacity so we do not expect any messages.
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert!(v.is_empty());
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
3,
|
|
|| {},
|
|
|| {
|
|
let v = HrmpOutboundMessages::<Test>::get();
|
|
assert_eq!(
|
|
v,
|
|
vec![OutboundHrmpMessage { recipient: ParaId::from(300), data: b"1".to_vec() }]
|
|
);
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn message_queue_chain() {
|
|
assert_eq!(MessageQueueChain::default().head(), H256::zero());
|
|
|
|
// Note that the resulting hashes are the same for HRMP and DMP. That's because even though
|
|
// the types are nominally different, they have the same structure and computation of the
|
|
// new head doesn't differ.
|
|
//
|
|
// These cases are taken from https://github.com/paritytech/polkadot/pull/2351
|
|
assert_eq!(
|
|
MessageQueueChain::default()
|
|
.extend_downward(&InboundDownwardMessage { sent_at: 2, msg: vec![1, 2, 3] })
|
|
.extend_downward(&InboundDownwardMessage { sent_at: 3, msg: vec![4, 5, 6] })
|
|
.head(),
|
|
hex!["88dc00db8cc9d22aa62b87807705831f164387dfa49f80a8600ed1cbe1704b6b"].into(),
|
|
);
|
|
assert_eq!(
|
|
MessageQueueChain::default()
|
|
.extend_hrmp(&InboundHrmpMessage { sent_at: 2, data: vec![1, 2, 3] })
|
|
.extend_hrmp(&InboundHrmpMessage { sent_at: 3, data: vec![4, 5, 6] })
|
|
.head(),
|
|
hex!["88dc00db8cc9d22aa62b87807705831f164387dfa49f80a8600ed1cbe1704b6b"].into(),
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
#[cfg(not(feature = "runtime-benchmarks"))]
|
|
fn receive_dmp() {
|
|
lazy_static::lazy_static! {
|
|
static ref MSG: InboundDownwardMessage = InboundDownwardMessage {
|
|
sent_at: 1,
|
|
msg: b"down".to_vec(),
|
|
};
|
|
}
|
|
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| match relay_block_num {
|
|
1 => {
|
|
sproof.dmq_mqc_head =
|
|
Some(MessageQueueChain::default().extend_downward(&MSG).head());
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| match relay_block_num {
|
|
1 => {
|
|
data.downward_messages.push(MSG.clone());
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.add(1, || {
|
|
HANDLED_DMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(&*m, &[(MSG.msg.clone())]);
|
|
m.clear();
|
|
});
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
#[cfg(not(feature = "runtime-benchmarks"))]
|
|
fn receive_dmp_after_pause() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| match relay_block_num {
|
|
1 => {
|
|
sproof.dmq_mqc_head =
|
|
Some(MessageQueueChain::default().extend_downward(&mk_dmp(1)).head());
|
|
},
|
|
2 => {
|
|
// no new messages, mqc stayed the same.
|
|
sproof.dmq_mqc_head =
|
|
Some(MessageQueueChain::default().extend_downward(&mk_dmp(1)).head());
|
|
},
|
|
3 => {
|
|
sproof.dmq_mqc_head = Some(
|
|
MessageQueueChain::default()
|
|
.extend_downward(&mk_dmp(1))
|
|
.extend_downward(&mk_dmp(3))
|
|
.head(),
|
|
);
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| match relay_block_num {
|
|
1 => {
|
|
data.downward_messages.push(mk_dmp(1));
|
|
},
|
|
2 => {
|
|
// no new messages
|
|
},
|
|
3 => {
|
|
data.downward_messages.push(mk_dmp(3));
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.add(1, || {
|
|
HANDLED_DMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(&*m, &[(mk_dmp(1).msg.clone())]);
|
|
m.clear();
|
|
});
|
|
})
|
|
.add(2, || {})
|
|
.add(3, || {
|
|
HANDLED_DMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(&*m, &[(mk_dmp(3).msg.clone())]);
|
|
m.clear();
|
|
});
|
|
});
|
|
}
|
|
|
|
// Sent up to 100 DMP messages per block over a period of 100 blocks.
|
|
#[test]
|
|
#[cfg(not(feature = "runtime-benchmarks"))]
|
|
fn receive_dmp_many() {
|
|
wasm_ext().execute_with(|| {
|
|
parameter_types! {
|
|
pub storage MqcHead: MessageQueueChain = Default::default();
|
|
pub storage SentInBlock: Vec<Vec<InboundDownwardMessage>> = Default::default();
|
|
}
|
|
|
|
let mut sent_in_block = vec![vec![]];
|
|
let mut rng = rand::thread_rng();
|
|
|
|
for block in 1..100 {
|
|
let mut msgs = vec![];
|
|
for _ in 1..=rng.gen_range(1..=100) {
|
|
// Just use the same message multiple times per block.
|
|
msgs.push(mk_dmp(block));
|
|
}
|
|
sent_in_block.push(msgs);
|
|
}
|
|
SentInBlock::set(&sent_in_block);
|
|
|
|
let mut tester = BlockTests::new_without_externalities()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| {
|
|
let mut new_hash = MqcHead::get();
|
|
|
|
for msg in SentInBlock::get()[relay_block_num as usize].iter() {
|
|
new_hash.extend_downward(&msg);
|
|
}
|
|
|
|
sproof.dmq_mqc_head = Some(new_hash.head());
|
|
MqcHead::set(&new_hash);
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| {
|
|
for msg in SentInBlock::get()[relay_block_num as usize].iter() {
|
|
data.downward_messages.push(msg.clone());
|
|
}
|
|
});
|
|
|
|
for block in 1..100 {
|
|
tester = tester.add(block, move || {
|
|
HANDLED_DMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
let msgs = SentInBlock::get()[block as usize]
|
|
.iter()
|
|
.map(|m| m.msg.clone())
|
|
.collect::<Vec<_>>();
|
|
assert_eq!(&*m, &msgs);
|
|
m.clear();
|
|
});
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn receive_hrmp() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| match relay_block_num {
|
|
1 => {
|
|
// 200 - doesn't exist yet
|
|
// 300 - one new message
|
|
sproof.upsert_inbound_channel(ParaId::from(300)).mqc_head =
|
|
Some(MessageQueueChain::default().extend_hrmp(&mk_hrmp(1)).head());
|
|
},
|
|
2 => {
|
|
// 200 - now present with one message
|
|
// 300 - two new messages
|
|
sproof.upsert_inbound_channel(ParaId::from(200)).mqc_head =
|
|
Some(MessageQueueChain::default().extend_hrmp(&mk_hrmp(4)).head());
|
|
sproof.upsert_inbound_channel(ParaId::from(300)).mqc_head = Some(
|
|
MessageQueueChain::default()
|
|
.extend_hrmp(&mk_hrmp(1))
|
|
.extend_hrmp(&mk_hrmp(2))
|
|
.extend_hrmp(&mk_hrmp(3))
|
|
.head(),
|
|
);
|
|
},
|
|
3 => {
|
|
// 200 - no new messages
|
|
// 300 - is gone
|
|
sproof.upsert_inbound_channel(ParaId::from(200)).mqc_head =
|
|
Some(MessageQueueChain::default().extend_hrmp(&mk_hrmp(4)).head());
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| match relay_block_num {
|
|
1 => {
|
|
data.horizontal_messages.insert(ParaId::from(300), vec![mk_hrmp(1)]);
|
|
},
|
|
2 => {
|
|
data.horizontal_messages.insert(
|
|
ParaId::from(300),
|
|
vec![
|
|
// can't be sent at the block 1 actually. However, we cheat here
|
|
// because we want to test the case where there are multiple messages
|
|
// but the harness at the moment doesn't support block skipping.
|
|
mk_hrmp(2).clone(),
|
|
mk_hrmp(3).clone(),
|
|
],
|
|
);
|
|
data.horizontal_messages.insert(ParaId::from(200), vec![mk_hrmp(4)]);
|
|
},
|
|
3 => {},
|
|
_ => unreachable!(),
|
|
})
|
|
.add(1, || {
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(&*m, &[(ParaId::from(300), 1, b"1".to_vec())]);
|
|
m.clear();
|
|
});
|
|
})
|
|
.add(2, || {
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(
|
|
&*m,
|
|
&[
|
|
(ParaId::from(300), 2, b"2".to_vec()),
|
|
(ParaId::from(300), 3, b"3".to_vec()),
|
|
(ParaId::from(200), 4, b"4".to_vec()),
|
|
]
|
|
);
|
|
m.clear();
|
|
});
|
|
})
|
|
.add(3, || {});
|
|
}
|
|
|
|
#[test]
|
|
fn receive_hrmp_empty_channel() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| match relay_block_num {
|
|
1 => {
|
|
// no channels
|
|
},
|
|
2 => {
|
|
// one new channel
|
|
sproof.upsert_inbound_channel(ParaId::from(300)).mqc_head =
|
|
Some(MessageQueueChain::default().head());
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.add(1, || {})
|
|
.add(2, || {});
|
|
}
|
|
|
|
#[test]
|
|
fn receive_hrmp_after_pause() {
|
|
const ALICE: ParaId = ParaId::new(300);
|
|
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| match relay_block_num {
|
|
1 => {
|
|
sproof.upsert_inbound_channel(ALICE).mqc_head =
|
|
Some(MessageQueueChain::default().extend_hrmp(&mk_hrmp(1)).head());
|
|
},
|
|
2 => {
|
|
// 300 - no new messages, mqc stayed the same.
|
|
sproof.upsert_inbound_channel(ALICE).mqc_head =
|
|
Some(MessageQueueChain::default().extend_hrmp(&mk_hrmp(1)).head());
|
|
},
|
|
3 => {
|
|
// 300 - new message.
|
|
sproof.upsert_inbound_channel(ALICE).mqc_head = Some(
|
|
MessageQueueChain::default()
|
|
.extend_hrmp(&mk_hrmp(1))
|
|
.extend_hrmp(&mk_hrmp(3))
|
|
.head(),
|
|
);
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| match relay_block_num {
|
|
1 => {
|
|
data.horizontal_messages.insert(ALICE, vec![mk_hrmp(1)]);
|
|
},
|
|
2 => {
|
|
// no new messages
|
|
},
|
|
3 => {
|
|
data.horizontal_messages.insert(ALICE, vec![mk_hrmp(3)]);
|
|
},
|
|
_ => unreachable!(),
|
|
})
|
|
.add(1, || {
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(&*m, &[(ALICE, 1, b"1".to_vec())]);
|
|
m.clear();
|
|
});
|
|
})
|
|
.add(2, || {})
|
|
.add(3, || {
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
assert_eq!(&*m, &[(ALICE, 3, b"3".to_vec())]);
|
|
m.clear();
|
|
});
|
|
});
|
|
}
|
|
|
|
// Sent up to 100 HRMP messages per block over a period of 100 blocks.
|
|
#[test]
|
|
fn receive_hrmp_many() {
|
|
const ALICE: ParaId = ParaId::new(300);
|
|
|
|
wasm_ext().execute_with(|| {
|
|
parameter_types! {
|
|
pub storage MqcHead: MessageQueueChain = Default::default();
|
|
pub storage SentInBlock: Vec<Vec<InboundHrmpMessage>> = Default::default();
|
|
}
|
|
|
|
let mut sent_in_block = vec![vec![]];
|
|
let mut rng = rand::thread_rng();
|
|
|
|
for block in 1..100 {
|
|
let mut msgs = vec![];
|
|
for _ in 1..=rng.gen_range(1..=100) {
|
|
// Just use the same message multiple times per block.
|
|
msgs.push(mk_hrmp(block));
|
|
}
|
|
sent_in_block.push(msgs);
|
|
}
|
|
SentInBlock::set(&sent_in_block);
|
|
|
|
let mut tester = BlockTests::new_without_externalities()
|
|
.with_relay_sproof_builder(|_, relay_block_num, sproof| {
|
|
let mut new_hash = MqcHead::get();
|
|
|
|
for msg in SentInBlock::get()[relay_block_num as usize].iter() {
|
|
new_hash.extend_hrmp(&msg);
|
|
}
|
|
|
|
sproof.upsert_inbound_channel(ALICE).mqc_head = Some(new_hash.head());
|
|
MqcHead::set(&new_hash);
|
|
})
|
|
.with_inherent_data(|_, relay_block_num, data| {
|
|
// TODO use vector for dmp as well
|
|
data.horizontal_messages
|
|
.insert(ALICE, SentInBlock::get()[relay_block_num as usize].clone());
|
|
});
|
|
|
|
for block in 1..100 {
|
|
tester = tester.add(block, move || {
|
|
HANDLED_XCMP_MESSAGES.with(|m| {
|
|
let mut m = m.borrow_mut();
|
|
let msgs = SentInBlock::get()[block as usize]
|
|
.iter()
|
|
.map(|m| (ALICE, m.sent_at, m.data.clone()))
|
|
.collect::<Vec<_>>();
|
|
assert_eq!(&*m, &msgs);
|
|
m.clear();
|
|
});
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn upgrade_version_checks_should_work() {
|
|
use codec::Encode;
|
|
use sp_runtime::DispatchErrorWithPostInfo;
|
|
use sp_version::RuntimeVersion;
|
|
|
|
let test_data = vec![
|
|
("test", 0, 1, Err(frame_system::Error::<Test>::SpecVersionNeedsToIncrease)),
|
|
("test", 1, 0, Err(frame_system::Error::<Test>::SpecVersionNeedsToIncrease)),
|
|
("test", 1, 1, Err(frame_system::Error::<Test>::SpecVersionNeedsToIncrease)),
|
|
("test", 1, 2, Err(frame_system::Error::<Test>::SpecVersionNeedsToIncrease)),
|
|
("test2", 1, 1, Err(frame_system::Error::<Test>::InvalidSpecName)),
|
|
];
|
|
|
|
for (spec_name, spec_version, impl_version, expected) in test_data.into_iter() {
|
|
let version = RuntimeVersion {
|
|
spec_name: spec_name.into(),
|
|
spec_version,
|
|
impl_version,
|
|
..Default::default()
|
|
};
|
|
let read_runtime_version = ReadRuntimeVersion(version.encode());
|
|
|
|
let mut ext = new_test_ext();
|
|
ext.register_extension(sp_core::traits::ReadRuntimeVersionExt::new(read_runtime_version));
|
|
ext.execute_with(|| {
|
|
let new_code = vec![1, 2, 3, 4];
|
|
let new_code_hash = H256(sp_core::blake2_256(&new_code));
|
|
|
|
let _authorize =
|
|
ParachainSystem::authorize_upgrade(RawOrigin::Root.into(), new_code_hash, true);
|
|
let res = ParachainSystem::enact_authorized_upgrade(RawOrigin::None.into(), new_code);
|
|
|
|
assert_eq!(expected.map_err(DispatchErrorWithPostInfo::from), res);
|
|
});
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn deposits_relay_parent_storage_root() {
|
|
BlockTests::new().add_with_post_test(
|
|
123,
|
|
|| {},
|
|
|| {
|
|
let digest = System::digest();
|
|
assert!(cumulus_primitives_core::rpsr_digest::extract_relay_parent_storage_root(
|
|
&digest
|
|
)
|
|
.is_some());
|
|
},
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn ump_fee_factor_increases_and_decreases() {
|
|
BlockTests::new()
|
|
.with_relay_sproof_builder(|_, _, sproof| {
|
|
sproof.host_config.max_upward_queue_size = 100;
|
|
sproof.host_config.max_upward_message_num_per_candidate = 1;
|
|
})
|
|
.add_with_post_test(
|
|
1,
|
|
|| {
|
|
// Fee factor increases in `send_upward_message`
|
|
ParachainSystem::send_upward_message(b"Test".to_vec()).unwrap();
|
|
assert_eq!(UpwardDeliveryFeeFactor::<Test>::get(), FixedU128::from_u32(1));
|
|
|
|
ParachainSystem::send_upward_message(
|
|
b"This message will be enough to increase the fee factor".to_vec(),
|
|
)
|
|
.unwrap();
|
|
assert_eq!(
|
|
UpwardDeliveryFeeFactor::<Test>::get(),
|
|
FixedU128::from_rational(105, 100)
|
|
);
|
|
},
|
|
|| {
|
|
// Factor decreases in `on_finalize`, but only if we are below the threshold
|
|
let messages = UpwardMessages::<Test>::get();
|
|
assert_eq!(messages, vec![b"Test".to_vec()]);
|
|
assert_eq!(
|
|
UpwardDeliveryFeeFactor::<Test>::get(),
|
|
FixedU128::from_rational(105, 100)
|
|
);
|
|
},
|
|
)
|
|
.add_with_post_test(
|
|
2,
|
|
|| {
|
|
// We do nothing here
|
|
},
|
|
|| {
|
|
let messages = UpwardMessages::<Test>::get();
|
|
assert_eq!(
|
|
messages,
|
|
vec![b"This message will be enough to increase the fee factor".to_vec(),]
|
|
);
|
|
// Now the delivery fee factor is decreased, since we are below the threshold
|
|
assert_eq!(UpwardDeliveryFeeFactor::<Test>::get(), FixedU128::from_u32(1));
|
|
},
|
|
);
|
|
}
|