mirror of
https://github.com/pezkuwichain/pezkuwi-subxt.git
synced 2026-04-26 14:37:57 +00:00
93e8ffed55
* Introduce preimages module in traits * Multisize Preimages * Len not actually necessary * Tweaks to the preimage API * Fixes * Get Scheduler building with new API * Scheduler tests pass * Bounded Scheduler 🎉 * Use Agenda holes and introduce IncompleteSince to avoid need to reschedule * Tests pass with new weight system * New benchmarks * Add missing file * Drop preimage when permenantly overeight * Drop preimage when permenantly overeight * Referenda uses latest preimage API * Testing ok * Adding tests Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * fmt Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Add preimage migration Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Docs * Remove dbg * Refactor Democracy * Refactor Democracy * Add final MEL * Remove silly maps * Fixes * Minor refactor * Formatting * Fixes * Fixes * Fixes * Update frame/preimage/src/lib.rs Co-authored-by: Shawn Tabrizi <shawntabrizi@gmail.com> * Add migrations to Democracy * WIP Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Resolve conflicts Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Revert "Resolve conflicts" This reverts commit 734d66d69e54553471ffa54fa52e3e304dc8f106. * Undo wrong resolves... Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * WIP Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Make compile Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * massage clippy Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * More clippy Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * clippy annoyance Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * clippy annoyance Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix benchmarks Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * add missing file * Test <Preimage as QueryPreimage> Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * More tests Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Clippy harassment Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Add test Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * clippy Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fixup tests Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Remove old stuff Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * fmt Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Test <Scheduler as Anon> trait functions Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Update pallet-ui tests Why is this needed? Should not be the case unless master is broken... Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * More scheduler trait test Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * More tests Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Apply review suggestion Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Beauty fixes Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Add Scheduler test migration_v3_to_v4_works Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Merge fixup Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Keep referenda benchmarks instantiatable Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Update weights Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Use new scheduler weight functions Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Use new democracy weight functions Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Use weight compare functions Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Update pallet-ui tests Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * More renaming… Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * More renaming… Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Add comment Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Implement OnRuntimeUpgrade for scheduler::v3_to_v4 migration Put the migration into a proper `MigrateToV4` struct and implement the OnRuntimeUpgrade hooks for it. Also move the test to use that instead. This should make it easier for adding it to Polkadot. Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Clippy Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Handle undecodable Agendas Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Remove trash Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix test Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Use new OnRuntimeUpgrade functions Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * fix test Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix BoundedSlice::truncate_from Co-authored-by: jakoblell Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix pre_upgrade hook return values Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Add more error logging Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Find too large preimages in the pre_upgrade hook Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Test that too large Calls in agendas are ignored Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Use new OnRuntimeUpgrade hooks Why did the CI not catch this?! Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * works fine - just more logs Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix staking migration Causing issues on Kusama... Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix UI tests No idea why this is needed. This is actually undoing an earlier change. Maybe the CI has different rustc versions!? Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Remove multisig's Calls (#12072) * Remove multisig's Calls * Multisig: Fix tests and re-introduce reserve logic (#12241) * Fix tests and re-introduce reserve logic * fix benches * add todo * remove irrelevant bench * [Feature] Add a migration that drains and refunds stored calls (#12313) * [Feature] Add a migration that drains and refunds stored calls * migration fixes * fixes * address review comments * consume the whole block weight * fix assertions * license header * fix interface Co-authored-by: parity-processbot <> Co-authored-by: parity-processbot <> Co-authored-by: Roman Useinov <roman.useinov@gmail.com> * Fix test Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * Fix multisig benchmarks Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> * ".git/.scripts/bench-bot.sh" pallet dev pallet_democracy * ".git/.scripts/bench-bot.sh" pallet dev pallet_scheduler * ".git/.scripts/bench-bot.sh" pallet dev pallet_preimage Signed-off-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> Co-authored-by: Oliver Tale-Yazdi <oliver.tale-yazdi@parity.io> Co-authored-by: Shawn Tabrizi <shawntabrizi@gmail.com> Co-authored-by: parity-processbot <> Co-authored-by: Roman Useinov <roman.useinov@gmail.com>
318 lines
10 KiB
Rust
318 lines
10 KiB
Rust
// This file is part of Substrate.
|
|
|
|
// Copyright (C) 2019-2022 Parity Technologies (UK) Ltd.
|
|
// SPDX-License-Identifier: Apache-2.0
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
//! Stuff for dealing with 32-byte hashed preimages.
|
|
|
|
use codec::{Decode, Encode, EncodeLike, MaxEncodedLen};
|
|
use sp_core::{RuntimeDebug, H256};
|
|
use sp_io::hashing::blake2_256;
|
|
use sp_runtime::{traits::ConstU32, DispatchError};
|
|
use sp_std::borrow::Cow;
|
|
|
|
pub type Hash = H256;
|
|
pub type BoundedInline = crate::BoundedVec<u8, ConstU32<128>>;
|
|
|
|
#[derive(
|
|
Encode, Decode, MaxEncodedLen, Clone, Eq, PartialEq, scale_info::TypeInfo, RuntimeDebug,
|
|
)]
|
|
#[codec(mel_bound())]
|
|
pub enum Bounded<T> {
|
|
/// A Blake2 256 hash with no preimage length. We
|
|
/// do not support creation of this except for transitioning from legacy state.
|
|
/// In the future we will make this a pure `Dummy` item storing only the final `dummy` field.
|
|
Legacy { hash: Hash, dummy: sp_std::marker::PhantomData<T> },
|
|
/// A an bounded `Call`. Its encoding must be at most 128 bytes.
|
|
Inline(BoundedInline),
|
|
/// A Blake2-256 hash of the call together with an upper limit for its size.
|
|
Lookup { hash: Hash, len: u32 },
|
|
}
|
|
|
|
impl<T> Bounded<T> {
|
|
/// Casts the wrapped type into something that encodes alike.
|
|
///
|
|
/// # Examples
|
|
/// ```
|
|
/// use frame_support::traits::Bounded;
|
|
///
|
|
/// // Transmute from `String` to `&str`.
|
|
/// let x: Bounded<String> = Bounded::Inline(Default::default());
|
|
/// let _: Bounded<&str> = x.transmute();
|
|
/// ```
|
|
pub fn transmute<S: Encode>(self) -> Bounded<S>
|
|
where
|
|
T: Encode + EncodeLike<S>,
|
|
{
|
|
use Bounded::*;
|
|
match self {
|
|
Legacy { hash, .. } => Legacy { hash, dummy: sp_std::marker::PhantomData },
|
|
Inline(x) => Inline(x),
|
|
Lookup { hash, len } => Lookup { hash, len },
|
|
}
|
|
}
|
|
|
|
/// Returns the hash of the preimage.
|
|
///
|
|
/// The hash is re-calculated every time if the preimage is inlined.
|
|
pub fn hash(&self) -> H256 {
|
|
use Bounded::*;
|
|
match self {
|
|
Legacy { hash, .. } => *hash,
|
|
Inline(x) => blake2_256(x.as_ref()).into(),
|
|
Lookup { hash, .. } => *hash,
|
|
}
|
|
}
|
|
}
|
|
|
|
// The maximum we expect a single legacy hash lookup to be.
|
|
const MAX_LEGACY_LEN: u32 = 1_000_000;
|
|
|
|
impl<T> Bounded<T> {
|
|
/// Returns the length of the preimage or `None` if the length is unknown.
|
|
pub fn len(&self) -> Option<u32> {
|
|
match self {
|
|
Self::Legacy { .. } => None,
|
|
Self::Inline(i) => Some(i.len() as u32),
|
|
Self::Lookup { len, .. } => Some(*len),
|
|
}
|
|
}
|
|
|
|
/// Returns whether the image will require a lookup to be peeked.
|
|
pub fn lookup_needed(&self) -> bool {
|
|
match self {
|
|
Self::Inline(..) => false,
|
|
Self::Legacy { .. } | Self::Lookup { .. } => true,
|
|
}
|
|
}
|
|
|
|
/// The maximum length of the lookup that is needed to peek `Self`.
|
|
pub fn lookup_len(&self) -> Option<u32> {
|
|
match self {
|
|
Self::Inline(..) => None,
|
|
Self::Legacy { .. } => Some(MAX_LEGACY_LEN),
|
|
Self::Lookup { len, .. } => Some(*len),
|
|
}
|
|
}
|
|
|
|
/// Constructs a `Lookup` bounded item.
|
|
pub fn unrequested(hash: Hash, len: u32) -> Self {
|
|
Self::Lookup { hash, len }
|
|
}
|
|
|
|
/// Constructs a `Legacy` bounded item.
|
|
#[deprecated = "This API is only for transitioning to Scheduler v3 API"]
|
|
pub fn from_legacy_hash(hash: impl Into<Hash>) -> Self {
|
|
Self::Legacy { hash: hash.into(), dummy: sp_std::marker::PhantomData }
|
|
}
|
|
}
|
|
|
|
pub type FetchResult = Result<Cow<'static, [u8]>, DispatchError>;
|
|
|
|
/// A interface for looking up preimages from their hash on chain.
|
|
pub trait QueryPreimage {
|
|
/// Returns whether a preimage exists for a given hash and if so its length.
|
|
fn len(hash: &Hash) -> Option<u32>;
|
|
|
|
/// Returns the preimage for a given hash. If given, `len` must be the size of the preimage.
|
|
fn fetch(hash: &Hash, len: Option<u32>) -> FetchResult;
|
|
|
|
/// Returns whether a preimage request exists for a given hash.
|
|
fn is_requested(hash: &Hash) -> bool;
|
|
|
|
/// Request that someone report a preimage. Providers use this to optimise the economics for
|
|
/// preimage reporting.
|
|
fn request(hash: &Hash);
|
|
|
|
/// Cancel a previous preimage request.
|
|
fn unrequest(hash: &Hash);
|
|
|
|
/// Request that the data required for decoding the given `bounded` value is made available.
|
|
fn hold<T>(bounded: &Bounded<T>) {
|
|
use Bounded::*;
|
|
match bounded {
|
|
Inline(..) => {},
|
|
Legacy { hash, .. } | Lookup { hash, .. } => Self::request(hash),
|
|
}
|
|
}
|
|
|
|
/// No longer request that the data required for decoding the given `bounded` value is made
|
|
/// available.
|
|
fn drop<T>(bounded: &Bounded<T>) {
|
|
use Bounded::*;
|
|
match bounded {
|
|
Inline(..) => {},
|
|
Legacy { hash, .. } | Lookup { hash, .. } => Self::unrequest(hash),
|
|
}
|
|
}
|
|
|
|
/// Check to see if all data required for the given `bounded` value is available for its
|
|
/// decoding.
|
|
fn have<T>(bounded: &Bounded<T>) -> bool {
|
|
use Bounded::*;
|
|
match bounded {
|
|
Inline(..) => true,
|
|
Legacy { hash, .. } | Lookup { hash, .. } => Self::len(hash).is_some(),
|
|
}
|
|
}
|
|
|
|
/// Create a `Bounded` instance based on the `hash` and `len` of the encoded value. This may not
|
|
/// be `peek`-able or `realize`-able.
|
|
fn pick<T>(hash: Hash, len: u32) -> Bounded<T> {
|
|
Self::request(&hash);
|
|
Bounded::Lookup { hash, len }
|
|
}
|
|
|
|
/// Convert the given `bounded` instance back into its original instance, also returning the
|
|
/// exact size of its encoded form if it needed to be looked-up from a stored preimage).
|
|
///
|
|
/// NOTE: This does not remove any data needed for realization. If you will no longer use the
|
|
/// `bounded`, call `realize` instead or call `drop` afterwards.
|
|
fn peek<T: Decode>(bounded: &Bounded<T>) -> Result<(T, Option<u32>), DispatchError> {
|
|
use Bounded::*;
|
|
match bounded {
|
|
Inline(data) => T::decode(&mut &data[..]).ok().map(|x| (x, None)),
|
|
Lookup { hash, len } => {
|
|
let data = Self::fetch(hash, Some(*len))?;
|
|
T::decode(&mut &data[..]).ok().map(|x| (x, Some(data.len() as u32)))
|
|
},
|
|
Legacy { hash, .. } => {
|
|
let data = Self::fetch(hash, None)?;
|
|
T::decode(&mut &data[..]).ok().map(|x| (x, Some(data.len() as u32)))
|
|
},
|
|
}
|
|
.ok_or(DispatchError::Corruption)
|
|
}
|
|
|
|
/// Convert the given `bounded` value back into its original instance. If successful,
|
|
/// `drop` any data backing it. This will not break the realisability of independently
|
|
/// created instances of `Bounded` which happen to have identical data.
|
|
fn realize<T: Decode>(bounded: &Bounded<T>) -> Result<(T, Option<u32>), DispatchError> {
|
|
let r = Self::peek(bounded)?;
|
|
Self::drop(bounded);
|
|
Ok(r)
|
|
}
|
|
}
|
|
|
|
/// A interface for managing preimages to hashes on chain.
|
|
///
|
|
/// Note that this API does not assume any underlying user is calling, and thus
|
|
/// does not handle any preimage ownership or fees. Other system level logic that
|
|
/// uses this API should implement that on their own side.
|
|
pub trait StorePreimage: QueryPreimage {
|
|
/// The maximum length of preimage we can store.
|
|
///
|
|
/// This is the maximum length of the *encoded* value that can be passed to `bound`.
|
|
const MAX_LENGTH: usize;
|
|
|
|
/// Request and attempt to store the bytes of a preimage on chain.
|
|
///
|
|
/// May return `DispatchError::Exhausted` if the preimage is just too big.
|
|
fn note(bytes: Cow<[u8]>) -> Result<Hash, DispatchError>;
|
|
|
|
/// Attempt to clear a previously noted preimage. Exactly the same as `unrequest` but is
|
|
/// provided for symmetry.
|
|
fn unnote(hash: &Hash) {
|
|
Self::unrequest(hash)
|
|
}
|
|
|
|
/// Convert an otherwise unbounded or large value into a type ready for placing in storage. The
|
|
/// result is a type whose `MaxEncodedLen` is 131 bytes.
|
|
///
|
|
/// NOTE: Once this API is used, you should use either `drop` or `realize`.
|
|
fn bound<T: Encode>(t: T) -> Result<Bounded<T>, DispatchError> {
|
|
let data = t.encode();
|
|
let len = data.len() as u32;
|
|
Ok(match BoundedInline::try_from(data) {
|
|
Ok(bounded) => Bounded::Inline(bounded),
|
|
Err(unbounded) => Bounded::Lookup { hash: Self::note(unbounded.into())?, len },
|
|
})
|
|
}
|
|
}
|
|
|
|
impl QueryPreimage for () {
|
|
fn len(_: &Hash) -> Option<u32> {
|
|
None
|
|
}
|
|
fn fetch(_: &Hash, _: Option<u32>) -> FetchResult {
|
|
Err(DispatchError::Unavailable)
|
|
}
|
|
fn is_requested(_: &Hash) -> bool {
|
|
false
|
|
}
|
|
fn request(_: &Hash) {}
|
|
fn unrequest(_: &Hash) {}
|
|
}
|
|
|
|
impl StorePreimage for () {
|
|
const MAX_LENGTH: usize = 0;
|
|
fn note(_: Cow<[u8]>) -> Result<Hash, DispatchError> {
|
|
Err(DispatchError::Exhausted)
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
use crate::{bounded_vec, BoundedVec};
|
|
|
|
#[test]
|
|
fn bounded_size_is_correct() {
|
|
assert_eq!(<Bounded<Vec<u8>> as MaxEncodedLen>::max_encoded_len(), 131);
|
|
}
|
|
|
|
#[test]
|
|
fn bounded_basic_works() {
|
|
let data: BoundedVec<u8, _> = bounded_vec![b'a', b'b', b'c'];
|
|
let len = data.len() as u32;
|
|
let hash = blake2_256(&data).into();
|
|
|
|
// Inline works
|
|
{
|
|
let bound: Bounded<Vec<u8>> = Bounded::Inline(data.clone());
|
|
assert_eq!(bound.hash(), hash);
|
|
assert_eq!(bound.len(), Some(len));
|
|
assert!(!bound.lookup_needed());
|
|
assert_eq!(bound.lookup_len(), None);
|
|
}
|
|
// Legacy works
|
|
{
|
|
let bound: Bounded<Vec<u8>> = Bounded::Legacy { hash, dummy: Default::default() };
|
|
assert_eq!(bound.hash(), hash);
|
|
assert_eq!(bound.len(), None);
|
|
assert!(bound.lookup_needed());
|
|
assert_eq!(bound.lookup_len(), Some(1_000_000));
|
|
}
|
|
// Lookup works
|
|
{
|
|
let bound: Bounded<Vec<u8>> = Bounded::Lookup { hash, len: data.len() as u32 };
|
|
assert_eq!(bound.hash(), hash);
|
|
assert_eq!(bound.len(), Some(len));
|
|
assert!(bound.lookup_needed());
|
|
assert_eq!(bound.lookup_len(), Some(len));
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn bounded_transmuting_works() {
|
|
let data: BoundedVec<u8, _> = bounded_vec![b'a', b'b', b'c'];
|
|
|
|
// Transmute a `String` into a `&str`.
|
|
let x: Bounded<String> = Bounded::Inline(data.clone());
|
|
let y: Bounded<&str> = x.transmute();
|
|
assert_eq!(y, Bounded::Inline(data));
|
|
}
|
|
}
|