// This file is part of Substrate. // Copyright (C) Parity Technologies (UK) Ltd. // SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0 // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see . //! A consensus proposer for "basic" chains which use the primitive inherent-data. // FIXME #1021 move this into sp-consensus use codec::Encode; use futures::{ channel::oneshot, future, future::{Future, FutureExt}, select, }; use log::{debug, error, info, trace, warn}; use sc_block_builder::{BlockBuilderApi, BlockBuilderBuilder}; use sc_telemetry::{telemetry, TelemetryHandle, CONSENSUS_INFO}; use sc_transaction_pool_api::{InPoolTransaction, TransactionPool}; use sp_api::{ApiExt, CallApiAt, ProvideRuntimeApi}; use sp_blockchain::{ApplyExtrinsicFailed::Validity, Error::ApplyExtrinsicFailed, HeaderBackend}; use sp_consensus::{DisableProofRecording, EnableProofRecording, ProofRecording, Proposal}; use sp_core::traits::SpawnNamed; use sp_inherents::InherentData; use sp_runtime::{ traits::{BlakeTwo256, Block as BlockT, Hash as HashT, Header as HeaderT}, Digest, ExtrinsicInclusionMode, Percent, SaturatedConversion, }; use std::{marker::PhantomData, pin::Pin, sync::Arc, time}; use prometheus_endpoint::Registry as PrometheusRegistry; use sc_proposer_metrics::{EndProposingReason, MetricsLink as PrometheusMetrics}; /// Default block size limit in bytes used by [`Proposer`]. /// /// Can be overwritten by [`ProposerFactory::set_default_block_size_limit`]. /// /// Be aware that there is also an upper packet size on what the networking code /// will accept. If the block doesn't fit in such a package, it can not be /// transferred to other nodes. pub const DEFAULT_BLOCK_SIZE_LIMIT: usize = 4 * 1024 * 1024 + 512; const DEFAULT_SOFT_DEADLINE_PERCENT: Percent = Percent::from_percent(50); const LOG_TARGET: &'static str = "basic-authorship"; /// [`Proposer`] factory. pub struct ProposerFactory { spawn_handle: Box, /// The client instance. client: Arc, /// The transaction pool. transaction_pool: Arc, /// Prometheus Link, metrics: PrometheusMetrics, /// The default block size limit. /// /// If no `block_size_limit` is passed to [`sp_consensus::Proposer::propose`], this block size /// limit will be used. default_block_size_limit: usize, /// Soft deadline percentage of hard deadline. /// /// The value is used to compute soft deadline during block production. /// The soft deadline indicates where we should stop attempting to add transactions /// to the block, which exhaust resources. After soft deadline is reached, /// we switch to a fixed-amount mode, in which after we see `MAX_SKIPPED_TRANSACTIONS` /// transactions which exhaust resources, we will conclude that the block is full. soft_deadline_percent: Percent, telemetry: Option, /// When estimating the block size, should the proof be included? include_proof_in_block_size_estimation: bool, /// phantom member to pin the `ProofRecording` type. _phantom: PhantomData, } impl Clone for ProposerFactory { fn clone(&self) -> Self { Self { spawn_handle: self.spawn_handle.clone(), client: self.client.clone(), transaction_pool: self.transaction_pool.clone(), metrics: self.metrics.clone(), default_block_size_limit: self.default_block_size_limit, soft_deadline_percent: self.soft_deadline_percent, telemetry: self.telemetry.clone(), include_proof_in_block_size_estimation: self.include_proof_in_block_size_estimation, _phantom: self._phantom, } } } impl ProposerFactory { /// Create a new proposer factory. /// /// Proof recording will be disabled when using proposers built by this instance to build /// blocks. pub fn new( spawn_handle: impl SpawnNamed + 'static, client: Arc, transaction_pool: Arc, prometheus: Option<&PrometheusRegistry>, telemetry: Option, ) -> Self { ProposerFactory { spawn_handle: Box::new(spawn_handle), transaction_pool, metrics: PrometheusMetrics::new(prometheus), default_block_size_limit: DEFAULT_BLOCK_SIZE_LIMIT, soft_deadline_percent: DEFAULT_SOFT_DEADLINE_PERCENT, telemetry, client, include_proof_in_block_size_estimation: false, _phantom: PhantomData, } } } impl ProposerFactory { /// Create a new proposer factory with proof recording enabled. /// /// Each proposer created by this instance will record a proof while building a block. /// /// This will also include the proof into the estimation of the block size. This can be disabled /// by calling [`ProposerFactory::disable_proof_in_block_size_estimation`]. pub fn with_proof_recording( spawn_handle: impl SpawnNamed + 'static, client: Arc, transaction_pool: Arc, prometheus: Option<&PrometheusRegistry>, telemetry: Option, ) -> Self { ProposerFactory { client, spawn_handle: Box::new(spawn_handle), transaction_pool, metrics: PrometheusMetrics::new(prometheus), default_block_size_limit: DEFAULT_BLOCK_SIZE_LIMIT, soft_deadline_percent: DEFAULT_SOFT_DEADLINE_PERCENT, telemetry, include_proof_in_block_size_estimation: true, _phantom: PhantomData, } } /// Disable the proof inclusion when estimating the block size. pub fn disable_proof_in_block_size_estimation(&mut self) { self.include_proof_in_block_size_estimation = false; } } impl ProposerFactory { /// Set the default block size limit in bytes. /// /// The default value for the block size limit is: /// [`DEFAULT_BLOCK_SIZE_LIMIT`]. /// /// If there is no block size limit passed to [`sp_consensus::Proposer::propose`], this value /// will be used. pub fn set_default_block_size_limit(&mut self, limit: usize) { self.default_block_size_limit = limit; } /// Set soft deadline percentage. /// /// The value is used to compute soft deadline during block production. /// The soft deadline indicates where we should stop attempting to add transactions /// to the block, which exhaust resources. After soft deadline is reached, /// we switch to a fixed-amount mode, in which after we see `MAX_SKIPPED_TRANSACTIONS` /// transactions which exhaust resources, we will conclude that the block is full. /// /// Setting the value too low will significantly limit the amount of transactions /// we try in case they exhaust resources. Setting the value too high can /// potentially open a DoS vector, where many "exhaust resources" transactions /// are being tried with no success, hence block producer ends up creating an empty block. pub fn set_soft_deadline(&mut self, percent: Percent) { self.soft_deadline_percent = percent; } } impl ProposerFactory where A: TransactionPool + 'static, Block: BlockT, C: HeaderBackend + ProvideRuntimeApi + Send + Sync + 'static, C::Api: ApiExt + BlockBuilderApi, { fn init_with_now( &mut self, parent_header: &::Header, now: Box time::Instant + Send + Sync>, ) -> Proposer { let parent_hash = parent_header.hash(); info!("🙌 Starting consensus session on top of parent {:?}", parent_hash); let proposer = Proposer::<_, _, _, PR> { spawn_handle: self.spawn_handle.clone(), client: self.client.clone(), parent_hash, parent_number: *parent_header.number(), transaction_pool: self.transaction_pool.clone(), now, metrics: self.metrics.clone(), default_block_size_limit: self.default_block_size_limit, soft_deadline_percent: self.soft_deadline_percent, telemetry: self.telemetry.clone(), _phantom: PhantomData, include_proof_in_block_size_estimation: self.include_proof_in_block_size_estimation, }; proposer } } impl sp_consensus::Environment for ProposerFactory where A: TransactionPool + 'static, Block: BlockT, C: HeaderBackend + ProvideRuntimeApi + CallApiAt + Send + Sync + 'static, C::Api: ApiExt + BlockBuilderApi, PR: ProofRecording, { type CreateProposer = future::Ready>; type Proposer = Proposer; type Error = sp_blockchain::Error; fn init(&mut self, parent_header: &::Header) -> Self::CreateProposer { future::ready(Ok(self.init_with_now(parent_header, Box::new(time::Instant::now)))) } } /// The proposer logic. pub struct Proposer { spawn_handle: Box, client: Arc, parent_hash: Block::Hash, parent_number: <::Header as HeaderT>::Number, transaction_pool: Arc, now: Box time::Instant + Send + Sync>, metrics: PrometheusMetrics, default_block_size_limit: usize, include_proof_in_block_size_estimation: bool, soft_deadline_percent: Percent, telemetry: Option, _phantom: PhantomData, } impl sp_consensus::Proposer for Proposer where A: TransactionPool + 'static, Block: BlockT, C: HeaderBackend + ProvideRuntimeApi + CallApiAt + Send + Sync + 'static, C::Api: ApiExt + BlockBuilderApi, PR: ProofRecording, { type Proposal = Pin, Self::Error>> + Send>>; type Error = sp_blockchain::Error; type ProofRecording = PR; type Proof = PR::Proof; fn propose( self, inherent_data: InherentData, inherent_digests: Digest, max_duration: time::Duration, block_size_limit: Option, ) -> Self::Proposal { let (tx, rx) = oneshot::channel(); let spawn_handle = self.spawn_handle.clone(); spawn_handle.spawn_blocking( "basic-authorship-proposer", None, Box::pin(async move { // leave some time for evaluation and block finalization (33%) let deadline = (self.now)() + max_duration - max_duration / 3; let res = self .propose_with(inherent_data, inherent_digests, deadline, block_size_limit) .await; if tx.send(res).is_err() { trace!( target: LOG_TARGET, "Could not send block production result to proposer!" ); } }), ); async move { rx.await? }.boxed() } } /// If the block is full we will attempt to push at most /// this number of transactions before quitting for real. /// It allows us to increase block utilization. const MAX_SKIPPED_TRANSACTIONS: usize = 8; impl Proposer where A: TransactionPool, Block: BlockT, C: HeaderBackend + ProvideRuntimeApi + CallApiAt + Send + Sync + 'static, C::Api: ApiExt + BlockBuilderApi, PR: ProofRecording, { async fn propose_with( self, inherent_data: InherentData, inherent_digests: Digest, deadline: time::Instant, block_size_limit: Option, ) -> Result, sp_blockchain::Error> { let block_timer = time::Instant::now(); let mut block_builder = BlockBuilderBuilder::new(&*self.client) .on_parent_block(self.parent_hash) .with_parent_block_number(self.parent_number) .with_proof_recording(PR::ENABLED) .with_inherent_digests(inherent_digests) .build()?; self.apply_inherents(&mut block_builder, inherent_data)?; let mode = block_builder.extrinsic_inclusion_mode(); let end_reason = match mode { ExtrinsicInclusionMode::AllExtrinsics => self.apply_extrinsics(&mut block_builder, deadline, block_size_limit).await?, ExtrinsicInclusionMode::OnlyInherents => EndProposingReason::TransactionForbidden, }; let (block, storage_changes, proof) = block_builder.build()?.into_inner(); let block_took = block_timer.elapsed(); let proof = PR::into_proof(proof).map_err(|e| sp_blockchain::Error::Application(Box::new(e)))?; self.print_summary(&block, end_reason, block_took, block_timer.elapsed()); Ok(Proposal { block, proof, storage_changes }) } /// Apply all inherents to the block. fn apply_inherents( &self, block_builder: &mut sc_block_builder::BlockBuilder<'_, Block, C>, inherent_data: InherentData, ) -> Result<(), sp_blockchain::Error> { let create_inherents_start = time::Instant::now(); let inherents = block_builder.create_inherents(inherent_data)?; let create_inherents_end = time::Instant::now(); self.metrics.report(|metrics| { metrics.create_inherents_time.observe( create_inherents_end .saturating_duration_since(create_inherents_start) .as_secs_f64(), ); }); for inherent in inherents { match block_builder.push(inherent) { Err(ApplyExtrinsicFailed(Validity(e))) if e.exhausted_resources() => { warn!( target: LOG_TARGET, "⚠️ Dropping non-mandatory inherent from overweight block." ) }, Err(ApplyExtrinsicFailed(Validity(e))) if e.was_mandatory() => { error!( "❌️ Mandatory inherent extrinsic returned error. Block cannot be produced." ); return Err(ApplyExtrinsicFailed(Validity(e))) }, Err(e) => { warn!( target: LOG_TARGET, "❗️ Inherent extrinsic returned unexpected error: {}. Dropping.", e ); }, Ok(_) => {}, } } Ok(()) } /// Apply as many extrinsics as possible to the block. async fn apply_extrinsics( &self, block_builder: &mut sc_block_builder::BlockBuilder<'_, Block, C>, deadline: time::Instant, block_size_limit: Option, ) -> Result { // proceed with transactions // We calculate soft deadline used only in case we start skipping transactions. let now = (self.now)(); let left = deadline.saturating_duration_since(now); let left_micros: u64 = left.as_micros().saturated_into(); let soft_deadline = now + time::Duration::from_micros(self.soft_deadline_percent.mul_floor(left_micros)); let mut skipped = 0; let mut unqueue_invalid = Vec::new(); let mut t1 = self.transaction_pool.ready_at(self.parent_number).fuse(); let mut t2 = futures_timer::Delay::new(deadline.saturating_duration_since((self.now)()) / 8).fuse(); let mut pending_iterator = select! { res = t1 => res, _ = t2 => { warn!(target: LOG_TARGET, "Timeout fired waiting for transaction pool at block #{}. \ Proceeding with production.", self.parent_number, ); self.transaction_pool.ready() }, }; let block_size_limit = block_size_limit.unwrap_or(self.default_block_size_limit); debug!(target: LOG_TARGET, "Attempting to push transactions from the pool."); debug!(target: LOG_TARGET, "Pool status: {:?}", self.transaction_pool.status()); let mut transaction_pushed = false; let end_reason = loop { let pending_tx = if let Some(pending_tx) = pending_iterator.next() { pending_tx } else { debug!( target: LOG_TARGET, "No more transactions, proceeding with proposing." ); break EndProposingReason::NoMoreTransactions }; let now = (self.now)(); if now > deadline { debug!( target: LOG_TARGET, "Consensus deadline reached when pushing block transactions, \ proceeding with proposing." ); break EndProposingReason::HitDeadline } let pending_tx_data = pending_tx.data().clone(); let pending_tx_hash = pending_tx.hash().clone(); let block_size = block_builder.estimate_block_size(self.include_proof_in_block_size_estimation); if block_size + pending_tx_data.encoded_size() > block_size_limit { pending_iterator.report_invalid(&pending_tx); if skipped < MAX_SKIPPED_TRANSACTIONS { skipped += 1; debug!( target: LOG_TARGET, "Transaction would overflow the block size limit, \ but will try {} more transactions before quitting.", MAX_SKIPPED_TRANSACTIONS - skipped, ); continue } else if now < soft_deadline { debug!( target: LOG_TARGET, "Transaction would overflow the block size limit, \ but we still have time before the soft deadline, so \ we will try a bit more." ); continue } else { debug!( target: LOG_TARGET, "Reached block size limit, proceeding with proposing." ); break EndProposingReason::HitBlockSizeLimit } } trace!(target: LOG_TARGET, "[{:?}] Pushing to the block.", pending_tx_hash); match sc_block_builder::BlockBuilder::push(block_builder, pending_tx_data) { Ok(()) => { transaction_pushed = true; debug!(target: LOG_TARGET, "[{:?}] Pushed to the block.", pending_tx_hash); }, Err(ApplyExtrinsicFailed(Validity(e))) if e.exhausted_resources() => { pending_iterator.report_invalid(&pending_tx); if skipped < MAX_SKIPPED_TRANSACTIONS { skipped += 1; debug!(target: LOG_TARGET, "Block seems full, but will try {} more transactions before quitting.", MAX_SKIPPED_TRANSACTIONS - skipped, ); } else if (self.now)() < soft_deadline { debug!(target: LOG_TARGET, "Block seems full, but we still have time before the soft deadline, \ so we will try a bit more before quitting." ); } else { debug!( target: LOG_TARGET, "Reached block weight limit, proceeding with proposing." ); break EndProposingReason::HitBlockWeightLimit } }, Err(e) => { pending_iterator.report_invalid(&pending_tx); debug!( target: LOG_TARGET, "[{:?}] Invalid transaction: {}", pending_tx_hash, e ); unqueue_invalid.push(pending_tx_hash); }, } }; if matches!(end_reason, EndProposingReason::HitBlockSizeLimit) && !transaction_pushed { warn!( target: LOG_TARGET, "Hit block size limit of `{}` without including any transaction!", block_size_limit, ); } self.transaction_pool.remove_invalid(&unqueue_invalid); Ok(end_reason) } /// Prints a summary and does telemetry + metrics. /// /// - `block`: The block that was build. /// - `end_reason`: Why did we stop producing the block? /// - `block_took`: How long did it took to produce the actual block? /// - `propose_took`: How long did the entire proposing took? fn print_summary( &self, block: &Block, end_reason: EndProposingReason, block_took: time::Duration, propose_took: time::Duration, ) { let extrinsics = block.extrinsics(); self.metrics.report(|metrics| { metrics.number_of_transactions.set(extrinsics.len() as u64); metrics.block_constructed.observe(block_took.as_secs_f64()); metrics.report_end_proposing_reason(end_reason); metrics.create_block_proposal_time.observe(propose_took.as_secs_f64()); }); let extrinsics_summary = if extrinsics.is_empty() { "no extrinsics".to_string() } else { format!( "extrinsics ({}): [{}]", extrinsics.len(), extrinsics .iter() .map(|xt| BlakeTwo256::hash_of(xt).to_string()) .collect::>() .join(", ") ) }; info!( "🎁 Prepared block for proposing at {} ({} ms) [hash: {:?}; parent_hash: {}; {extrinsics_summary}", block.header().number(), block_took.as_millis(), ::Hash::from(block.header().hash()), block.header().parent_hash(), ); telemetry!( self.telemetry; CONSENSUS_INFO; "prepared_block_for_proposing"; "number" => ?block.header().number(), "hash" => ?::Hash::from(block.header().hash()), ); } } #[cfg(test)] mod tests { use super::*; use futures::executor::block_on; use parking_lot::Mutex; use sc_client_api::Backend; use sc_transaction_pool::BasicPool; use sc_transaction_pool_api::{ChainEvent, MaintainedTransactionPool, TransactionSource}; use sp_api::Core; use sp_blockchain::HeaderBackend; use sp_consensus::{BlockOrigin, Environment, Proposer}; use sp_runtime::{generic::BlockId, traits::NumberFor, Perbill}; use substrate_test_runtime_client::{ prelude::*, runtime::{Block as TestBlock, Extrinsic, ExtrinsicBuilder, Transfer}, TestClientBuilder, TestClientBuilderExt, }; const SOURCE: TransactionSource = TransactionSource::External; // Note: // Maximum normal extrinsic size for `substrate_test_runtime` is ~65% of max_block (refer to // `substrate_test_runtime::RuntimeBlockWeights` for details). // This extrinsic sizing allows for: // - one huge xts + a lot of tiny dust // - one huge, no medium, // - two medium xts // This is widely exploited in following tests. const HUGE: u32 = 649000000; const MEDIUM: u32 = 250000000; const TINY: u32 = 1000; fn extrinsic(nonce: u64) -> Extrinsic { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(TINY)).nonce(nonce).build() } fn chain_event(header: B::Header) -> ChainEvent where NumberFor: From, { ChainEvent::NewBestBlock { hash: header.hash(), tree_route: None } } #[test] fn should_cease_building_block_when_deadline_is_reached() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let hashof0 = client.info().genesis_hash; block_on(txpool.submit_at(hashof0, SOURCE, vec![extrinsic(0), extrinsic(1)])).unwrap(); block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let cell = Mutex::new((false, time::Instant::now())); let proposer = proposer_factory.init_with_now( &client.expect_header(hashof0).unwrap(), Box::new(move || { let mut value = cell.lock(); if !value.0 { value.0 = true; return value.1 } let old = value.1; let new = old + time::Duration::from_secs(1); *value = (true, new); old }), ); // when let deadline = time::Duration::from_secs(3); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then // block should have some extrinsics although we have some more in the pool. assert_eq!(block.extrinsics().len(), 1); assert_eq!(txpool.ready().count(), 2); } #[test] fn should_not_panic_when_deadline_is_reached() { let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let cell = Mutex::new((false, time::Instant::now())); let proposer = proposer_factory.init_with_now( &client.expect_header(client.info().genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); if !value.0 { value.0 = true; return value.1 } let new = value.1 + time::Duration::from_secs(160); *value = (true, new); new }), ); let deadline = time::Duration::from_secs(1); block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); } #[test] fn proposed_storage_changes_should_match_execute_block_storage_changes() { let (client, backend) = TestClientBuilder::new().build_with_backend(); let client = Arc::new(client); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let genesis_hash = client.info().best_hash; block_on(txpool.submit_at(genesis_hash, SOURCE, vec![extrinsic(0)])).unwrap(); block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let proposer = proposer_factory.init_with_now( &client.header(genesis_hash).unwrap().unwrap(), Box::new(move || time::Instant::now()), ); let deadline = time::Duration::from_secs(9); let proposal = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .unwrap(); assert_eq!(proposal.block.extrinsics().len(), 1); let api = client.runtime_api(); api.execute_block(genesis_hash, proposal.block).unwrap(); let state = backend.state_at(genesis_hash).unwrap(); let storage_changes = api.into_storage_changes(&state, genesis_hash).unwrap(); assert_eq!( proposal.storage_changes.transaction_storage_root, storage_changes.transaction_storage_root, ); } // This test ensures that if one transaction of a user was rejected, because for example // the weight limit was hit, we don't mark the other transactions of the user as invalid because // the nonce is not matching. #[test] fn should_not_remove_invalid_transactions_from_the_same_sender_after_one_was_invalid() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let medium = |nonce| { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(MEDIUM)) .nonce(nonce) .build() }; let huge = |nonce| { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(HUGE)).nonce(nonce).build() }; block_on(txpool.submit_at( client.info().genesis_hash, SOURCE, vec![medium(0), medium(1), huge(2), medium(3), huge(4), medium(5), medium(6)], )) .unwrap(); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let mut propose_block = |client: &TestClient, parent_number, expected_block_extrinsics, expected_pool_transactions| { let hash = client.expect_block_hash_from_id(&BlockId::Number(parent_number)).unwrap(); let proposer = proposer_factory.init_with_now( &client.expect_header(hash).unwrap(), Box::new(move || time::Instant::now()), ); // when let deadline = time::Duration::from_secs(900); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then // block should have some extrinsics although we have some more in the pool. assert_eq!( txpool.ready().count(), expected_pool_transactions, "at block: {}", block.header.number ); assert_eq!( block.extrinsics().len(), expected_block_extrinsics, "at block: {}", block.header.number ); block }; let import_and_maintain = |mut client: Arc, block: TestBlock| { let hash = block.hash(); block_on(client.import(BlockOrigin::Own, block)).unwrap(); block_on(txpool.maintain(chain_event( client.expect_header(hash).expect("there should be header"), ))); }; block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); assert_eq!(txpool.ready().count(), 7); // let's create one block and import it let block = propose_block(&client, 0, 2, 7); import_and_maintain(client.clone(), block); assert_eq!(txpool.ready().count(), 5); // now let's make sure that we can still make some progress let block = propose_block(&client, 1, 1, 5); import_and_maintain(client.clone(), block); assert_eq!(txpool.ready().count(), 4); // again let's make sure that we can still make some progress let block = propose_block(&client, 2, 1, 4); import_and_maintain(client.clone(), block); assert_eq!(txpool.ready().count(), 3); // again let's make sure that we can still make some progress let block = propose_block(&client, 3, 1, 3); import_and_maintain(client.clone(), block); assert_eq!(txpool.ready().count(), 2); // again let's make sure that we can still make some progress let block = propose_block(&client, 4, 2, 2); import_and_maintain(client.clone(), block); assert_eq!(txpool.ready().count(), 0); } #[test] fn should_cease_building_block_when_block_limit_is_reached() { let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let genesis_hash = client.info().genesis_hash; let genesis_header = client.expect_header(genesis_hash).expect("there should be header"); let extrinsics_num = 5; let extrinsics = std::iter::once( Transfer { from: AccountKeyring::Alice.into(), to: AccountKeyring::Bob.into(), amount: 100, nonce: 0, } .into_unchecked_extrinsic(), ) .chain((1..extrinsics_num as u64).map(extrinsic)) .collect::>(); let block_limit = genesis_header.encoded_size() + extrinsics .iter() .take(extrinsics_num - 1) .map(Encode::encoded_size) .sum::() + Vec::::new().encoded_size(); block_on(txpool.submit_at(genesis_hash, SOURCE, extrinsics.clone())).unwrap(); block_on(txpool.maintain(chain_event(genesis_header.clone()))); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let proposer = block_on(proposer_factory.init(&genesis_header)).unwrap(); // Give it enough time let deadline = time::Duration::from_secs(300); let block = block_on(proposer.propose( Default::default(), Default::default(), deadline, Some(block_limit), )) .map(|r| r.block) .unwrap(); // Based on the block limit, one transaction shouldn't be included. assert_eq!(block.extrinsics().len(), extrinsics_num - 1); let proposer = block_on(proposer_factory.init(&genesis_header)).unwrap(); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // Without a block limit we should include all of them assert_eq!(block.extrinsics().len(), extrinsics_num); let mut proposer_factory = ProposerFactory::with_proof_recording( spawner.clone(), client.clone(), txpool.clone(), None, None, ); let proposer = block_on(proposer_factory.init(&genesis_header)).unwrap(); // Exact block_limit, which includes: // 99 (header_size) + 718 (proof@initialize_block) + 246 (one Transfer extrinsic) let block_limit = { let builder = BlockBuilderBuilder::new(&*client) .on_parent_block(genesis_header.hash()) .with_parent_block_number(0) .enable_proof_recording() .build() .unwrap(); builder.estimate_block_size(true) + extrinsics[0].encoded_size() }; let block = block_on(proposer.propose( Default::default(), Default::default(), deadline, Some(block_limit), )) .map(|r| r.block) .unwrap(); // The block limit was increased, but we now include the proof in the estimation of the // block size and thus, only the `Transfer` will fit into the block. It reads more data // than we have reserved in the block limit. assert_eq!(block.extrinsics().len(), 1); } #[test] fn should_keep_adding_transactions_after_exhausts_resources_before_soft_deadline() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let genesis_hash = client.info().genesis_hash; let tiny = |nonce| { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(TINY)).nonce(nonce).build() }; let huge = |who| { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(HUGE)) .signer(AccountKeyring::numeric(who)) .build() }; block_on( txpool.submit_at( genesis_hash, SOURCE, // add 2 * MAX_SKIPPED_TRANSACTIONS that exhaust resources (0..MAX_SKIPPED_TRANSACTIONS * 2) .into_iter() .map(huge) // and some transactions that are okay. .chain((0..MAX_SKIPPED_TRANSACTIONS as u64).into_iter().map(tiny)) .collect(), ), ) .unwrap(); block_on(txpool.maintain(chain_event( client.expect_header(genesis_hash).expect("there should be header"), ))); assert_eq!(txpool.ready().count(), MAX_SKIPPED_TRANSACTIONS * 3); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let cell = Mutex::new(time::Instant::now()); let proposer = proposer_factory.init_with_now( &client.expect_header(genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); let old = *value; *value = old + time::Duration::from_secs(1); old }), ); // when // give it enough time so that deadline is never triggered. let deadline = time::Duration::from_secs(900); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then block should have all non-exhaust resources extrinsics (+ the first one). assert_eq!(block.extrinsics().len(), MAX_SKIPPED_TRANSACTIONS + 1); } #[test] fn should_only_skip_up_to_some_limit_after_soft_deadline() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let genesis_hash = client.info().genesis_hash; let tiny = |who| { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(TINY)) .signer(AccountKeyring::numeric(who)) .nonce(1) .build() }; let huge = |who| { ExtrinsicBuilder::new_fill_block(Perbill::from_parts(HUGE)) .signer(AccountKeyring::numeric(who)) .build() }; block_on( txpool.submit_at( genesis_hash, SOURCE, (0..MAX_SKIPPED_TRANSACTIONS + 2) .into_iter() .map(huge) // and some transactions that are okay. .chain((0..MAX_SKIPPED_TRANSACTIONS + 2).into_iter().map(tiny)) .collect(), ), ) .unwrap(); block_on(txpool.maintain(chain_event( client.expect_header(genesis_hash).expect("there should be header"), ))); assert_eq!(txpool.ready().count(), MAX_SKIPPED_TRANSACTIONS * 2 + 4); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let deadline = time::Duration::from_secs(600); let cell = Arc::new(Mutex::new((0, time::Instant::now()))); let cell2 = cell.clone(); let proposer = proposer_factory.init_with_now( &client.expect_header(genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); let (called, old) = *value; // add time after deadline is calculated internally (hence 1) let increase = if called == 1 { // we start after the soft_deadline should have already been reached. deadline / 2 } else { // but we make sure to never reach the actual deadline time::Duration::from_millis(0) }; *value = (called + 1, old + increase); old }), ); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then the block should have one or two transactions. This maybe random as they are // processed in parallel. The same signer and consecutive nonces for huge and tiny // transactions guarantees that max two transactions will get to the block. assert!( (1..3).contains(&block.extrinsics().len()), "Block shall contain one or two extrinsics." ); assert!( cell2.lock().0 > MAX_SKIPPED_TRANSACTIONS, "Not enough calls to current time, which indicates the test might have ended because of deadline, not soft deadline" ); } }