// This file is part of Substrate. // Copyright (C) Parity Technologies (UK) Ltd. // SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0 // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // You should have received a copy of the GNU General Public License // along with this program. If not, see . //! A consensus proposer for "basic" chains which use the primitive inherent-data. // FIXME #1021 move this into sp-consensus use codec::Encode; use futures::{ channel::oneshot, future, future::{Future, FutureExt}, select, }; use log::{debug, error, info, trace, warn}; use sc_block_builder::{BlockBuilderApi, BlockBuilderProvider}; use sc_client_api::backend; use sc_telemetry::{telemetry, TelemetryHandle, CONSENSUS_INFO}; use sc_transaction_pool_api::{InPoolTransaction, TransactionPool}; use sp_api::{ApiExt, ProvideRuntimeApi}; use sp_blockchain::{ApplyExtrinsicFailed::Validity, Error::ApplyExtrinsicFailed, HeaderBackend}; use sp_consensus::{DisableProofRecording, EnableProofRecording, ProofRecording, Proposal}; use sp_core::traits::SpawnNamed; use sp_inherents::InherentData; use sp_runtime::{ traits::{BlakeTwo256, Block as BlockT, Hash as HashT, Header as HeaderT}, Digest, Percent, SaturatedConversion, }; use std::{marker::PhantomData, pin::Pin, sync::Arc, time}; use prometheus_endpoint::Registry as PrometheusRegistry; use sc_proposer_metrics::{EndProposingReason, MetricsLink as PrometheusMetrics}; /// Default block size limit in bytes used by [`Proposer`]. /// /// Can be overwritten by [`ProposerFactory::set_default_block_size_limit`]. /// /// Be aware that there is also an upper packet size on what the networking code /// will accept. If the block doesn't fit in such a package, it can not be /// transferred to other nodes. pub const DEFAULT_BLOCK_SIZE_LIMIT: usize = 4 * 1024 * 1024 + 512; const DEFAULT_SOFT_DEADLINE_PERCENT: Percent = Percent::from_percent(50); /// [`Proposer`] factory. pub struct ProposerFactory { spawn_handle: Box, /// The client instance. client: Arc, /// The transaction pool. transaction_pool: Arc, /// Prometheus Link, metrics: PrometheusMetrics, /// The default block size limit. /// /// If no `block_size_limit` is passed to [`sp_consensus::Proposer::propose`], this block size /// limit will be used. default_block_size_limit: usize, /// Soft deadline percentage of hard deadline. /// /// The value is used to compute soft deadline during block production. /// The soft deadline indicates where we should stop attempting to add transactions /// to the block, which exhaust resources. After soft deadline is reached, /// we switch to a fixed-amount mode, in which after we see `MAX_SKIPPED_TRANSACTIONS` /// transactions which exhaust resrouces, we will conclude that the block is full. soft_deadline_percent: Percent, telemetry: Option, /// When estimating the block size, should the proof be included? include_proof_in_block_size_estimation: bool, /// phantom member to pin the `Backend`/`ProofRecording` type. _phantom: PhantomData<(B, PR)>, } impl ProposerFactory { /// Create a new proposer factory. /// /// Proof recording will be disabled when using proposers built by this instance to build /// blocks. pub fn new( spawn_handle: impl SpawnNamed + 'static, client: Arc, transaction_pool: Arc, prometheus: Option<&PrometheusRegistry>, telemetry: Option, ) -> Self { ProposerFactory { spawn_handle: Box::new(spawn_handle), transaction_pool, metrics: PrometheusMetrics::new(prometheus), default_block_size_limit: DEFAULT_BLOCK_SIZE_LIMIT, soft_deadline_percent: DEFAULT_SOFT_DEADLINE_PERCENT, telemetry, client, include_proof_in_block_size_estimation: false, _phantom: PhantomData, } } } impl ProposerFactory { /// Create a new proposer factory with proof recording enabled. /// /// Each proposer created by this instance will record a proof while building a block. /// /// This will also include the proof into the estimation of the block size. This can be disabled /// by calling [`ProposerFactory::disable_proof_in_block_size_estimation`]. pub fn with_proof_recording( spawn_handle: impl SpawnNamed + 'static, client: Arc, transaction_pool: Arc, prometheus: Option<&PrometheusRegistry>, telemetry: Option, ) -> Self { ProposerFactory { client, spawn_handle: Box::new(spawn_handle), transaction_pool, metrics: PrometheusMetrics::new(prometheus), default_block_size_limit: DEFAULT_BLOCK_SIZE_LIMIT, soft_deadline_percent: DEFAULT_SOFT_DEADLINE_PERCENT, telemetry, include_proof_in_block_size_estimation: true, _phantom: PhantomData, } } /// Disable the proof inclusion when estimating the block size. pub fn disable_proof_in_block_size_estimation(&mut self) { self.include_proof_in_block_size_estimation = false; } } impl ProposerFactory { /// Set the default block size limit in bytes. /// /// The default value for the block size limit is: /// [`DEFAULT_BLOCK_SIZE_LIMIT`]. /// /// If there is no block size limit passed to [`sp_consensus::Proposer::propose`], this value /// will be used. pub fn set_default_block_size_limit(&mut self, limit: usize) { self.default_block_size_limit = limit; } /// Set soft deadline percentage. /// /// The value is used to compute soft deadline during block production. /// The soft deadline indicates where we should stop attempting to add transactions /// to the block, which exhaust resources. After soft deadline is reached, /// we switch to a fixed-amount mode, in which after we see `MAX_SKIPPED_TRANSACTIONS` /// transactions which exhaust resrouces, we will conclude that the block is full. /// /// Setting the value too low will significantly limit the amount of transactions /// we try in case they exhaust resources. Setting the value too high can /// potentially open a DoS vector, where many "exhaust resources" transactions /// are being tried with no success, hence block producer ends up creating an empty block. pub fn set_soft_deadline(&mut self, percent: Percent) { self.soft_deadline_percent = percent; } } impl ProposerFactory where A: TransactionPool + 'static, B: backend::Backend + Send + Sync + 'static, Block: BlockT, C: BlockBuilderProvider + HeaderBackend + ProvideRuntimeApi + Send + Sync + 'static, C::Api: ApiExt> + BlockBuilderApi, { fn init_with_now( &mut self, parent_header: &::Header, now: Box time::Instant + Send + Sync>, ) -> Proposer { let parent_hash = parent_header.hash(); info!("🙌 Starting consensus session on top of parent {:?}", parent_hash); let proposer = Proposer::<_, _, _, _, PR> { spawn_handle: self.spawn_handle.clone(), client: self.client.clone(), parent_hash, parent_number: *parent_header.number(), transaction_pool: self.transaction_pool.clone(), now, metrics: self.metrics.clone(), default_block_size_limit: self.default_block_size_limit, soft_deadline_percent: self.soft_deadline_percent, telemetry: self.telemetry.clone(), _phantom: PhantomData, include_proof_in_block_size_estimation: self.include_proof_in_block_size_estimation, }; proposer } } impl sp_consensus::Environment for ProposerFactory where A: TransactionPool + 'static, B: backend::Backend + Send + Sync + 'static, Block: BlockT, C: BlockBuilderProvider + HeaderBackend + ProvideRuntimeApi + Send + Sync + 'static, C::Api: ApiExt> + BlockBuilderApi, PR: ProofRecording, { type CreateProposer = future::Ready>; type Proposer = Proposer; type Error = sp_blockchain::Error; fn init(&mut self, parent_header: &::Header) -> Self::CreateProposer { future::ready(Ok(self.init_with_now(parent_header, Box::new(time::Instant::now)))) } } /// The proposer logic. pub struct Proposer { spawn_handle: Box, client: Arc, parent_hash: Block::Hash, parent_number: <::Header as HeaderT>::Number, transaction_pool: Arc, now: Box time::Instant + Send + Sync>, metrics: PrometheusMetrics, default_block_size_limit: usize, include_proof_in_block_size_estimation: bool, soft_deadline_percent: Percent, telemetry: Option, _phantom: PhantomData<(B, PR)>, } impl sp_consensus::Proposer for Proposer where A: TransactionPool + 'static, B: backend::Backend + Send + Sync + 'static, Block: BlockT, C: BlockBuilderProvider + HeaderBackend + ProvideRuntimeApi + Send + Sync + 'static, C::Api: ApiExt> + BlockBuilderApi, PR: ProofRecording, { type Transaction = backend::TransactionFor; type Proposal = Pin< Box< dyn Future, Self::Error>> + Send, >, >; type Error = sp_blockchain::Error; type ProofRecording = PR; type Proof = PR::Proof; fn propose( self, inherent_data: InherentData, inherent_digests: Digest, max_duration: time::Duration, block_size_limit: Option, ) -> Self::Proposal { let (tx, rx) = oneshot::channel(); let spawn_handle = self.spawn_handle.clone(); spawn_handle.spawn_blocking( "basic-authorship-proposer", None, Box::pin(async move { // leave some time for evaluation and block finalization (33%) let deadline = (self.now)() + max_duration - max_duration / 3; let res = self .propose_with(inherent_data, inherent_digests, deadline, block_size_limit) .await; if tx.send(res).is_err() { trace!("Could not send block production result to proposer!"); } }), ); async move { rx.await? }.boxed() } } /// If the block is full we will attempt to push at most /// this number of transactions before quitting for real. /// It allows us to increase block utilization. const MAX_SKIPPED_TRANSACTIONS: usize = 8; impl Proposer where A: TransactionPool, B: backend::Backend + Send + Sync + 'static, Block: BlockT, C: BlockBuilderProvider + HeaderBackend + ProvideRuntimeApi + Send + Sync + 'static, C::Api: ApiExt> + BlockBuilderApi, PR: ProofRecording, { async fn propose_with( self, inherent_data: InherentData, inherent_digests: Digest, deadline: time::Instant, block_size_limit: Option, ) -> Result, PR::Proof>, sp_blockchain::Error> { let propose_with_start = time::Instant::now(); let mut block_builder = self.client.new_block_at(self.parent_hash, inherent_digests, PR::ENABLED)?; let create_inherents_start = time::Instant::now(); let inherents = block_builder.create_inherents(inherent_data)?; let create_inherents_end = time::Instant::now(); self.metrics.report(|metrics| { metrics.create_inherents_time.observe( create_inherents_end .saturating_duration_since(create_inherents_start) .as_secs_f64(), ); }); for inherent in inherents { match block_builder.push(inherent) { Err(ApplyExtrinsicFailed(Validity(e))) if e.exhausted_resources() => { warn!("⚠️ Dropping non-mandatory inherent from overweight block.") }, Err(ApplyExtrinsicFailed(Validity(e))) if e.was_mandatory() => { error!( "❌️ Mandatory inherent extrinsic returned error. Block cannot be produced." ); return Err(ApplyExtrinsicFailed(Validity(e))) }, Err(e) => { warn!("❗️ Inherent extrinsic returned unexpected error: {}. Dropping.", e); }, Ok(_) => {}, } } // proceed with transactions // We calculate soft deadline used only in case we start skipping transactions. let now = (self.now)(); let left = deadline.saturating_duration_since(now); let left_micros: u64 = left.as_micros().saturated_into(); let soft_deadline = now + time::Duration::from_micros(self.soft_deadline_percent.mul_floor(left_micros)); let block_timer = time::Instant::now(); let mut skipped = 0; let mut unqueue_invalid = Vec::new(); let mut t1 = self.transaction_pool.ready_at(self.parent_number).fuse(); let mut t2 = futures_timer::Delay::new(deadline.saturating_duration_since((self.now)()) / 8).fuse(); let mut pending_iterator = select! { res = t1 => res, _ = t2 => { log::warn!( "Timeout fired waiting for transaction pool at block #{}. \ Proceeding with production.", self.parent_number, ); self.transaction_pool.ready() }, }; let block_size_limit = block_size_limit.unwrap_or(self.default_block_size_limit); debug!("Attempting to push transactions from the pool."); debug!("Pool status: {:?}", self.transaction_pool.status()); let mut transaction_pushed = false; let end_reason = loop { let pending_tx = if let Some(pending_tx) = pending_iterator.next() { pending_tx } else { break EndProposingReason::NoMoreTransactions }; let now = (self.now)(); if now > deadline { debug!( "Consensus deadline reached when pushing block transactions, \ proceeding with proposing." ); break EndProposingReason::HitDeadline } let pending_tx_data = pending_tx.data().clone(); let pending_tx_hash = pending_tx.hash().clone(); let block_size = block_builder.estimate_block_size(self.include_proof_in_block_size_estimation); if block_size + pending_tx_data.encoded_size() > block_size_limit { pending_iterator.report_invalid(&pending_tx); if skipped < MAX_SKIPPED_TRANSACTIONS { skipped += 1; debug!( "Transaction would overflow the block size limit, \ but will try {} more transactions before quitting.", MAX_SKIPPED_TRANSACTIONS - skipped, ); continue } else if now < soft_deadline { debug!( "Transaction would overflow the block size limit, \ but we still have time before the soft deadline, so \ we will try a bit more." ); continue } else { debug!("Reached block size limit, proceeding with proposing."); break EndProposingReason::HitBlockSizeLimit } } trace!("[{:?}] Pushing to the block.", pending_tx_hash); match sc_block_builder::BlockBuilder::push(&mut block_builder, pending_tx_data) { Ok(()) => { transaction_pushed = true; debug!("[{:?}] Pushed to the block.", pending_tx_hash); }, Err(ApplyExtrinsicFailed(Validity(e))) if e.exhausted_resources() => { pending_iterator.report_invalid(&pending_tx); if skipped < MAX_SKIPPED_TRANSACTIONS { skipped += 1; debug!( "Block seems full, but will try {} more transactions before quitting.", MAX_SKIPPED_TRANSACTIONS - skipped, ); } else if (self.now)() < soft_deadline { debug!( "Block seems full, but we still have time before the soft deadline, \ so we will try a bit more before quitting." ); } else { debug!("Reached block weight limit, proceeding with proposing."); break EndProposingReason::HitBlockWeightLimit } }, Err(e) if skipped > 0 => { pending_iterator.report_invalid(&pending_tx); trace!( "[{:?}] Ignoring invalid transaction when skipping: {}", pending_tx_hash, e ); }, Err(e) => { pending_iterator.report_invalid(&pending_tx); debug!("[{:?}] Invalid transaction: {}", pending_tx_hash, e); unqueue_invalid.push(pending_tx_hash); }, } }; if matches!(end_reason, EndProposingReason::HitBlockSizeLimit) && !transaction_pushed { warn!( "Hit block size limit of `{}` without including any transaction!", block_size_limit, ); } self.transaction_pool.remove_invalid(&unqueue_invalid); let (block, storage_changes, proof) = block_builder.build()?.into_inner(); self.metrics.report(|metrics| { metrics.number_of_transactions.set(block.extrinsics().len() as u64); metrics.block_constructed.observe(block_timer.elapsed().as_secs_f64()); metrics.report_end_proposing_reason(end_reason); }); info!( "🎁 Prepared block for proposing at {} ({} ms) [hash: {:?}; parent_hash: {}; extrinsics ({}): [{}]]", block.header().number(), block_timer.elapsed().as_millis(), ::Hash::from(block.header().hash()), block.header().parent_hash(), block.extrinsics().len(), block.extrinsics() .iter() .map(|xt| BlakeTwo256::hash_of(xt).to_string()) .collect::>() .join(", ") ); telemetry!( self.telemetry; CONSENSUS_INFO; "prepared_block_for_proposing"; "number" => ?block.header().number(), "hash" => ?::Hash::from(block.header().hash()), ); let proof = PR::into_proof(proof).map_err(|e| sp_blockchain::Error::Application(Box::new(e)))?; let propose_with_end = time::Instant::now(); self.metrics.report(|metrics| { metrics.create_block_proposal_time.observe( propose_with_end.saturating_duration_since(propose_with_start).as_secs_f64(), ); }); Ok(Proposal { block, proof, storage_changes }) } } #[cfg(test)] mod tests { use super::*; use futures::executor::block_on; use parking_lot::Mutex; use sc_client_api::Backend; use sc_transaction_pool::BasicPool; use sc_transaction_pool_api::{ChainEvent, MaintainedTransactionPool, TransactionSource}; use sp_api::Core; use sp_blockchain::HeaderBackend; use sp_consensus::{BlockOrigin, Environment, Proposer}; use sp_core::Pair; use sp_runtime::{generic::BlockId, traits::NumberFor}; use substrate_test_runtime_client::{ prelude::*, runtime::{Extrinsic, Transfer}, TestClientBuilder, TestClientBuilderExt, }; const SOURCE: TransactionSource = TransactionSource::External; fn extrinsic(nonce: u64) -> Extrinsic { Transfer { amount: Default::default(), nonce, from: AccountKeyring::Alice.into(), to: AccountKeyring::Bob.into(), } .into_signed_tx() } fn exhausts_resources_extrinsic_from(who: usize) -> Extrinsic { let pair = AccountKeyring::numeric(who); let transfer = Transfer { // increase the amount to bump priority amount: 1, nonce: 0, from: pair.public(), to: AccountKeyring::Bob.into(), }; let signature = pair.sign(&transfer.encode()).into(); Extrinsic::Transfer { transfer, signature, exhaust_resources_when_not_first: true } } fn chain_event(header: B::Header) -> ChainEvent where NumberFor: From, { ChainEvent::NewBestBlock { hash: header.hash(), tree_route: None } } #[test] fn should_cease_building_block_when_deadline_is_reached() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); block_on(txpool.submit_at(&BlockId::number(0), SOURCE, vec![extrinsic(0), extrinsic(1)])) .unwrap(); block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let cell = Mutex::new((false, time::Instant::now())); let proposer = proposer_factory.init_with_now( &client.expect_header(client.info().genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); if !value.0 { value.0 = true; return value.1 } let old = value.1; let new = old + time::Duration::from_secs(1); *value = (true, new); old }), ); // when let deadline = time::Duration::from_secs(3); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then // block should have some extrinsics although we have some more in the pool. assert_eq!(block.extrinsics().len(), 1); assert_eq!(txpool.ready().count(), 2); } #[test] fn should_not_panic_when_deadline_is_reached() { let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let cell = Mutex::new((false, time::Instant::now())); let proposer = proposer_factory.init_with_now( &client.expect_header(client.info().genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); if !value.0 { value.0 = true; return value.1 } let new = value.1 + time::Duration::from_secs(160); *value = (true, new); new }), ); let deadline = time::Duration::from_secs(1); block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); } #[test] fn proposed_storage_changes_should_match_execute_block_storage_changes() { let (client, backend) = TestClientBuilder::new().build_with_backend(); let client = Arc::new(client); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let genesis_hash = client.info().best_hash; block_on(txpool.submit_at(&BlockId::number(0), SOURCE, vec![extrinsic(0)])).unwrap(); block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let proposer = proposer_factory.init_with_now( &client.header(genesis_hash).unwrap().unwrap(), Box::new(move || time::Instant::now()), ); let deadline = time::Duration::from_secs(9); let proposal = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .unwrap(); assert_eq!(proposal.block.extrinsics().len(), 1); let api = client.runtime_api(); api.execute_block(genesis_hash, proposal.block).unwrap(); let state = backend.state_at(genesis_hash).unwrap(); let storage_changes = api.into_storage_changes(&state, genesis_hash).unwrap(); assert_eq!( proposal.storage_changes.transaction_storage_root, storage_changes.transaction_storage_root, ); } #[test] fn should_not_remove_invalid_transactions_when_skipping() { // given let mut client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); block_on(txpool.submit_at( &BlockId::number(0), SOURCE, vec![ extrinsic(0), extrinsic(1), Transfer { amount: Default::default(), nonce: 2, from: AccountKeyring::Alice.into(), to: AccountKeyring::Bob.into(), }.into_resources_exhausting_tx(), extrinsic(3), Transfer { amount: Default::default(), nonce: 4, from: AccountKeyring::Alice.into(), to: AccountKeyring::Bob.into(), }.into_resources_exhausting_tx(), extrinsic(5), extrinsic(6), ], )) .unwrap(); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let mut propose_block = |client: &TestClient, number, expected_block_extrinsics, expected_pool_transactions| { let hash = client.expect_block_hash_from_id(&BlockId::Number(number)).unwrap(); let proposer = proposer_factory.init_with_now( &client.expect_header(hash).unwrap(), Box::new(move || time::Instant::now()), ); // when let deadline = time::Duration::from_secs(900); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then // block should have some extrinsics although we have some more in the pool. assert_eq!(txpool.ready().count(), expected_pool_transactions); assert_eq!(block.extrinsics().len(), expected_block_extrinsics); block }; block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); assert_eq!(txpool.ready().count(), 7); // let's create one block and import it let block = propose_block(&client, 0, 2, 7); let hashof1 = block.hash(); block_on(client.import(BlockOrigin::Own, block)).unwrap(); block_on( txpool.maintain(chain_event( client.expect_header(hashof1).expect("there should be header"), )), ); assert_eq!(txpool.ready().count(), 5); // now let's make sure that we can still make some progress let block = propose_block(&client, 1, 2, 5); block_on(client.import(BlockOrigin::Own, block)).unwrap(); } #[test] fn should_cease_building_block_when_block_limit_is_reached() { let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); let genesis_header = client .expect_header(client.info().genesis_hash) .expect("there should be header"); let extrinsics_num = 5; let extrinsics = std::iter::once( Transfer { from: AccountKeyring::Alice.into(), to: AccountKeyring::Bob.into(), amount: 100, nonce: 0, } .into_signed_tx(), ) .chain((0..extrinsics_num - 1).map(|v| Extrinsic::IncludeData(vec![v as u8; 10]))) .collect::>(); let block_limit = genesis_header.encoded_size() + extrinsics .iter() .take(extrinsics_num - 1) .map(Encode::encoded_size) .sum::() + Vec::::new().encoded_size(); block_on(txpool.submit_at(&BlockId::number(0), SOURCE, extrinsics)).unwrap(); block_on(txpool.maintain(chain_event(genesis_header.clone()))); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let proposer = block_on(proposer_factory.init(&genesis_header)).unwrap(); // Give it enough time let deadline = time::Duration::from_secs(300); let block = block_on(proposer.propose( Default::default(), Default::default(), deadline, Some(block_limit), )) .map(|r| r.block) .unwrap(); // Based on the block limit, one transaction shouldn't be included. assert_eq!(block.extrinsics().len(), extrinsics_num - 1); let proposer = block_on(proposer_factory.init(&genesis_header)).unwrap(); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // Without a block limit we should include all of them assert_eq!(block.extrinsics().len(), extrinsics_num); let mut proposer_factory = ProposerFactory::with_proof_recording( spawner.clone(), client.clone(), txpool.clone(), None, None, ); let proposer = block_on(proposer_factory.init(&genesis_header)).unwrap(); // Give it enough time let block = block_on(proposer.propose( Default::default(), Default::default(), deadline, Some(block_limit), )) .map(|r| r.block) .unwrap(); // The block limit didn't changed, but we now include the proof in the estimation of the // block size and thus, only the `Transfer` will fit into the block. It reads more data // than we have reserved in the block limit. assert_eq!(block.extrinsics().len(), 1); } #[test] fn should_keep_adding_transactions_after_exhausts_resources_before_soft_deadline() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); block_on( txpool.submit_at( &BlockId::number(0), SOURCE, // add 2 * MAX_SKIPPED_TRANSACTIONS that exhaust resources (0..MAX_SKIPPED_TRANSACTIONS * 2) .into_iter() .map(|i| exhausts_resources_extrinsic_from(i)) // and some transactions that are okay. .chain((0..MAX_SKIPPED_TRANSACTIONS).into_iter().map(|i| extrinsic(i as _))) .collect(), ), ) .unwrap(); block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); assert_eq!(txpool.ready().count(), MAX_SKIPPED_TRANSACTIONS * 3); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let cell = Mutex::new(time::Instant::now()); let proposer = proposer_factory.init_with_now( &client.expect_header(client.info().genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); let old = *value; *value = old + time::Duration::from_secs(1); old }), ); // when // give it enough time so that deadline is never triggered. let deadline = time::Duration::from_secs(900); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then block should have all non-exhaust resources extrinsics (+ the first one). assert_eq!(block.extrinsics().len(), MAX_SKIPPED_TRANSACTIONS + 1); } #[test] fn should_only_skip_up_to_some_limit_after_soft_deadline() { // given let client = Arc::new(substrate_test_runtime_client::new()); let spawner = sp_core::testing::TaskExecutor::new(); let txpool = BasicPool::new_full( Default::default(), true.into(), None, spawner.clone(), client.clone(), ); block_on( txpool.submit_at( &BlockId::number(0), SOURCE, (0..MAX_SKIPPED_TRANSACTIONS + 2) .into_iter() .map(|i| exhausts_resources_extrinsic_from(i)) // and some transactions that are okay. .chain((0..MAX_SKIPPED_TRANSACTIONS).into_iter().map(|i| extrinsic(i as _))) .collect(), ), ) .unwrap(); block_on( txpool.maintain(chain_event( client .expect_header(client.info().genesis_hash) .expect("there should be header"), )), ); assert_eq!(txpool.ready().count(), MAX_SKIPPED_TRANSACTIONS * 2 + 2); let mut proposer_factory = ProposerFactory::new(spawner.clone(), client.clone(), txpool.clone(), None, None); let deadline = time::Duration::from_secs(600); let cell = Arc::new(Mutex::new((0, time::Instant::now()))); let cell2 = cell.clone(); let proposer = proposer_factory.init_with_now( &client.expect_header(client.info().genesis_hash).unwrap(), Box::new(move || { let mut value = cell.lock(); let (called, old) = *value; // add time after deadline is calculated internally (hence 1) let increase = if called == 1 { // we start after the soft_deadline should have already been reached. deadline / 2 } else { // but we make sure to never reach the actual deadline time::Duration::from_millis(0) }; *value = (called + 1, old + increase); old }), ); let block = block_on(proposer.propose(Default::default(), Default::default(), deadline, None)) .map(|r| r.block) .unwrap(); // then the block should have no transactions despite some in the pool assert_eq!(block.extrinsics().len(), 1); assert!( cell2.lock().0 > MAX_SKIPPED_TRANSACTIONS, "Not enough calls to current time, which indicates the test might have ended because of deadline, not soft deadline" ); } }