Files
pezkuwi-subxt/substrate/client/db/benches/state_access.rs
T
Bastian Köcher 73d9ae3284 Introduce trie level cache and remove state cache (#11407)
* trie state cache

* Also cache missing access on read.

* fix comp

* bis

* fix

* use has_lru

* remove local storage cache on size 0.

* No cache.

* local cache only

* trie cache and local cache

* storage cache (with local)

* trie cache no local cache

* Add state access benchmark

* Remove warnings etc

* Add trie cache benchmark

* No extra "clone" required

* Change benchmark to use multiple blocks

* Use patches

* Integrate shitty implementation

* More stuff

* Revert "Merge branch 'master' into trie_state_cache"

This reverts commit 947cd8e6d43fced10e21b76d5b92ffa57b57c318, reversing
changes made to 29ff036463.

* Improve benchmark

* Adapt to latest changes

* Adapt to changes in trie

* Add a test that uses iterator

* Start fixing it

* Remove obsolete file

* Make it compile

* Start rewriting the trie node cache

* More work on the cache

* More docs and code etc

* Make data cache an optional

* Tests

* Remove debug stuff

* Recorder

* Some docs and a simple test for the recorder

* Compile fixes

* Make it compile

* More fixes

* More fixes

* Fix fix fix

* Make sure cache and recorder work together for basic stuff

* Test that data caching and recording works

* Test `TrieDBMut` with caching

* Try something

* Fixes, fixes, fixes

* Forward the recorder

* Make it compile

* Use recorder in more places

* Switch to new `with_optional_recorder` fn

* Refactor and cleanups

* Move `ProvingBackend` tests

* Simplify

* Move over all functionality to the essence

* Fix compilation

* Implement estimate encoded size for StorageProof

* Start using the `cache` everywhere

* Use the cache everywhere

* Fix compilation

* Fix tests

* Adds `TrieBackendBuilder` and enhances the tests

* Ensure that recorder drain checks that values are found as expected

* Switch over to `TrieBackendBuilder`

* Start fixing the problem with child tries and recording

* Fix recording of child tries

* Make it compile

* Overwrite `storage_hash` in `TrieBackend`

* Add `storage_cache` to  the benchmarks

* Fix `no_std` build

* Speed up cache lookup

* Extend the state access benchmark to also hash a runtime

* Fix build

* Fix compilation

* Rewrite value cache

* Add lru cache

* Ensure that the cache lru works

* Value cache should not be optional

* Add support for keeping the shared node cache in its bounds

* Make the cache configurable

* Check that the cache respects the bounds

* Adds a new test

* Fixes

* Docs and some renamings

* More docs

* Start using the new recorder

* Fix more code

* Take `self` argument

* Remove warnings

* Fix benchmark

* Fix accounting

* Rip off the state cache

* Start fixing fallout after removing the state cache

* Make it compile after trie changes

* Fix test

* Add some logging

* Some docs

* Some fixups and clean ups

* Fix benchmark

* Remove unneeded file

* Use git for patching

* Make CI happy

* Update primitives/trie/Cargo.toml

Co-authored-by: Koute <koute@users.noreply.github.com>

* Update primitives/state-machine/src/trie_backend.rs

Co-authored-by: cheme <emericchevalier.pro@gmail.com>

* Introduce new `AsTrieBackend` trait

* Make the LocalTrieCache not clonable

* Make it work in no_std and add docs

* Remove duplicate dependency

* Switch to ahash for better performance

* Speedup value cache merge

* Output errors on underflow

* Ensure the internal LRU map doesn't grow too much

* Use const fn to calculate the value cache element size

* Remove cache configuration

* Fix

* Clear the cache in between for more testing

* Try to come up with a failing test case

* Make the test fail

* Fix the child trie recording

* Make everything compile after the changes to trie

* Adapt to latest trie-db changes

* Fix on stable

* Update primitives/trie/src/cache.rs

Co-authored-by: cheme <emericchevalier.pro@gmail.com>

* Fix wrong merge

* Docs

* Fix warnings

* Cargo.lock

* Bump pin-project

* Fix warnings

* Switch to released crate version

* More fixes

* Make clippy and rustdocs happy

* More clippy

* Print error when using deprecated `--state-cache-size`

* 🤦

* Fixes

* Fix storage_hash linkings

* Update client/rpc/src/dev/mod.rs

Co-authored-by: Arkadiy Paronyan <arkady.paronyan@gmail.com>

* Review feedback

* encode bound

* Rework the shared value cache

Instead of using a `u64` to represent the key we now use an `Arc<[u8]>`. This arc is also stored in
some extra `HashSet`. We store the key are in an extra `HashSet` to de-duplicate the keys accross
different storage roots. When the latest key usage is dropped in the lru, we also remove the key
from the `HashSet`.

* Improve of the cache by merging the old and new solution

* FMT

* Please stop coming back all the time :crying:

* Update primitives/trie/src/cache/shared_cache.rs

Co-authored-by: Arkadiy Paronyan <arkady.paronyan@gmail.com>

* Fixes

* Make clippy happy

* Ensure we don't deadlock

* Only use one lock to simplify the code

* Do not depend on `Hasher`

* Fix tests

* FMT

* Clippy 🤦

Co-authored-by: cheme <emericchevalier.pro@gmail.com>
Co-authored-by: Koute <koute@users.noreply.github.com>
Co-authored-by: Arkadiy Paronyan <arkady.paronyan@gmail.com>
2022-08-18 18:59:22 +00:00

313 lines
8.2 KiB
Rust

// This file is part of Substrate.
// Copyright (C) 2021 Parity Technologies (UK) Ltd.
// SPDX-License-Identifier: GPL-3.0-or-later WITH Classpath-exception-2.0
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU General Public License for more details.
// You should have received a copy of the GNU General Public License
// along with this program. If not, see <https://www.gnu.org/licenses/>.
use criterion::{criterion_group, criterion_main, BatchSize, Criterion};
use rand::{distributions::Uniform, rngs::StdRng, Rng, SeedableRng};
use sc_client_api::{Backend as _, BlockImportOperation, NewBlockState, StateBackend};
use sc_client_db::{Backend, BlocksPruning, DatabaseSettings, DatabaseSource, PruningMode};
use sp_core::H256;
use sp_runtime::{
generic::BlockId,
testing::{Block as RawBlock, ExtrinsicWrapper, Header},
StateVersion, Storage,
};
use tempfile::TempDir;
pub(crate) type Block = RawBlock<ExtrinsicWrapper<u64>>;
fn insert_blocks(db: &Backend<Block>, storage: Vec<(Vec<u8>, Vec<u8>)>) -> H256 {
let mut op = db.begin_operation().unwrap();
let mut header = Header {
number: 0,
parent_hash: Default::default(),
state_root: Default::default(),
digest: Default::default(),
extrinsics_root: Default::default(),
};
header.state_root = op
.set_genesis_state(
Storage {
top: vec![(
sp_core::storage::well_known_keys::CODE.to_vec(),
kitchensink_runtime::wasm_binary_unwrap().to_vec(),
)]
.into_iter()
.collect(),
children_default: Default::default(),
},
true,
StateVersion::V1,
)
.unwrap();
op.set_block_data(header.clone(), Some(vec![]), None, None, NewBlockState::Best)
.unwrap();
db.commit_operation(op).unwrap();
let mut number = 1;
let mut parent_hash = header.hash();
for i in 0..10 {
let mut op = db.begin_operation().unwrap();
db.begin_state_operation(&mut op, BlockId::Hash(parent_hash)).unwrap();
let mut header = Header {
number,
parent_hash,
state_root: Default::default(),
digest: Default::default(),
extrinsics_root: Default::default(),
};
let changes = storage
.iter()
.skip(i * 100_000)
.take(100_000)
.map(|(k, v)| (k.clone(), Some(v.clone())))
.collect::<Vec<_>>();
let (state_root, tx) = db.state_at(BlockId::Number(number - 1)).unwrap().storage_root(
changes.iter().map(|(k, v)| (k.as_slice(), v.as_deref())),
StateVersion::V1,
);
header.state_root = state_root;
op.update_db_storage(tx).unwrap();
op.update_storage(changes.clone(), Default::default()).unwrap();
op.set_block_data(header.clone(), Some(vec![]), None, None, NewBlockState::Best)
.unwrap();
db.commit_operation(op).unwrap();
number += 1;
parent_hash = header.hash();
}
parent_hash
}
enum BenchmarkConfig {
NoCache,
TrieNodeCache,
}
fn create_backend(config: BenchmarkConfig, temp_dir: &TempDir) -> Backend<Block> {
let path = temp_dir.path().to_owned();
let trie_cache_maximum_size = match config {
BenchmarkConfig::NoCache => None,
BenchmarkConfig::TrieNodeCache => Some(2 * 1024 * 1024 * 1024),
};
let settings = DatabaseSettings {
trie_cache_maximum_size,
state_pruning: Some(PruningMode::ArchiveAll),
source: DatabaseSource::ParityDb { path },
blocks_pruning: BlocksPruning::All,
};
Backend::new(settings, 100).expect("Creates backend")
}
/// Generate the storage that will be used for the benchmark
///
/// Returns the `Vec<key>` and the `Vec<(key, value)>`
fn generate_storage() -> (Vec<Vec<u8>>, Vec<(Vec<u8>, Vec<u8>)>) {
let mut rng = StdRng::seed_from_u64(353893213);
let mut storage = Vec::new();
let mut keys = Vec::new();
for _ in 0..1_000_000 {
let key_len: usize = rng.gen_range(32..128);
let key = (&mut rng)
.sample_iter(Uniform::new_inclusive(0, 255))
.take(key_len)
.collect::<Vec<u8>>();
let value_len: usize = rng.gen_range(20..60);
let value = (&mut rng)
.sample_iter(Uniform::new_inclusive(0, 255))
.take(value_len)
.collect::<Vec<u8>>();
keys.push(key.clone());
storage.push((key, value));
}
(keys, storage)
}
fn state_access_benchmarks(c: &mut Criterion) {
sp_tracing::try_init_simple();
let (keys, storage) = generate_storage();
let path = TempDir::new().expect("Creates temporary directory");
let block_hash = {
let backend = create_backend(BenchmarkConfig::NoCache, &path);
insert_blocks(&backend, storage.clone())
};
let mut group = c.benchmark_group("Reading entire state");
group.sample_size(20);
let mut bench_multiple_values = |config, desc, multiplier| {
let backend = create_backend(config, &path);
group.bench_function(desc, |b| {
b.iter_batched(
|| backend.state_at(BlockId::Hash(block_hash)).expect("Creates state"),
|state| {
for key in keys.iter().cycle().take(keys.len() * multiplier) {
let _ = state.storage(&key).expect("Doesn't fail").unwrap();
}
},
BatchSize::SmallInput,
)
});
};
bench_multiple_values(
BenchmarkConfig::TrieNodeCache,
"with trie node cache and reading each key once",
1,
);
bench_multiple_values(BenchmarkConfig::NoCache, "no cache and reading each key once", 1);
bench_multiple_values(
BenchmarkConfig::TrieNodeCache,
"with trie node cache and reading 4 times each key in a row",
4,
);
bench_multiple_values(
BenchmarkConfig::NoCache,
"no cache and reading 4 times each key in a row",
4,
);
group.finish();
let mut group = c.benchmark_group("Reading a single value");
let mut bench_single_value = |config, desc, multiplier| {
let backend = create_backend(config, &path);
group.bench_function(desc, |b| {
b.iter_batched(
|| backend.state_at(BlockId::Hash(block_hash)).expect("Creates state"),
|state| {
for key in keys.iter().take(1).cycle().take(multiplier) {
let _ = state.storage(&key).expect("Doesn't fail").unwrap();
}
},
BatchSize::SmallInput,
)
});
};
bench_single_value(
BenchmarkConfig::TrieNodeCache,
"with trie node cache and reading the key once",
1,
);
bench_single_value(BenchmarkConfig::NoCache, "no cache and reading the key once", 1);
bench_single_value(
BenchmarkConfig::TrieNodeCache,
"with trie node cache and reading 4 times each key in a row",
4,
);
bench_single_value(
BenchmarkConfig::NoCache,
"no cache and reading 4 times each key in a row",
4,
);
group.finish();
let mut group = c.benchmark_group("Hashing a value");
let mut bench_single_value = |config, desc, multiplier| {
let backend = create_backend(config, &path);
group.bench_function(desc, |b| {
b.iter_batched(
|| backend.state_at(BlockId::Hash(block_hash)).expect("Creates state"),
|state| {
for key in keys.iter().take(1).cycle().take(multiplier) {
let _ = state.storage_hash(&key).expect("Doesn't fail").unwrap();
}
},
BatchSize::SmallInput,
)
});
};
bench_single_value(
BenchmarkConfig::TrieNodeCache,
"with trie node cache and hashing the key once",
1,
);
bench_single_value(BenchmarkConfig::NoCache, "no cache and hashing the key once", 1);
bench_single_value(
BenchmarkConfig::TrieNodeCache,
"with trie node cache and hashing 4 times each key in a row",
4,
);
bench_single_value(
BenchmarkConfig::NoCache,
"no cache and hashing 4 times each key in a row",
4,
);
group.finish();
let mut group = c.benchmark_group("Hashing `:code`");
let mut bench_single_value = |config, desc| {
let backend = create_backend(config, &path);
group.bench_function(desc, |b| {
b.iter_batched(
|| backend.state_at(BlockId::Hash(block_hash)).expect("Creates state"),
|state| {
let _ = state
.storage_hash(sp_core::storage::well_known_keys::CODE)
.expect("Doesn't fail")
.unwrap();
},
BatchSize::SmallInput,
)
});
};
bench_single_value(BenchmarkConfig::TrieNodeCache, "with trie node cache");
bench_single_value(BenchmarkConfig::NoCache, "no cache");
group.finish();
}
criterion_group!(benches, state_access_benchmarks);
criterion_main!(benches);