Core db and aristo updates for destructor and tx logic (#1894)
* Disable `TransactionID` related functions from `state_db.nim`
why:
Functions `getCommittedStorage()` and `updateOriginalRoot()` from
the `state_db` module are nowhere used. The emulation of a legacy
`TransactionID` type functionality is administratively expensive to
provide by `Aristo` (the legacy DB version is only partially
implemented, anyway).
As there is no other place where `TransactionID`s are used, they will
not be provided by the `Aristo` variant of the `CoreDb`. For the
legacy DB API, nothing will change.
* Fix copyright headers in source code
* Get rid of compiler warning
* Update Aristo code, remove unused `merge()` variant, export `hashify()`
why:
Adapt to upcoming `CoreDb` wrapper
* Remove synced tx feature from `Aristo`
why:
+ This feature allowed to synchronise transaction methods like begin,
commit, and rollback for a group of descriptors.
+ The feature is over engineered and not needed for `CoreDb`, neither
is it complete (some convergence features missing.)
* Add debugging helpers to `Kvt`
also:
Update database iterator, add count variable yield argument similar
to `Aristo`.
* Provide optional destructors for `CoreDb` API
why;
For the upcoming Aristo wrapper, this allows to control when certain
smart destruction and update can take place. The auto destructor works
fine in general when the storage/cache strategy is known and acceptable
when creating descriptors.
* Add update option for `CoreDb` API function `hash()`
why;
The hash function is typically used to get the state root of the MPT.
Due to lazy hashing, this might be not available on the `Aristo` DB.
So the `update` function asks for re-hashing the gurrent state changes
if needed.
* Update API tracking log mode: `info` => `debug
* Use shared `Kvt` descriptor in new Ledger API
why:
No need to create a new descriptor all the time
2023-11-16 19:35:03 +00:00
|
|
|
# Nimbus
|
2024-02-04 14:28:20 +00:00
|
|
|
# Copyright (c) 2022-2024 Status Research & Development GmbH
|
2023-01-23 16:09:12 +00:00
|
|
|
# Licensed under either of
|
|
|
|
# * Apache License, version 2.0, ([LICENSE-APACHE](LICENSE-APACHE) or
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0)
|
|
|
|
# * MIT license ([LICENSE-MIT](LICENSE-MIT) or
|
|
|
|
# http://opensource.org/licenses/MIT)
|
|
|
|
# at your option. This file may not be copied, modified, or
|
|
|
|
# distributed except according to those terms.
|
|
|
|
|
|
|
|
## Snap sync components tester and TDD environment
|
2023-02-02 13:27:09 +00:00
|
|
|
##
|
|
|
|
## This module provides test bodies for storing chain chain data directly
|
|
|
|
## rather than derive them by executing the EVM. Here, only accounts are
|
|
|
|
## considered.
|
|
|
|
##
|
|
|
|
## The `snap/1` protocol allows to fetch data for a certain account range. The
|
|
|
|
## following boundary conditions apply to the received data:
|
|
|
|
##
|
|
|
|
## * `State root`: All data are relaive to the same state root.
|
|
|
|
##
|
|
|
|
## * `Accounts`: There is an accounts interval sorted in strictly increasing
|
|
|
|
## order. The accounts are required consecutive, i.e. without holes in
|
|
|
|
## between although this cannot be verified immediately.
|
|
|
|
##
|
|
|
|
## * `Lower bound`: There is a start value which might be lower than the first
|
|
|
|
## account hash. There must be no other account between this start value and
|
|
|
|
## the first account (not verifyable yet.) For all practicat purposes, this
|
|
|
|
## value is mostly ignored but carried through.
|
|
|
|
##
|
|
|
|
## * `Proof`: There is a list of hexary nodes which allow to build a partial
|
|
|
|
## Patricia-Merkle trie starting at the state root with all the account
|
|
|
|
## leaves. There are enough nodes that show that there is no account before
|
|
|
|
## the least account (which is currently ignored.)
|
|
|
|
##
|
|
|
|
## There are test data samples on the sub-directory `test_sync_snap`. These
|
|
|
|
## are complete replies for some (admittedly snap) test requests from a `kiln#`
|
|
|
|
## session.
|
|
|
|
##
|
|
|
|
## There are three tests:
|
|
|
|
##
|
|
|
|
## 1. Run the `test_accountsImport()` function which is the all-in-one
|
|
|
|
## production function processoing the data described above. The test
|
|
|
|
## applies it sequentially to all argument data sets.
|
|
|
|
##
|
|
|
|
## 2. With `test_accountsMergeProofs()` individual items are tested which are
|
|
|
|
## hidden in test 1. while merging the sample data.
|
|
|
|
## * Load/accumulate `proofs` data from several samples
|
|
|
|
## * Load/accumulate accounts (needs some unique sorting)
|
|
|
|
## * Build/complete hexary trie for accounts
|
|
|
|
## * Save/bulk-store hexary trie on disk. If rocksdb is available, data
|
|
|
|
## are bulk stored via sst.
|
|
|
|
##
|
|
|
|
## 3. The function `test_accountsRevisitStoredItems()` traverses trie nodes
|
|
|
|
## stored earlier. The accounts from test 2 are re-visted using the account
|
|
|
|
## hash as access path.
|
|
|
|
##
|
2023-01-23 16:09:12 +00:00
|
|
|
|
|
|
|
import
|
2023-01-30 17:50:58 +00:00
|
|
|
std/algorithm,
|
|
|
|
eth/[common, p2p],
|
2023-01-23 16:09:12 +00:00
|
|
|
unittest2,
|
2023-02-15 10:14:40 +00:00
|
|
|
../../nimbus/sync/protocol,
|
2023-01-23 16:09:12 +00:00
|
|
|
../../nimbus/sync/snap/range_desc,
|
2023-03-17 14:46:50 +00:00
|
|
|
../../nimbus/sync/snap/worker/db/[
|
|
|
|
hexary_debug, hexary_desc, hexary_error,
|
|
|
|
snapdb_accounts, snapdb_debug, snapdb_desc],
|
2023-01-23 16:09:12 +00:00
|
|
|
../replay/[pp, undump_accounts],
|
|
|
|
./test_helpers
|
|
|
|
|
|
|
|
# ------------------------------------------------------------------------------
|
|
|
|
# Private helpers
|
|
|
|
# ------------------------------------------------------------------------------
|
|
|
|
|
2023-02-15 10:14:40 +00:00
|
|
|
proc flatten(list: openArray[seq[SnapProof]]): seq[SnapProof] =
|
2023-01-23 16:09:12 +00:00
|
|
|
for w in list:
|
|
|
|
result.add w
|
|
|
|
|
|
|
|
# ------------------------------------------------------------------------------
|
|
|
|
# Public test function
|
|
|
|
# ------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
proc test_accountsImport*(
|
|
|
|
inList: seq[UndumpAccounts];
|
|
|
|
desc: SnapDbAccountsRef;
|
2023-01-30 17:50:58 +00:00
|
|
|
persistent: bool;
|
2023-01-23 16:09:12 +00:00
|
|
|
) =
|
|
|
|
## Import accounts
|
|
|
|
for n,w in inList:
|
|
|
|
check desc.importAccounts(w.base, w.data, persistent).isImportOk
|
|
|
|
|
|
|
|
|
|
|
|
proc test_accountsMergeProofs*(
|
|
|
|
inList: seq[UndumpAccounts];
|
|
|
|
desc: SnapDbAccountsRef;
|
|
|
|
accKeys: var seq[NodeKey];
|
|
|
|
) =
|
|
|
|
## Merge account proofs
|
|
|
|
# Load/accumulate data from several samples (needs some particular sort)
|
2023-03-17 14:46:50 +00:00
|
|
|
let
|
|
|
|
getFn = desc.getAccountFn
|
|
|
|
baseTag = inList.mapIt(it.base).sortMerge
|
|
|
|
packed = PackedAccountRange(
|
|
|
|
accounts: inList.mapIt(it.data.accounts).sortMerge,
|
|
|
|
proof: inList.mapIt(it.data.proof).flatten)
|
|
|
|
nAccounts = packed.accounts.len
|
2023-01-23 16:09:12 +00:00
|
|
|
# Merging intervals will produce gaps, so the result is expected OK but
|
|
|
|
# different from `.isImportOk`
|
|
|
|
check desc.importAccounts(baseTag, packed, true).isOk
|
|
|
|
|
2023-03-07 14:23:22 +00:00
|
|
|
# for debugging, make sure that state root ~ "$0"
|
|
|
|
desc.hexaDb.assignPrettyKeys(desc.root)
|
2023-01-23 16:09:12 +00:00
|
|
|
|
|
|
|
# Update list of accounts. There might be additional accounts in the set
|
|
|
|
# of proof nodes, typically before the `lowerBound` of each block. As
|
|
|
|
# there is a list of account ranges (that were merged for testing), one
|
|
|
|
# need to check for additional records only on either end of a range.
|
|
|
|
var keySet = packed.accounts.mapIt(it.accKey).toHashSet
|
|
|
|
for w in inList:
|
2023-03-17 14:46:50 +00:00
|
|
|
var key = desc.prevAccountsChainDbKey(w.data.accounts[0].accKey, getFn)
|
2023-01-23 16:09:12 +00:00
|
|
|
while key.isOk and key.value notin keySet:
|
|
|
|
keySet.incl key.value
|
2023-03-17 14:46:50 +00:00
|
|
|
let newKey = desc.prevAccountsChainDbKey(key.value, getFn)
|
2023-01-23 16:09:12 +00:00
|
|
|
check newKey != key
|
|
|
|
key = newKey
|
2023-03-17 14:46:50 +00:00
|
|
|
key = desc.nextAccountsChainDbKey(w.data.accounts[^1].accKey, getFn)
|
2023-01-23 16:09:12 +00:00
|
|
|
while key.isOk and key.value notin keySet:
|
|
|
|
keySet.incl key.value
|
2023-03-17 14:46:50 +00:00
|
|
|
let newKey = desc.nextAccountsChainDbKey(key.value, getFn)
|
2023-01-23 16:09:12 +00:00
|
|
|
check newKey != key
|
|
|
|
key = newKey
|
|
|
|
accKeys = toSeq(keySet).mapIt(it.to(NodeTag)).sorted(cmp)
|
|
|
|
.mapIt(it.to(NodeKey))
|
2023-03-17 14:46:50 +00:00
|
|
|
# Some database samples have a few more account keys which come in by the
|
|
|
|
# proof nodes.
|
|
|
|
check nAccounts <= accKeys.len
|
|
|
|
|
|
|
|
# Verify against table importer
|
|
|
|
let
|
|
|
|
xDb = HexaryTreeDbRef.init() # Can dump database with `.pp(xDb)`
|
|
|
|
rc = xDb.fromPersistent(desc.root, getFn, accKeys.len + 100)
|
|
|
|
check rc == Result[int,HexaryError].ok(accKeys.len)
|
2023-01-23 16:09:12 +00:00
|
|
|
|
|
|
|
|
|
|
|
proc test_accountsRevisitStoredItems*(
|
|
|
|
accKeys: seq[NodeKey];
|
|
|
|
desc: SnapDbAccountsRef;
|
|
|
|
noisy = false;
|
|
|
|
) =
|
|
|
|
## Revisit stored items on ChainDBRef
|
2023-03-17 14:46:50 +00:00
|
|
|
let
|
|
|
|
getFn = desc.getAccountFn
|
2023-01-23 16:09:12 +00:00
|
|
|
var
|
|
|
|
nextAccount = accKeys[0]
|
|
|
|
prevAccount: NodeKey
|
|
|
|
count = 0
|
|
|
|
for accKey in accKeys:
|
|
|
|
count.inc
|
|
|
|
let
|
|
|
|
pfx = $count & "#"
|
2023-03-17 14:46:50 +00:00
|
|
|
byChainDB = desc.getAccountsData(accKey, persistent=true)
|
|
|
|
byNextKey = desc.nextAccountsChainDbKey(accKey, getFn)
|
|
|
|
byPrevKey = desc.prevAccountsChainDbKey(accKey, getFn)
|
|
|
|
if byChainDB.isErr:
|
|
|
|
noisy.say "*** find",
|
|
|
|
"<", count, "> byChainDb=", byChainDB.pp
|
|
|
|
check byChainDB.isOk
|
2023-01-23 16:09:12 +00:00
|
|
|
|
|
|
|
# Check `next` traversal funcionality. If `byNextKey.isOk` fails, the
|
|
|
|
# `nextAccount` value is still the old one and will be different from
|
|
|
|
# the account in the next for-loop cycle (if any.)
|
|
|
|
check pfx & accKey.pp(false) == pfx & nextAccount.pp(false)
|
|
|
|
if byNextKey.isOk:
|
|
|
|
nextAccount = byNextKey.get(otherwise = NodeKey.default)
|
|
|
|
|
|
|
|
# Check `prev` traversal funcionality
|
|
|
|
if prevAccount != NodeKey.default:
|
|
|
|
check byPrevKey.isOk
|
|
|
|
if byPrevKey.isOk:
|
|
|
|
check pfx & byPrevKey.value.pp(false) == pfx & prevAccount.pp(false)
|
|
|
|
prevAccount = accKey
|
|
|
|
|
|
|
|
# ------------------------------------------------------------------------------
|
|
|
|
# End
|
|
|
|
# ------------------------------------------------------------------------------
|