827 lines
28 KiB
Nim
827 lines
28 KiB
Nim
# Nimbus - Types, data structures and shared utilities used in network sync
|
|
#
|
|
# Copyright (c) 2018-2021 Status Research & Development GmbH
|
|
# Licensed under either of
|
|
# * Apache License, version 2.0, ([LICENSE-APACHE](LICENSE-APACHE) or
|
|
# http://www.apache.org/licenses/LICENSE-2.0)
|
|
# * MIT license ([LICENSE-MIT](LICENSE-MIT) or
|
|
# http://opensource.org/licenses/MIT)
|
|
# at your option. This file may not be copied, modified, or
|
|
# distributed except according to those terms.
|
|
|
|
## Snap sync components tester
|
|
|
|
import
|
|
std/[algorithm, distros, hashes, math, os,
|
|
sequtils, strformat, strutils, tables, times],
|
|
chronicles,
|
|
eth/[common/eth_types, p2p, rlp, trie/db],
|
|
rocksdb,
|
|
stint,
|
|
stew/[byteutils, results],
|
|
unittest2,
|
|
../nimbus/[chain_config, config, genesis],
|
|
../nimbus/db/[db_chain, select_backend, storage_types],
|
|
../nimbus/p2p/chain,
|
|
../nimbus/sync/[types, protocol],
|
|
../nimbus/sync/snap/range_desc,
|
|
../nimbus/sync/snap/worker/[accounts_db, db/hexary_desc, db/rocky_bulk_load],
|
|
../nimbus/utils/prettify,
|
|
./replay/[pp, undump],
|
|
./test_sync_snap/sample0
|
|
|
|
const
|
|
baseDir = [".", "..", ".."/"..", $DirSep]
|
|
repoDir = [".", "tests"/"replay", "tests"/"test_sync_snap",
|
|
"nimbus-eth1-blobs"/"replay"]
|
|
|
|
nTestDbInstances = 9
|
|
|
|
type
|
|
CaptureSpecs = tuple
|
|
name: string ## sample name, also used as sub-directory for db separation
|
|
network: NetworkId
|
|
file: string ## name of capture file
|
|
numBlocks: int ## Number of blocks to load
|
|
|
|
AccountsProofSample = object
|
|
name: string ## sample name, also used as sub-directory for db separation
|
|
root: Hash256
|
|
data: seq[TestSample]
|
|
|
|
TestSample = tuple
|
|
## Data layout provided by the data dump `sample0.nim`
|
|
base: Hash256
|
|
accounts: seq[(Hash256,uint64,UInt256,Hash256,Hash256)]
|
|
proofs: seq[Blob]
|
|
|
|
TestItem = object
|
|
## Palatable input format for test function
|
|
base: NodeTag
|
|
data: SnapAccountRange
|
|
|
|
TestDbs = object
|
|
## Provide enough spare empty databases
|
|
persistent: bool
|
|
dbDir: string
|
|
cdb: array[nTestDbInstances,ChainDb]
|
|
|
|
when defined(linux):
|
|
# The `detectOs(Ubuntu)` directive is not Windows compatible, causes an
|
|
# error when running the system command `lsb_release -d` in the background.
|
|
let isUbuntu32bit = detectOs(Ubuntu) and int.sizeof == 4
|
|
else:
|
|
const isUbuntu32bit = false
|
|
|
|
const
|
|
goerliCapture: CaptureSpecs = (
|
|
name: "goerli",
|
|
network: GoerliNet,
|
|
file: "goerli68161.txt.gz",
|
|
numBlocks: 1_000)
|
|
|
|
accSample0 = AccountsProofSample(
|
|
name: "sample0",
|
|
root: sample0.snapRoot,
|
|
data: sample0.snapProofData)
|
|
|
|
let
|
|
# Forces `check()` to print the error (as opposed when using `isOk()`)
|
|
OkHexDb = Result[void,HexaryDbError].ok()
|
|
|
|
# There was a problem with the Github/CI which results in spurious crashes
|
|
# when leaving the `runner()` if the persistent BaseChainDB initialisation
|
|
# was present, see `test_custom_network` for more details.
|
|
disablePersistentDB = isUbuntu32bit
|
|
|
|
var
|
|
xTmpDir: string
|
|
xDbs: TestDbs # for repeated storage/overwrite tests
|
|
xTab32: Table[ByteArray32,Blob] # extracted data
|
|
xTab33: Table[ByteArray33,Blob]
|
|
xVal32Sum, xVal32SqSum: float # statistics
|
|
xVal33Sum, xVal33SqSum: float
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# Helpers
|
|
# ------------------------------------------------------------------------------
|
|
|
|
proc isOk(rc: ValidationResult): bool =
|
|
rc == ValidationResult.OK
|
|
|
|
proc findFilePath(file: string;
|
|
baseDir, repoDir: openArray[string]): Result[string,void] =
|
|
for dir in baseDir:
|
|
for repo in repoDir:
|
|
let path = dir / repo / file
|
|
if path.fileExists:
|
|
return ok(path)
|
|
err()
|
|
|
|
proc getTmpDir(sampleDir = "sample0.nim"): string =
|
|
sampleDir.findFilePath(baseDir,repoDir).value.splitFile.dir
|
|
|
|
proc pp(d: Duration): string =
|
|
if 40 < d.inSeconds:
|
|
d.ppMins
|
|
elif 200 < d.inMilliseconds:
|
|
d.ppSecs
|
|
elif 200 < d.inMicroseconds:
|
|
d.ppMs
|
|
else:
|
|
d.ppUs
|
|
|
|
proc pp(d: AccountLoadStats): string =
|
|
"[" & d.size.toSeq.mapIt(it.toSI).join(",") & "," &
|
|
d.dura.toSeq.mapIt(it.pp).join(",") & "]"
|
|
|
|
proc pp(rc: Result[Account,HexaryDbError]): string =
|
|
if rc.isErr: $rc.error else: rc.value.pp
|
|
|
|
proc ppKvPc(w: openArray[(string,int)]): string =
|
|
w.mapIt(&"{it[0]}={it[1]}%").join(", ")
|
|
|
|
proc say*(noisy = false; pfx = "***"; args: varargs[string, `$`]) =
|
|
if noisy:
|
|
if args.len == 0:
|
|
echo "*** ", pfx
|
|
elif 0 < pfx.len and pfx[^1] != ' ':
|
|
echo pfx, " ", args.toSeq.join
|
|
else:
|
|
echo pfx, args.toSeq.join
|
|
|
|
proc setTraceLevel =
|
|
discard
|
|
when defined(chronicles_runtime_filtering) and loggingEnabled:
|
|
setLogLevel(LogLevel.TRACE)
|
|
|
|
proc setErrorLevel =
|
|
discard
|
|
when defined(chronicles_runtime_filtering) and loggingEnabled:
|
|
setLogLevel(LogLevel.ERROR)
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# Private functions
|
|
# ------------------------------------------------------------------------------
|
|
|
|
proc to(data: seq[TestSample]; T: type seq[TestItem]): T =
|
|
## Convert test data into usable format
|
|
for r in data:
|
|
result.add TestItem(
|
|
base: r.base.to(NodeTag),
|
|
data: SnapAccountRange(
|
|
proof: r.proofs,
|
|
accounts: r.accounts.mapIt(
|
|
SnapAccount(
|
|
accHash: it[0],
|
|
accBody: Account(
|
|
nonce: it[1],
|
|
balance: it[2],
|
|
storageRoot: it[3],
|
|
codeHash: it[4])))))
|
|
|
|
proc to(b: openArray[byte]; T: type ByteArray32): T =
|
|
## Convert to other representation (or exception)
|
|
if b.len == 32:
|
|
(addr result[0]).copyMem(unsafeAddr b[0], 32)
|
|
else:
|
|
doAssert b.len == 32
|
|
|
|
proc to(b: openArray[byte]; T: type ByteArray33): T =
|
|
## Convert to other representation (or exception)
|
|
if b.len == 33:
|
|
(addr result[0]).copyMem(unsafeAddr b[0], 33)
|
|
else:
|
|
doAssert b.len == 33
|
|
|
|
proc to(b: ByteArray32|ByteArray33; T: type Blob): T =
|
|
b.toSeq
|
|
|
|
proc to(b: openArray[byte]; T: type NodeTag): T =
|
|
## Convert from serialised equivalent
|
|
UInt256.fromBytesBE(b).T
|
|
|
|
proc to(w: (byte, NodeTag); T: type Blob): T =
|
|
let (b,t) = w
|
|
@[b] & toSeq(t.UInt256.toBytesBE)
|
|
|
|
proc to(t: NodeTag; T: type Blob): T =
|
|
toSeq(t.UInt256.toBytesBE)
|
|
|
|
proc flushDbDir(s: string; subDir = "") =
|
|
if s != "":
|
|
let baseDir = s / "tmp"
|
|
for n in 0 ..< nTestDbInstances:
|
|
let instDir = if subDir == "": baseDir / $n else: baseDir / subDir / $n
|
|
if (instDir / "nimbus" / "data").dirExists:
|
|
# Typically under Windows: there might be stale file locks.
|
|
try: instDir.removeDir except: discard
|
|
try: (baseDir / subDir).removeDir except: discard
|
|
block dontClearUnlessEmpty:
|
|
for w in baseDir.walkDir:
|
|
break dontClearUnlessEmpty
|
|
try: baseDir.removeDir except: discard
|
|
|
|
proc testDbs(workDir = ""; subDir = ""): TestDbs =
|
|
if disablePersistentDB or workDir == "":
|
|
result.persistent = false
|
|
result.dbDir = "*notused*"
|
|
else:
|
|
result.persistent = true
|
|
if subDir != "":
|
|
result.dbDir = workDir / "tmp" / subDir
|
|
else:
|
|
result.dbDir = workDir / "tmp"
|
|
if result.persistent:
|
|
result.dbDir.flushDbDir
|
|
for n in 0 ..< result.cdb.len:
|
|
result.cdb[n] = (result.dbDir / $n).newChainDB
|
|
|
|
proc lastTwo(a: openArray[string]): seq[string] =
|
|
if 1 < a.len: @[a[^2],a[^1]] else: a.toSeq
|
|
|
|
proc thisRecord(r: rocksdb_iterator_t): (Blob,Blob) =
|
|
var kLen, vLen: csize_t
|
|
let
|
|
kData = r.rocksdb_iter_key(addr kLen)
|
|
vData = r.rocksdb_iter_value(addr vLen)
|
|
if not kData.isNil and not vData.isNil:
|
|
let
|
|
key = string.fromBytes(toOpenArrayByte(kData,0,int(kLen)-1))
|
|
value = string.fromBytes(toOpenArrayByte(vData,0,int(vLen)-1))
|
|
return (key.mapIt(it.byte),value.mapIt(it.byte))
|
|
|
|
proc meanStdDev(sum, sqSum: float; length: int): (float,float) =
|
|
if 0 < length:
|
|
result[0] = sum / length.float
|
|
result[1] = sqrt(sqSum / length.float - result[0] * result[0])
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# Test Runners
|
|
# ------------------------------------------------------------------------------
|
|
|
|
proc accountsRunner(noisy = true; persistent = true; sample = accSample0) =
|
|
let
|
|
peer = Peer.new
|
|
root = sample.root
|
|
testItemLst = sample.data.to(seq[TestItem])
|
|
tmpDir = getTmpDir()
|
|
db = if persistent: tmpDir.testDbs(sample.name) else: testDbs()
|
|
dbDir = db.dbDir.split($DirSep).lastTwo.join($DirSep)
|
|
info = if db.persistent: &"persistent db on \"{dbDir}\""
|
|
else: "in-memory db"
|
|
|
|
defer:
|
|
if db.persistent:
|
|
tmpDir.flushDbDir(sample.name)
|
|
|
|
suite &"SyncSnap: {sample.name} accounts and proofs for {info}":
|
|
var
|
|
desc: AccountsDbSessionRef
|
|
accounts: seq[SnapAccount]
|
|
|
|
test &"Snap-proofing {testItemLst.len} items for state root ..{root.pp}":
|
|
let dbBase = if persistent: AccountsDbRef.init(db.cdb[0])
|
|
else: AccountsDbRef.init(newMemoryDB())
|
|
for n,w in testItemLst:
|
|
check dbBase.importAccounts(
|
|
peer, root, w.base, w.data, storeData = persistent) == OkHexDb
|
|
noisy.say "***", "import stats=", dbBase.dbImportStats.pp
|
|
|
|
test &"Merging {testItemLst.len} proofs for state root ..{root.pp}":
|
|
let dbBase = if persistent: AccountsDbRef.init(db.cdb[1])
|
|
else: AccountsDbRef.init(newMemoryDB())
|
|
desc = AccountsDbSessionRef.init(dbBase, root, peer)
|
|
for w in testItemLst:
|
|
check desc.merge(w.data.proof) == OkHexDb
|
|
let base = testItemLst.mapIt(it.base).sortMerge
|
|
accounts = testItemLst.mapIt(it.data.accounts).sortMerge
|
|
check desc.merge(base, accounts) == OkHexDb
|
|
desc.assignPrettyKeys() # for debugging (if any)
|
|
check desc.interpolate() == OkHexDb
|
|
|
|
check desc.dbImports() == OkHexDb
|
|
noisy.say "***", "import stats=", desc.dbImportStats.pp
|
|
|
|
test &"Revisting {accounts.len} items stored items on BaseChainDb":
|
|
for acc in accounts:
|
|
let
|
|
byChainDB = desc.getChainDbAccount(acc.accHash)
|
|
byBulker = desc.getBulkDbXAccount(acc.accHash)
|
|
noisy.say "*** find",
|
|
"byChainDb=", byChainDB.pp, " inBulker=", byBulker.pp
|
|
check byChainDB.isOk
|
|
if desc.dbBackendRocksDb():
|
|
check byBulker.isOk
|
|
check byChainDB == byBulker
|
|
|
|
#noisy.say "***", "database dump\n ", desc.dumpProofsDB.join("\n ")
|
|
|
|
|
|
proc importRunner(noisy = true; persistent = true; capture = goerliCapture) =
|
|
|
|
let
|
|
fileInfo = capture.file.splitFile.name.split(".")[0]
|
|
filePath = capture.file.findFilePath(baseDir,repoDir).value
|
|
tmpDir = getTmpDir()
|
|
db = if persistent: tmpDir.testDbs(capture.name) else: testDbs()
|
|
numBlocksInfo = if capture.numBlocks == high(int): ""
|
|
else: $capture.numBlocks & " "
|
|
loadNoise = noisy
|
|
|
|
defer:
|
|
if db.persistent:
|
|
tmpDir.flushDbDir(capture.name)
|
|
|
|
suite &"SyncSnap: using {fileInfo} capture for testing db timings":
|
|
var
|
|
ddb: BaseChainDB # perstent DB on disk
|
|
chn: Chain
|
|
|
|
test &"Create persistent BaseChainDB on {tmpDir}":
|
|
let chainDb = if db.persistent: db.cdb[0].trieDB
|
|
else: newMemoryDB()
|
|
|
|
# Constructor ...
|
|
ddb = newBaseChainDB(
|
|
chainDb,
|
|
id = capture.network,
|
|
pruneTrie = true,
|
|
params = capture.network.networkParams)
|
|
|
|
ddb.initializeEmptyDb
|
|
chn = ddb.newChain
|
|
|
|
test &"Storing {numBlocksInfo}persistent blocks from dump":
|
|
for w in filePath.undumpNextGroup:
|
|
let (fromBlock, toBlock) = (w[0][0].blockNumber, w[0][^1].blockNumber)
|
|
if fromBlock == 0.u256:
|
|
doAssert w[0][0] == ddb.getBlockHeader(0.u256)
|
|
continue
|
|
# Message if [fromBlock,toBlock] contains a multiple of 700
|
|
if fromBlock + (toBlock mod 900) <= toBlock:
|
|
loadNoise.say "***", &"processing ...[#{fromBlock},#{toBlock}]..."
|
|
check chn.persistBlocks(w[0], w[1]).isOk
|
|
if capture.numBlocks.toBlockNumber <= w[0][^1].blockNumber:
|
|
break
|
|
|
|
test "Extract key-value records into memory tables via rocksdb iterator":
|
|
# Implicit test: if not persistent => db.cdb[0] is nil
|
|
if db.cdb[0].rocksStoreRef.isNil:
|
|
skip()
|
|
else:
|
|
let
|
|
rdb = db.cdb[0].rocksStoreRef
|
|
rop = rdb.store.readOptions
|
|
rit = rdb.store.db.rocksdb_create_iterator(rop)
|
|
check not rit.isNil
|
|
|
|
xTab32.clear
|
|
xTab33.clear
|
|
|
|
rit.rocksdb_iter_seek_to_first()
|
|
while rit.rocksdb_iter_valid() != 0:
|
|
let (key,val) = rit.thisRecord()
|
|
rit.rocksdb_iter_next()
|
|
if key.len == 32:
|
|
xTab32[key.to(ByteArray32)] = val
|
|
xVal32Sum += val.len.float
|
|
xVal32SqSum += val.len.float * val.len.float
|
|
check key.to(ByteArray32).to(Blob) == key
|
|
elif key.len == 33:
|
|
xTab33[key.to(ByteArray33)] = val
|
|
xVal33Sum += val.len.float
|
|
xVal33SqSum += val.len.float * val.len.float
|
|
check key.to(ByteArray33).to(Blob) == key
|
|
else:
|
|
noisy.say "***", "ignoring key=", key.toHex
|
|
|
|
rit.rocksdb_iter_destroy()
|
|
|
|
var
|
|
(mean32, stdv32) = meanStdDev(xVal32Sum, xVal32SqSum, xTab32.len)
|
|
(mean33, stdv33) = meanStdDev(xVal33Sum, xVal33SqSum, xTab33.len)
|
|
noisy.say "***",
|
|
"key 32 table: ",
|
|
&"size={xTab32.len} valLen={(mean32+0.5).int}({(stdv32+0.5).int})",
|
|
", key 33 table: ",
|
|
&"size={xTab33.len} valLen={(mean33+0.5).int}({(stdv33+0.5).int})"
|
|
|
|
|
|
proc storeRunner(noisy = true; persistent = true; cleanUp = true) =
|
|
let
|
|
fullNoise = false
|
|
var
|
|
emptyDb = "empty"
|
|
|
|
# Allows to repeat storing on existing data
|
|
if not xDbs.cdb[0].isNil:
|
|
emptyDb = "pre-loaded"
|
|
elif persistent:
|
|
xTmpDir = getTmpDir()
|
|
xDbs = xTmpDir.testDbs("store-runner")
|
|
else:
|
|
xDbs = testDbs()
|
|
|
|
defer:
|
|
if xDbs.persistent and cleanUp:
|
|
xTmpDir.flushDbDir("store-runner")
|
|
xDbs.reset
|
|
|
|
suite &"SyncSnap: storage tests on {emptyDb} databases":
|
|
#
|
|
# `xDbs` instance slots layout:
|
|
#
|
|
# * cdb[0] -- direct db, key length 32, no transaction
|
|
# * cdb[1] -- direct db, key length 32 as 33, no transaction
|
|
#
|
|
# * cdb[2] -- direct db, key length 32, transaction based
|
|
# * cdb[3] -- direct db, key length 32 as 33, transaction based
|
|
#
|
|
# * cdb[4] -- direct db, key length 33, no transaction
|
|
# * cdb[5] -- direct db, key length 33, transaction based
|
|
#
|
|
# * cdb[6] -- rocksdb, key length 32
|
|
# * cdb[7] -- rocksdb, key length 32 as 33
|
|
# * cdb[8] -- rocksdb, key length 33
|
|
#
|
|
doAssert 9 <= nTestDbInstances
|
|
|
|
if xTab32.len == 0 or xTab33.len == 0:
|
|
test &"Both tables with 32 byte keys(size={xTab32.len}), " &
|
|
&"33 byte keys(size={xTab32.len}) must be non-empty":
|
|
skip()
|
|
else:
|
|
# cdb[0] -- direct db, key length 32, no transaction
|
|
test &"Directly store {xTab32.len} records " &
|
|
&"(key length 32) into {emptyDb} trie database":
|
|
var ela: Duration
|
|
let tdb = xDbs.cdb[0].trieDB
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Standard db loader(keyLen 32)", ela):
|
|
for (key,val) in xTab32.pairs:
|
|
tdb.put(key, val)
|
|
|
|
if ela.inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.inNanoseconds.float
|
|
perRec = ((elaNs / xTab32.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab32.len, ", ",
|
|
"perRecord=", perRec.pp
|
|
|
|
# cdb[1] -- direct db, key length 32 as 33, no transaction
|
|
test &"Directly store {xTab32.len} records " &
|
|
&"(key length 33) into {emptyDb} trie database":
|
|
var ela = initDuration()
|
|
let tdb = xDbs.cdb[1].trieDB
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Standard db loader(keyLen 32 as 33)", ela):
|
|
for (key,val) in xTab32.pairs:
|
|
tdb.put(@[99.byte] & key.toSeq, val)
|
|
|
|
if ela.inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.inNanoseconds.float
|
|
perRec = ((elaNs / xTab32.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab32.len, ", ",
|
|
"perRecord=", perRec.pp
|
|
|
|
# cdb[2] -- direct db, key length 32, transaction based
|
|
test &"Transactionally store {xTab32.len} records " &
|
|
&"(key length 32) into {emptyDb} trie database":
|
|
var ela: Duration
|
|
let tdb = xDbs.cdb[2].trieDB
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Standard db loader(tx,keyLen 32)", ela):
|
|
let dbTx = tdb.beginTransaction
|
|
defer: dbTx.commit
|
|
|
|
for (key,val) in xTab32.pairs:
|
|
tdb.put(key, val)
|
|
|
|
if ela.inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.inNanoseconds.float
|
|
perRec = ((elaNs / xTab32.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab32.len, ", ",
|
|
"perRecord=", perRec.pp
|
|
|
|
# cdb[3] -- direct db, key length 32 as 33, transaction based
|
|
test &"Transactionally store {xTab32.len} records " &
|
|
&"(key length 33) into {emptyDb} trie database":
|
|
var ela: Duration
|
|
let tdb = xDbs.cdb[3].trieDB
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Standard db loader(tx,keyLen 32 as 33)", ela):
|
|
let dbTx = tdb.beginTransaction
|
|
defer: dbTx.commit
|
|
|
|
for (key,val) in xTab32.pairs:
|
|
tdb.put(@[99.byte] & key.toSeq, val)
|
|
|
|
if ela.inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.inNanoseconds.float
|
|
perRec = ((elaNs / xTab32.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab32.len, ", ",
|
|
"perRecord=", perRec.pp
|
|
|
|
# cdb[4] -- direct db, key length 33, no transaction
|
|
test &"Directly store {xTab33.len} records " &
|
|
&"(key length 33) into {emptyDb} trie database":
|
|
var ela: Duration
|
|
let tdb = xDbs.cdb[4].trieDB
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Standard db loader(keyLen 33)", ela):
|
|
for (key,val) in xTab33.pairs:
|
|
tdb.put(key, val)
|
|
|
|
if ela.inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.inNanoseconds.float
|
|
perRec = ((elaNs / xTab33.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab33.len, ", ",
|
|
"perRecord=", perRec.pp
|
|
|
|
# cdb[5] -- direct db, key length 33, transaction based
|
|
test &"Transactionally store {xTab33.len} records " &
|
|
&"(key length 33) into {emptyDb} trie database":
|
|
var ela: Duration
|
|
let tdb = xDbs.cdb[5].trieDB
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Standard db loader(tx,keyLen 33)", ela):
|
|
let dbTx = tdb.beginTransaction
|
|
defer: dbTx.commit
|
|
|
|
for (key,val) in xTab33.pairs:
|
|
tdb.put(key, val)
|
|
|
|
if ela.inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.inNanoseconds.float
|
|
perRec = ((elaNs / xTab33.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab33.len, ", ",
|
|
"perRecord=", perRec.pp
|
|
|
|
if xDbs.cdb[0].rocksStoreRef.isNil:
|
|
test "The rocksdb interface must be available": skip()
|
|
else:
|
|
# cdb[6] -- rocksdb, key length 32
|
|
test &"Store {xTab32.len} records " &
|
|
"(key length 32) into empty rocksdb table":
|
|
var
|
|
ela: array[4,Duration]
|
|
size: int64
|
|
let
|
|
rdb = xDbs.cdb[6].rocksStoreRef
|
|
|
|
# Note that 32 and 33 size keys cannot be usefiully merged into the
|
|
# same SST file. The keys must be added in a sorted mode. So playing
|
|
# safe, key sizes should be of
|
|
# equal length.
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Rocky bulk loader(keyLen 32)", ela[0]):
|
|
let bulker = RockyBulkLoadRef.init(rdb)
|
|
defer: bulker.destroy()
|
|
check bulker.begin("rocky-bulk-cache")
|
|
|
|
var
|
|
keyList = newSeq[NodeTag](xTab32.len)
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/32, sorter", ela[1]):
|
|
var inx = 0
|
|
for key in xTab32.keys:
|
|
keyList[inx] = key.to(NodeTag)
|
|
inx.inc
|
|
keyList.sort(cmp)
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/32, append", ela[2]):
|
|
for n,nodeTag in keyList:
|
|
let key = nodeTag.to(Blob)
|
|
check bulker.add(key, xTab32[key.to(ByteArray32)])
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/32, slurp", ela[3]):
|
|
let rc = bulker.finish()
|
|
if rc.isOk:
|
|
size = rc.value
|
|
else:
|
|
check bulker.lastError == "" # force printing error
|
|
|
|
fullNoise.say "***", " ela[]=", $ela.toSeq.mapIt(it.pp)
|
|
if ela[0].inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.toSeq.mapIt(it.inNanoseconds.float)
|
|
elaPc = elaNs.mapIt(((it / elaNs[0]) * 100 + 0.5).int)
|
|
perRec = ((elaNs[0] / xTab32.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab32.len, ", ",
|
|
"perRecord=", perRec.pp, ", ",
|
|
"sstSize=", size.uint64.toSI, ", ",
|
|
"perRecord=", ((size.float / xTab32.len.float) + 0.5).int, ", ",
|
|
["Total","Sorter","Append","Ingest"].zip(elaPc).ppKvPc
|
|
|
|
# cdb[7] -- rocksdb, key length 32 as 33
|
|
test &"Store {xTab32.len} records " &
|
|
"(key length 33) into empty rocksdb table":
|
|
var
|
|
ela: array[4,Duration]
|
|
size: int64
|
|
let
|
|
rdb = xDbs.cdb[7].rocksStoreRef
|
|
|
|
# Note that 32 and 33 size keys cannot be usefiully merged into the
|
|
# same SST file. The keys must be added in a sorted mode. So playing
|
|
# safe, key sizes should be of
|
|
# equal length.
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Rocky bulk loader(keyLen 32 as 33)", ela[0]):
|
|
let bulker = RockyBulkLoadRef.init(rdb)
|
|
defer: bulker.destroy()
|
|
check bulker.begin("rocky-bulk-cache")
|
|
|
|
var
|
|
keyList = newSeq[NodeTag](xTab32.len)
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/32 as 33, sorter", ela[1]):
|
|
var inx = 0
|
|
for key in xTab32.keys:
|
|
keyList[inx] = key.to(NodeTag)
|
|
inx.inc
|
|
keyList.sort(cmp)
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/32 as 33, append", ela[2]):
|
|
for n,nodeTag in keyList:
|
|
let key = nodeTag.to(Blob)
|
|
check bulker.add(@[99.byte] & key, xTab32[key.to(ByteArray32)])
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/32 as 33, slurp", ela[3]):
|
|
let rc = bulker.finish()
|
|
if rc.isOk:
|
|
size = rc.value
|
|
else:
|
|
check bulker.lastError == "" # force printing error
|
|
|
|
fullNoise.say "***", " ela[]=", $ela.toSeq.mapIt(it.pp)
|
|
if ela[0].inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.toSeq.mapIt(it.inNanoseconds.float)
|
|
elaPc = elaNs.mapIt(((it / elaNs[0]) * 100 + 0.5).int)
|
|
perRec = ((elaNs[0] / xTab32.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab32.len, ", ",
|
|
"perRecord=", perRec.pp, ", ",
|
|
"sstSize=", size.uint64.toSI, ", ",
|
|
"perRecord=", ((size.float / xTab32.len.float) + 0.5).int, ", ",
|
|
["Total","Sorter","Append","Ingest"].zip(elaPc).ppKvPc
|
|
|
|
|
|
# cdb[8] -- rocksdb, key length 33
|
|
test &"Store {xTab33.len} records " &
|
|
&"(key length 33) into {emptyDb} rocksdb table":
|
|
var
|
|
ela: array[4,Duration]
|
|
size: int64
|
|
let rdb = xDbs.cdb[8].rocksStoreRef
|
|
|
|
# Note that 32 and 33 size keys cannot be usefiully merged into the
|
|
# same SST file. The keys must be added in a sorted mode. So playing
|
|
# safe, key sizes should be of equal length.
|
|
|
|
if noisy: echo ""
|
|
noisy.showElapsed("Rocky bulk loader(keyLen 33)", ela[0]):
|
|
let bulker = RockyBulkLoadRef.init(rdb)
|
|
defer: bulker.destroy()
|
|
check bulker.begin("rocky-bulk-cache")
|
|
|
|
var
|
|
kKeys: seq[byte] # need to cacscade
|
|
kTab: Table[byte,seq[NodeTag]]
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/33, sorter", ela[1]):
|
|
for key in xTab33.keys:
|
|
if kTab.hasKey(key[0]):
|
|
kTab[key[0]].add key.toOpenArray(1,32).to(NodeTag)
|
|
else:
|
|
kTab[key[0]] = @[key.toOpenArray(1,32).to(NodeTag)]
|
|
|
|
kKeys = toSeq(kTab.keys).sorted
|
|
for w in kKeys:
|
|
kTab[w].sort(cmp)
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/33, append", ela[2]):
|
|
for w in kKeys:
|
|
fullNoise.say "***", " prefix=", w, " entries=", kTab[w].len
|
|
for n,nodeTag in kTab[w]:
|
|
let key = (w,nodeTag).to(Blob)
|
|
check bulker.add(key, xTab33[key.to(ByteArray33)])
|
|
|
|
fullNoise.showElapsed("Rocky bulk loader/33, slurp", ela[3]):
|
|
let rc = bulker.finish()
|
|
if rc.isOk:
|
|
size = rc.value
|
|
else:
|
|
check bulker.lastError == "" # force printing error
|
|
|
|
fullNoise.say "***", " ela[]=", $ela.toSeq.mapIt(it.pp)
|
|
if ela[0].inNanoseconds != 0:
|
|
let
|
|
elaNs = ela.toSeq.mapIt(it.inNanoseconds.float)
|
|
elaPc = elaNs.mapIt(((it / elaNs[0]) * 100 + 0.5).int)
|
|
perRec = ((elaNs[0] / xTab33.len.float) + 0.5).int.initDuration
|
|
noisy.say "***",
|
|
"nRecords=", xTab33.len, ", ",
|
|
"perRecord=", perRec.pp, ", ",
|
|
"sstSize=", size.uint64.toSI, ", ",
|
|
"perRecord=", ((size.float / xTab33.len.float) + 0.5).int, ", ",
|
|
["Total","Cascaded-Sorter","Append","Ingest"].zip(elaPc).ppKvPc
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# Main function(s)
|
|
# ------------------------------------------------------------------------------
|
|
|
|
proc syncSnapMain*(noisy = defined(debug)) =
|
|
# Caveat: running `accountsRunner(persistent=true)` twice will crash as the
|
|
# persistent database might not be fully cleared due to some stale
|
|
# locks.
|
|
noisy.accountsRunner(persistent=true)
|
|
noisy.accountsRunner(persistent=false)
|
|
noisy.importRunner() # small sample, just verify functionality
|
|
noisy.storeRunner()
|
|
|
|
when isMainModule:
|
|
const
|
|
noisy = defined(debug) or true
|
|
snapTest0 = accSample0
|
|
snapTest1 = AccountsProofSample(
|
|
name: "test1",
|
|
root: snapTest0.root,
|
|
data: snapTest0.data[0..0])
|
|
|
|
bulkTest0 = goerliCapture
|
|
bulkTest1: CaptureSpecs = (
|
|
name: "full-goerli",
|
|
network: goerliCapture.network,
|
|
file: goerliCapture.file,
|
|
numBlocks: high(int))
|
|
bulkTest2: CaptureSpecs = (
|
|
name: "more-goerli",
|
|
network: GoerliNet,
|
|
file: "goerli482304.txt.gz",
|
|
numBlocks: high(int))
|
|
bulkTest3: CaptureSpecs = (
|
|
name: "mainnet",
|
|
network: MainNet,
|
|
file: "mainnet332160.txt.gz",
|
|
numBlocks: high(int))
|
|
|
|
when false: # or true:
|
|
import ../../nimbus-eth1-blobs/replay/sync_sample1 as sample1
|
|
const
|
|
snapTest2 = AccountsProofSample(
|
|
name: "test2",
|
|
root: sample1.snapRoot,
|
|
data: sample1.snapProofData)
|
|
snapTest3 = AccountsProofSample(
|
|
name: "test3",
|
|
root: snapTest2.root,
|
|
data: snapTest2.data[0..0])
|
|
|
|
#setTraceLevel()
|
|
setErrorLevel()
|
|
|
|
false.accountsRunner(persistent=true, snapTest0)
|
|
false.accountsRunner(persistent=false, snapTest0)
|
|
#noisy.accountsRunner(persistent=true, snapTest1)
|
|
|
|
when declared(snapTest2):
|
|
noisy.accountsRunner(persistent=false, snapTest2)
|
|
#noisy.accountsRunner(persistent=true, snapTest3)
|
|
|
|
when true: # and false:
|
|
# ---- database storage timings -------
|
|
|
|
noisy.showElapsed("importRunner()"):
|
|
noisy.importRunner(capture = bulkTest0)
|
|
|
|
noisy.showElapsed("storeRunner()"):
|
|
true.storeRunner(cleanUp = false)
|
|
true.storeRunner()
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# End
|
|
# ------------------------------------------------------------------------------
|