mirror of
https://github.com/status-im/nimbus-eth1.git
synced 2025-01-18 00:01:07 +00:00
d39c589ec3
* replace rocksdb row cache with larger rdb lru caches - these serve the same purpose but are more efficient because they skips serialization, locking and rocksdb layering * don't append fresh items to cache - this has the effect of evicting the existing items and replacing them with low-value entries that might never be read - during write-heavy periods of processing, the newly-added entries were evicted during the store loop * allow tuning rdb lru size at runtime * add (hidden) option to print lru stats at exit (replacing the compile-time flag) pre: ``` INF 2024-09-03 15:07:01.136+02:00 Imported blocks blockNumber=20012001 blocks=12000 importedSlot=9216851 txs=1837042 mgas=181911.265 bps=11.675 tps=1870.397 mgps=176.819 avgBps=10.288 avgTps=1574.889 avgMGps=155.952 elapsed=19m26s458ms ``` post: ``` INF 2024-09-03 13:54:26.730+02:00 Imported blocks blockNumber=20012001 blocks=12000 importedSlot=9216851 txs=1837042 mgas=181911.265 bps=11.637 tps=1864.384 mgps=176.250 avgBps=11.202 avgTps=1714.920 avgMGps=169.818 elapsed=17m51s211ms ``` 9%:ish import perf improvement on similar mem usage :)
185 lines
6.3 KiB
Nim
185 lines
6.3 KiB
Nim
# nimbus-eth1
|
|
# Copyright (c) 2023-2024 Status Research & Development GmbH
|
|
# Licensed under either of
|
|
# * Apache License, version 2.0, ([LICENSE-APACHE](LICENSE-APACHE) or
|
|
# http://www.apache.org/licenses/LICENSE-2.0)
|
|
# * MIT license ([LICENSE-MIT](LICENSE-MIT) or
|
|
# http://opensource.org/licenses/MIT)
|
|
# at your option. This file may not be copied, modified, or distributed
|
|
# except according to those terms.
|
|
|
|
## Rocksdb constructor/destructor for Aristo DB
|
|
## ============================================
|
|
|
|
{.push raises: [].}
|
|
|
|
import
|
|
std/[exitprocs, sets, sequtils, strformat, os],
|
|
rocksdb,
|
|
results,
|
|
../../aristo_desc,
|
|
./rdb_desc,
|
|
../../../opts
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# Private constructor
|
|
# ------------------------------------------------------------------------------
|
|
|
|
const
|
|
lruOverhead = 32
|
|
# Approximate LRU cache overhead per entry - although `keyed_queue` which is
|
|
# currently used has a much larger overhead, 32 is an easily reachable
|
|
# number which likely can be reduced in the future
|
|
|
|
proc dumpCacheStats(keySize, vtxSize: int) =
|
|
block vtx:
|
|
var misses, hits: uint64
|
|
echo "vtxLru(", vtxSize, ")"
|
|
echo " state vtype miss hit total hitrate"
|
|
for state in RdbStateType:
|
|
for vtype in VertexType:
|
|
let
|
|
(miss, hit) = (
|
|
rdbVtxLruStats[state][vtype].get(false),
|
|
rdbVtxLruStats[state][vtype].get(true),
|
|
)
|
|
hitRate = float64(hit * 100) / (float64(hit + miss))
|
|
misses += miss
|
|
hits += hit
|
|
echo &"{state:>8} {vtype:>8} {miss:>10} {hit:>10} {miss+hit:>10} {hitRate:>6.2f}%"
|
|
let hitRate = float64(hits * 100) / (float64(hits + misses))
|
|
echo &" all all {misses:>10} {hits:>10} {misses+hits:>10} {hitRate:>6.2f}%"
|
|
|
|
block key:
|
|
var misses, hits: uint64
|
|
echo "keyLru(", keySize, ") "
|
|
|
|
echo " state miss hit total hitrate"
|
|
|
|
for state in RdbStateType:
|
|
let
|
|
(miss, hit) =
|
|
(rdbKeyLruStats[state].get(false), rdbKeyLruStats[state].get(true))
|
|
hitRate = float64(hit * 100) / (float64(hit + miss))
|
|
misses += miss
|
|
hits += hit
|
|
|
|
echo &"{state:>8} {miss:>10} {hit:>10} {miss+hit:>10} {hitRate:>5.2f}%"
|
|
|
|
let hitRate = float64(hits * 100) / (float64(hits + misses))
|
|
echo &" all {misses:>10} {hits:>10} {misses+hits:>10} {hitRate:>5.2f}%"
|
|
|
|
proc initImpl(
|
|
rdb: var RdbInst;
|
|
basePath: string;
|
|
opts: DbOptions;
|
|
dbOpts: DbOptionsRef,
|
|
cfOpts: ColFamilyOptionsRef;
|
|
guestCFs: openArray[ColFamilyDescriptor] = [];
|
|
): Result[seq[ColFamilyReadWrite],(AristoError,string)] =
|
|
## Database backend constructor
|
|
const initFailed = "RocksDB/init() failed"
|
|
|
|
rdb.basePath = basePath
|
|
|
|
# bytes -> entries based on overhead estimates
|
|
rdb.rdKeySize =
|
|
opts.rdbKeyCacheSize div (sizeof(VertexID) + sizeof(HashKey) + lruOverhead)
|
|
rdb.rdVtxSize =
|
|
opts.rdbVtxCacheSize div (sizeof(VertexID) + sizeof(default(VertexRef)[]) + lruOverhead)
|
|
|
|
if opts.rdbPrintStats:
|
|
let
|
|
ks = rdb.rdKeySize
|
|
vs = rdb.rdVtxSize
|
|
# TODO instead of dumping at exit, these stats could be logged or written
|
|
# to a file for better tracking over time - that said, this is mainly
|
|
# a debug utility at this point
|
|
addExitProc(proc() =
|
|
dumpCacheStats(ks, vs))
|
|
|
|
let
|
|
dataDir = rdb.dataDir
|
|
try:
|
|
dataDir.createDir
|
|
except OSError, IOError:
|
|
return err((RdbBeCantCreateDataDir, ""))
|
|
|
|
# Column familiy names to allocate when opening the database. This list
|
|
# might be extended below.
|
|
var useCFs = AristoCFs.mapIt($it).toHashSet
|
|
|
|
# The `guestCFs` list must not overwrite `AristoCFs` options
|
|
let guestCFs = guestCFs.filterIt(it.name notin useCFs)
|
|
|
|
# If the database exists already, check for missing column families and
|
|
# allocate them for opening. Otherwise rocksdb might reject the peristent
|
|
# database.
|
|
if (dataDir / "CURRENT").fileExists:
|
|
let hdCFs = dataDir.listColumnFamilies.valueOr:
|
|
raiseAssert initFailed & " cannot read existing CFs: " & error
|
|
# Update list of column families for opener.
|
|
useCFs = useCFs + hdCFs.toHashSet
|
|
|
|
# The `guestCFs` list might come with a different set of options. So it is
|
|
# temporarily removed from `useCFs` and will be re-added with appropriate
|
|
# options.
|
|
let guestCFq = @guestCFs
|
|
useCFs = useCFs - guestCFs.mapIt(it.name).toHashSet
|
|
|
|
# Finalise list of column families
|
|
let cfs = useCFs.toSeq.mapIt(it.initColFamilyDescriptor cfOpts) & guestCFq
|
|
|
|
# Open database for the extended family :)
|
|
let baseDb = openRocksDb(dataDir, dbOpts, columnFamilies=cfs).valueOr:
|
|
raiseAssert initFailed & " cannot create base descriptor: " & error
|
|
|
|
# Initialise column handlers (this stores implicitely `baseDb`)
|
|
rdb.admCol = baseDb.getColFamily($AdmCF).valueOr:
|
|
raiseAssert initFailed & " cannot initialise AdmCF descriptor: " & error
|
|
rdb.vtxCol = baseDb.getColFamily($VtxCF).valueOr:
|
|
raiseAssert initFailed & " cannot initialise VtxCF descriptor: " & error
|
|
rdb.keyCol = baseDb.getColFamily($KeyCF).valueOr:
|
|
raiseAssert initFailed & " cannot initialise KeyCF descriptor: " & error
|
|
|
|
ok(guestCFs.mapIt(baseDb.getColFamily(it.name).expect("loaded cf")))
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# Public constructor
|
|
# ------------------------------------------------------------------------------
|
|
|
|
proc init*(
|
|
rdb: var RdbInst;
|
|
basePath: string;
|
|
opts: DbOptions;
|
|
dbOpts: DbOptionsRef;
|
|
cfOpts: ColFamilyOptionsRef;
|
|
guestCFs: openArray[ColFamilyDescriptor];
|
|
): Result[seq[ColFamilyReadWrite],(AristoError,string)] =
|
|
## Temporarily define a guest CF list here.
|
|
rdb.initImpl(basePath, opts, dbOpts, cfOpts, guestCFs)
|
|
|
|
|
|
proc destroy*(rdb: var RdbInst; eradicate: bool) =
|
|
## Destructor
|
|
rdb.baseDb.close()
|
|
|
|
if eradicate:
|
|
try:
|
|
rdb.dataDir.removeDir
|
|
|
|
# Remove the base folder if it is empty
|
|
block done:
|
|
for w in rdb.baseDir.walkDirRec:
|
|
# Ignore backup files
|
|
if 0 < w.len and w[^1] != '~':
|
|
break done
|
|
rdb.baseDir.removeDir
|
|
|
|
except CatchableError:
|
|
discard
|
|
|
|
# ------------------------------------------------------------------------------
|
|
# End
|
|
# ------------------------------------------------------------------------------
|