Rename to HistoricalHashesAccumulator as per Portal spec (#2663)

This commit is contained in:
Kim De Mey 2024-09-27 22:25:26 +02:00 committed by GitHub
parent debb68b3a7
commit 438e183586
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
13 changed files with 79 additions and 61 deletions

View File

@ -19,7 +19,7 @@ type Era1DB* = ref object
## a linear history of pre-merge execution chain data. ## a linear history of pre-merge execution chain data.
path: string path: string
network: string network: string
accumulator: FinishedAccumulator accumulator: FinishedHistoricalHashesAccumulator
files: seq[Era1File] files: seq[Era1File]
proc getEra1File(db: Era1DB, era: Era1): Result[Era1File, string] = proc getEra1File(db: Era1DB, era: Era1): Result[Era1File, string] =
@ -51,7 +51,10 @@ proc getEra1File(db: Era1DB, era: Era1): Result[Era1File, string] =
ok(f) ok(f)
proc new*( proc new*(
T: type Era1DB, path: string, network: string, accumulator: FinishedAccumulator T: type Era1DB,
path: string,
network: string,
accumulator: FinishedHistoricalHashesAccumulator,
): Era1DB = ): Era1DB =
Era1DB(path: path, network: network, accumulator: accumulator) Era1DB(path: path, network: network, accumulator: accumulator)

View File

@ -19,11 +19,13 @@ export results
# Reading SSZ data from files # Reading SSZ data from files
proc readAccumulator*(file: string): Result[FinishedAccumulator, string] = proc readAccumulator*(
file: string
): Result[FinishedHistoricalHashesAccumulator, string] =
let encodedAccumulator = ?readAllFile(file).mapErr(toString) let encodedAccumulator = ?readAllFile(file).mapErr(toString)
try: try:
ok(SSZ.decode(encodedAccumulator, FinishedAccumulator)) ok(SSZ.decode(encodedAccumulator, FinishedHistoricalHashesAccumulator))
except SerializationError as e: except SerializationError as e:
err("Failed decoding accumulator: " & e.msg) err("Failed decoding accumulator: " & e.msg)

View File

@ -52,7 +52,7 @@ func toContentId(contentKey: ContentKeyByteList): ContentId =
readUintBE[256](idHash.data) readUintBE[256](idHash.data)
proc pruneDeprecatedAccumulatorRecords*( proc pruneDeprecatedAccumulatorRecords*(
accumulator: FinishedAccumulator, contentDB: ContentDB accumulator: FinishedHistoricalHashesAccumulator, contentDB: ContentDB
) = ) =
info "Pruning deprecated accumulator records" info "Pruning deprecated accumulator records"

View File

@ -27,18 +27,18 @@ const
type type
## BlockHeader types ## BlockHeader types
AccumulatorProof* = array[15, Digest] HistoricalHashesAccumulatorProof* = array[15, Digest]
BlockHeaderProofType* = enum BlockHeaderProofType* = enum
none = 0x00 # An SSZ Union None none = 0x00 # An SSZ Union None
accumulatorProof = 0x01 historicalHashesAccumulatorProof = 0x01
BlockHeaderProof* = object BlockHeaderProof* = object
case proofType*: BlockHeaderProofType case proofType*: BlockHeaderProofType
of none: of none:
discard discard
of accumulatorProof: of historicalHashesAccumulatorProof:
accumulatorProof*: AccumulatorProof historicalHashesAccumulatorProof*: HistoricalHashesAccumulatorProof
BlockHeaderWithProof* = object BlockHeaderWithProof* = object
header*: ByteList[MAX_HEADER_LENGTH] # RLP data header*: ByteList[MAX_HEADER_LENGTH] # RLP data
@ -68,8 +68,10 @@ type
ReceiptByteList* = ByteList[MAX_RECEIPT_LENGTH] # RLP data ReceiptByteList* = ByteList[MAX_RECEIPT_LENGTH] # RLP data
PortalReceipts* = List[ReceiptByteList, MAX_TRANSACTION_COUNT] PortalReceipts* = List[ReceiptByteList, MAX_TRANSACTION_COUNT]
func init*(T: type BlockHeaderProof, proof: AccumulatorProof): T = func init*(T: type BlockHeaderProof, proof: HistoricalHashesAccumulatorProof): T =
BlockHeaderProof(proofType: accumulatorProof, accumulatorProof: proof) BlockHeaderProof(
proofType: historicalHashesAccumulatorProof, historicalHashesAccumulatorProof: proof
)
func init*(T: type BlockHeaderProof): T = func init*(T: type BlockHeaderProof): T =
BlockHeaderProof(proofType: none) BlockHeaderProof(proofType: none)

View File

@ -31,7 +31,7 @@ type
portalProtocol*: PortalProtocol portalProtocol*: PortalProtocol
contentDB*: ContentDB contentDB*: ContentDB
contentQueue*: AsyncQueue[(Opt[NodeId], ContentKeysList, seq[seq[byte]])] contentQueue*: AsyncQueue[(Opt[NodeId], ContentKeysList, seq[seq[byte]])]
accumulator*: FinishedAccumulator accumulator*: FinishedHistoricalHashesAccumulator
historicalRoots*: HistoricalRoots historicalRoots*: HistoricalRoots
processContentLoop: Future[void] processContentLoop: Future[void]
statusLogLoop: Future[void] statusLogLoop: Future[void]
@ -636,7 +636,7 @@ proc new*(
baseProtocol: protocol.Protocol, baseProtocol: protocol.Protocol,
contentDB: ContentDB, contentDB: ContentDB,
streamManager: StreamManager, streamManager: StreamManager,
accumulator: FinishedAccumulator, accumulator: FinishedHistoricalHashesAccumulator,
historicalRoots: HistoricalRoots = loadHistoricalRoots(), historicalRoots: HistoricalRoots = loadHistoricalRoots(),
bootstrapRecords: openArray[Record] = [], bootstrapRecords: openArray[Record] = [],
portalConfig: PortalProtocolConfig = defaultPortalProtocolConfig, portalConfig: PortalProtocolConfig = defaultPortalProtocolConfig,

View File

@ -17,9 +17,8 @@ import
export ssz_serialization, merkleization, proofs, eth_types_rlp export ssz_serialization, merkleization, proofs, eth_types_rlp
# Header Accumulator, as per specification: # HistoricalHashesAccumulator, as per specification:
# https://github.com/ethereum/portal-network-specs/blob/master/history-network.md#the-header-accumulator # https://github.com/ethereum/portal-network-specs/blob/master/history/history-network.md#the-historical-hashes-accumulator
# But with the adjustment to finish the accumulator at merge point.
const const
EPOCH_SIZE* = 8192 # block roots per epoch record EPOCH_SIZE* = 8192 # block roots per epoch record
@ -60,18 +59,17 @@ type
# obviously much faster, so this second type is added for this usage. # obviously much faster, so this second type is added for this usage.
EpochRecordCached* = HashList[HeaderRecord, EPOCH_SIZE] EpochRecordCached* = HashList[HeaderRecord, EPOCH_SIZE]
# HistoricalHashesAccumulator HistoricalHashesAccumulator* = object
Accumulator* = object
historicalEpochs*: List[Bytes32, int(MAX_HISTORICAL_EPOCHS)] historicalEpochs*: List[Bytes32, int(MAX_HISTORICAL_EPOCHS)]
currentEpoch*: EpochRecord currentEpoch*: EpochRecord
# HistoricalHashesAccumulator in its final state # HistoricalHashesAccumulator in its final state
FinishedAccumulator* = object FinishedHistoricalHashesAccumulator* = object
historicalEpochs*: List[Bytes32, int(MAX_HISTORICAL_EPOCHS)] historicalEpochs*: List[Bytes32, int(MAX_HISTORICAL_EPOCHS)]
currentEpoch*: EpochRecord currentEpoch*: EpochRecord
func init*(T: type Accumulator): T = func init*(T: type HistoricalHashesAccumulator): T =
Accumulator( HistoricalHashesAccumulator(
historicalEpochs: List[Bytes32, int(MAX_HISTORICAL_EPOCHS)].init(@[]), historicalEpochs: List[Bytes32, int(MAX_HISTORICAL_EPOCHS)].init(@[]),
currentEpoch: EpochRecord.init(@[]), currentEpoch: EpochRecord.init(@[]),
) )
@ -81,7 +79,7 @@ func getEpochRecordRoot*(headerRecords: openArray[HeaderRecord]): Digest =
hash_tree_root(epochRecord) hash_tree_root(epochRecord)
func updateAccumulator*(a: var Accumulator, header: BlockHeader) = func updateAccumulator*(a: var HistoricalHashesAccumulator, header: BlockHeader) =
doAssert( doAssert(
header.number < mergeBlockNumber, "No post merge blocks for header accumulator" header.number < mergeBlockNumber, "No post merge blocks for header accumulator"
) )
@ -95,7 +93,7 @@ func updateAccumulator*(a: var Accumulator, header: BlockHeader) =
# TODO: It is a bit annoying to require an extra header + update call to # TODO: It is a bit annoying to require an extra header + update call to
# finish an epoch. However, if we were to move this after adding the # finish an epoch. However, if we were to move this after adding the
# `HeaderRecord`, there would be no way to get the current total difficulty, # `HeaderRecord`, there would be no way to get the current total difficulty,
# unless another field is introduced in the `Accumulator` object. # unless another field is introduced in the `HistoricalHashesAccumulator` object.
if a.currentEpoch.len() == EPOCH_SIZE: if a.currentEpoch.len() == EPOCH_SIZE:
let epochHash = hash_tree_root(a.currentEpoch) let epochHash = hash_tree_root(a.currentEpoch)
@ -110,17 +108,19 @@ func updateAccumulator*(a: var Accumulator, header: BlockHeader) =
let res = a.currentEpoch.add(headerRecord) let res = a.currentEpoch.add(headerRecord)
doAssert(res, "Can't fail because of currentEpoch length check") doAssert(res, "Can't fail because of currentEpoch length check")
func finishAccumulator*(a: var Accumulator): FinishedAccumulator = func finishAccumulator*(
a: var HistoricalHashesAccumulator
): FinishedHistoricalHashesAccumulator =
# doAssert(a.currentEpoch[^2].totalDifficulty < TERMINAL_TOTAL_DIFFICULTY) # doAssert(a.currentEpoch[^2].totalDifficulty < TERMINAL_TOTAL_DIFFICULTY)
# doAssert(a.currentEpoch[^1].totalDifficulty >= TERMINAL_TOTAL_DIFFICULTY) # doAssert(a.currentEpoch[^1].totalDifficulty >= TERMINAL_TOTAL_DIFFICULTY)
let epochHash = hash_tree_root(a.currentEpoch) let epochHash = hash_tree_root(a.currentEpoch)
doAssert(a.historicalEpochs.add(epochHash.data)) doAssert(a.historicalEpochs.add(epochHash.data))
FinishedAccumulator(historicalEpochs: a.historicalEpochs) FinishedHistoricalHashesAccumulator(historicalEpochs: a.historicalEpochs)
## Calls and helper calls for building header proofs and verifying headers ## Calls and helper calls for building header proofs and verifying headers
## against the Accumulator and the header proofs. ## against the HistoricalHashesAccumulator and the header proofs.
func getEpochIndex*(blockNumber: uint64): uint64 = func getEpochIndex*(blockNumber: uint64): uint64 =
blockNumber div EPOCH_SIZE blockNumber div EPOCH_SIZE
@ -144,7 +144,9 @@ func isPreMerge*(header: BlockHeader): bool =
isPreMerge(header.number) isPreMerge(header.number)
func verifyProof( func verifyProof(
a: FinishedAccumulator, header: BlockHeader, proof: openArray[Digest] a: FinishedHistoricalHashesAccumulator,
header: BlockHeader,
proof: openArray[Digest],
): bool = ): bool =
let let
epochIndex = getEpochIndex(header) epochIndex = getEpochIndex(header)
@ -153,13 +155,14 @@ func verifyProof(
leave = hash_tree_root(header.blockHash()) leave = hash_tree_root(header.blockHash())
headerRecordIndex = getHeaderRecordIndex(header, epochIndex) headerRecordIndex = getHeaderRecordIndex(header, epochIndex)
# TODO: Implement more generalized `get_generalized_index`
gIndex = GeneralizedIndex(EPOCH_SIZE * 2 * 2 + (headerRecordIndex * 2)) gIndex = GeneralizedIndex(EPOCH_SIZE * 2 * 2 + (headerRecordIndex * 2))
verify_merkle_multiproof(@[leave], proof, @[gIndex], epochRecordHash) verify_merkle_multiproof(@[leave], proof, @[gIndex], epochRecordHash)
func verifyAccumulatorProof*( func verifyAccumulatorProof*(
a: FinishedAccumulator, header: BlockHeader, proof: AccumulatorProof a: FinishedHistoricalHashesAccumulator,
header: BlockHeader,
proof: HistoricalHashesAccumulatorProof,
): Result[void, string] = ): Result[void, string] =
if header.isPreMerge(): if header.isPreMerge():
# Note: The proof is typed with correct depth, so no check on this is # Note: The proof is typed with correct depth, so no check on this is
@ -172,14 +175,14 @@ func verifyAccumulatorProof*(
err("Cannot verify post merge header with accumulator proof") err("Cannot verify post merge header with accumulator proof")
func verifyHeader*( func verifyHeader*(
a: FinishedAccumulator, header: BlockHeader, proof: BlockHeaderProof a: FinishedHistoricalHashesAccumulator, header: BlockHeader, proof: BlockHeaderProof
): Result[void, string] = ): Result[void, string] =
case proof.proofType case proof.proofType
of BlockHeaderProofType.accumulatorProof: of BlockHeaderProofType.historicalHashesAccumulatorProof:
a.verifyAccumulatorProof(header, proof.accumulatorProof) a.verifyAccumulatorProof(header, proof.historicalHashesAccumulatorProof)
of BlockHeaderProofType.none: of BlockHeaderProofType.none:
if header.isPreMerge(): if header.isPreMerge():
err("Pre merge header requires AccumulatorProof") err("Pre merge header requires HistoricalHashesAccumulatorProof")
else: else:
# TODO: # TODO:
# Currently there is no proof solution for verifying headers post-merge. # Currently there is no proof solution for verifying headers post-merge.
@ -191,17 +194,16 @@ func verifyHeader*(
func buildProof*( func buildProof*(
header: BlockHeader, epochRecord: EpochRecord | EpochRecordCached header: BlockHeader, epochRecord: EpochRecord | EpochRecordCached
): Result[AccumulatorProof, string] = ): Result[HistoricalHashesAccumulatorProof, string] =
doAssert(header.isPreMerge(), "Must be pre merge header") doAssert(header.isPreMerge(), "Must be pre merge header")
let let
epochIndex = getEpochIndex(header) epochIndex = getEpochIndex(header)
headerRecordIndex = getHeaderRecordIndex(header, epochIndex) headerRecordIndex = getHeaderRecordIndex(header, epochIndex)
# TODO: Implement more generalized `get_generalized_index`
gIndex = GeneralizedIndex(EPOCH_SIZE * 2 * 2 + (headerRecordIndex * 2)) gIndex = GeneralizedIndex(EPOCH_SIZE * 2 * 2 + (headerRecordIndex * 2))
var proof: AccumulatorProof var proof: HistoricalHashesAccumulatorProof
?epochRecord.build_proof(gIndex, proof) ?epochRecord.build_proof(gIndex, proof)
ok(proof) ok(proof)

View File

@ -52,9 +52,9 @@ const
historicalRootsSSZ* = slurp(portalConfigDir / "historical_roots.ssz") historicalRootsSSZ* = slurp(portalConfigDir / "historical_roots.ssz")
func loadAccumulator*(): FinishedAccumulator = func loadAccumulator*(): FinishedHistoricalHashesAccumulator =
try: try:
SSZ.decode(historicalHashesAccumulatorSSZ, FinishedAccumulator) SSZ.decode(historicalHashesAccumulatorSSZ, FinishedHistoricalHashesAccumulator)
except SerializationError as err: except SerializationError as err:
raiseAssert "Invalid baked-in accumulator: " & err.msg raiseAssert "Invalid baked-in accumulator: " & err.msg

View File

@ -57,7 +57,7 @@ suite "Historical Hashes Accumulator":
block: # Test invalid headers block: # Test invalid headers
# Post merge block number must fail (> than latest header in accumulator) # Post merge block number must fail (> than latest header in accumulator)
var proof: AccumulatorProof var proof: HistoricalHashesAccumulatorProof
let header = BlockHeader(number: mergeBlockNumber) let header = BlockHeader(number: mergeBlockNumber)
check verifyAccumulatorProof(accumulator, header, proof).isErr() check verifyAccumulatorProof(accumulator, header, proof).isErr()
@ -72,7 +72,7 @@ suite "Historical Hashes Accumulator":
check verifyAccumulatorProof(accumulator, header, proof.get()).isErr() check verifyAccumulatorProof(accumulator, header, proof.get()).isErr()
block: # Test invalid proofs block: # Test invalid proofs
var proof: AccumulatorProof var proof: HistoricalHashesAccumulatorProof
for i in headersToTest: for i in headersToTest:
check verifyAccumulatorProof(accumulator, headers[i], proof).isErr() check verifyAccumulatorProof(accumulator, headers[i], proof).isErr()

View File

@ -45,7 +45,7 @@ suite "Historical Hashes Accumulator Root":
let header = res.get() let header = res.get()
headers[header.number] = header headers[header.number] = header
var accumulator: Accumulator var accumulator: HistoricalHashesAccumulator
for i, hash in hashTreeRoots: for i, hash in hashTreeRoots:
updateAccumulator(accumulator, headers[i]) updateAccumulator(accumulator, headers[i])

View File

@ -24,7 +24,9 @@ type HistoryNode = ref object
historyNetwork*: HistoryNetwork historyNetwork*: HistoryNetwork
proc newHistoryNode( proc newHistoryNode(
rng: ref HmacDrbgContext, port: int, accumulator: FinishedAccumulator rng: ref HmacDrbgContext,
port: int,
accumulator: FinishedHistoricalHashesAccumulator,
): HistoryNode = ): HistoryNode =
let let
node = initDiscoveryNode(rng, PrivateKey.random(rng[]), localAddress(port)) node = initDiscoveryNode(rng, PrivateKey.random(rng[]), localAddress(port))

View File

@ -31,8 +31,10 @@ proc buildHeadersWithProof*(
ok(blockHeadersWithProof) ok(blockHeadersWithProof)
func buildAccumulator*(headers: seq[BlockHeader]): Result[FinishedAccumulator, string] = func buildAccumulator*(
var accumulator: Accumulator headers: seq[BlockHeader]
): Result[FinishedHistoricalHashesAccumulator, string] =
var accumulator: HistoricalHashesAccumulator
for header in headers: for header in headers:
updateAccumulator(accumulator, header) updateAccumulator(accumulator, header)
@ -43,8 +45,8 @@ func buildAccumulator*(headers: seq[BlockHeader]): Result[FinishedAccumulator, s
func buildAccumulatorData*( func buildAccumulatorData*(
headers: seq[BlockHeader] headers: seq[BlockHeader]
): Result[(FinishedAccumulator, seq[EpochRecord]), string] = ): Result[(FinishedHistoricalHashesAccumulator, seq[EpochRecord]), string] =
var accumulator: Accumulator var accumulator: HistoricalHashesAccumulator
var epochRecords: seq[EpochRecord] var epochRecords: seq[EpochRecord]
for header in headers: for header in headers:
updateAccumulator(accumulator, header) updateAccumulator(accumulator, header)
@ -61,7 +63,7 @@ func buildAccumulatorData*(
func buildProof*( func buildProof*(
header: BlockHeader, epochRecords: seq[EpochRecord] header: BlockHeader, epochRecords: seq[EpochRecord]
): Result[AccumulatorProof, string] = ): Result[HistoricalHashesAccumulatorProof, string] =
let epochIndex = getEpochIndex(header) let epochIndex = getEpochIndex(header)
doAssert(epochIndex < uint64(epochRecords.len())) doAssert(epochIndex < uint64(epochRecords.len()))
let epochRecord = epochRecords[epochIndex] let epochRecord = epochRecords[epochIndex]

View File

@ -119,7 +119,9 @@ proc newStateNode*(
"", uint32.high, RadiusConfig(kind: Dynamic), node.localNode.id, inMemory = true "", uint32.high, RadiusConfig(kind: Dynamic), node.localNode.id, inMemory = true
) )
sm = StreamManager.new(node) sm = StreamManager.new(node)
hn = HistoryNetwork.new(PortalNetwork.none, node, db, sm, FinishedAccumulator()) hn = HistoryNetwork.new(
PortalNetwork.none, node, db, sm, FinishedHistoricalHashesAccumulator()
)
sn = sn =
StateNetwork.new(PortalNetwork.none, node, db, sm, historyNetwork = Opt.some(hn)) StateNetwork.new(PortalNetwork.none, node, db, sm, historyNetwork = Opt.some(hn))

View File

@ -199,7 +199,7 @@ proc cmdExportEra1(config: ExporterConf) =
# error cases are fatal. But maybe we could throw proper errors still. # error cases are fatal. But maybe we could throw proper errors still.
var group = Era1Group.init(e2, startNumber).get() var group = Era1Group.init(e2, startNumber).get()
# Header records to build the accumulator root # Header records to build the HistoricalHashesAccumulator root
var headerRecords: seq[historical_hashes_accumulator.HeaderRecord] var headerRecords: seq[historical_hashes_accumulator.HeaderRecord]
for blockNumber in startNumber .. endNumber: for blockNumber in startNumber .. endNumber:
let blck = let blck =
@ -234,7 +234,7 @@ proc cmdExportEra1(config: ExporterConf) =
if completed: if completed:
let name = era1FileName("mainnet", era, accumulatorRoot) let name = era1FileName("mainnet", era, accumulatorRoot)
# We cannot check for the exact file any earlier as we need to know the # We cannot check for the exact file any earlier as we need to know the
# accumulator root. # HistoricalHashesAccumulator root.
# TODO: Could scan for file with era number in it. # TODO: Could scan for file with era number in it.
if isFile(name): if isFile(name):
info "Era1 file already exists", era, name info "Era1 file already exists", era, name
@ -381,7 +381,8 @@ when isMainModule:
# to build it. # to build it.
let accumulatorFile = dataDir / config.accumulatorFileName let accumulatorFile = dataDir / config.accumulatorFileName
if isFile(accumulatorFile): if isFile(accumulatorFile):
notice "Not building accumulator, file already exists", file = accumulatorFile notice "Not building HistoricalHashesAccumulator, file already exists",
file = accumulatorFile
quit 1 quit 1
# Lets verify if the necessary files exists before starting to build the # Lets verify if the necessary files exists before starting to build the
@ -394,8 +395,8 @@ when isMainModule:
proc buildAccumulator( proc buildAccumulator(
dataDir: string, writeEpochRecords = false dataDir: string, writeEpochRecords = false
): Result[FinishedAccumulator, string] = ): Result[FinishedHistoricalHashesAccumulator, string] =
var accumulator: Accumulator var accumulator: HistoricalHashesAccumulator
for i in 0 ..< preMergeEpochs: for i in 0 ..< preMergeEpochs:
let file = let file =
try: try:
@ -453,23 +454,24 @@ when isMainModule:
if blockHeader.number == mergeBlockNumber - 1: if blockHeader.number == mergeBlockNumber - 1:
let finishedAccumulator = finishAccumulator(accumulator) let finishedAccumulator = finishAccumulator(accumulator)
info "Updated last epoch, finished building master accumulator", info "Updated last epoch, finished building HistoricalHashesAccumulatorr",
epoch = i epoch = i
return ok(finishedAccumulator) return ok(finishedAccumulator)
else: else:
warn "Skipping record, not a block header", typ = toHex(header.typ) warn "Skipping record, not a block header", typ = toHex(header.typ)
err("Not enough headers provided to finish the accumulator") err("Not enough headers provided to finish the HistoricalHashesAccumulator")
let accumulatorRes = buildAccumulator(dataDir, config.writeEpochRecords) let accumulatorRes = buildAccumulator(dataDir, config.writeEpochRecords)
if accumulatorRes.isErr(): if accumulatorRes.isErr():
fatal "Could not build accumulator", error = accumulatorRes.error fatal "Could not build HistoricalHashesAccumulator",
error = accumulatorRes.error
quit 1 quit 1
let accumulator = accumulatorRes.get() let accumulator = accumulatorRes.get()
let res = io2.writeFile(accumulatorFile, SSZ.encode(accumulator)) let res = io2.writeFile(accumulatorFile, SSZ.encode(accumulator))
if res.isErr(): if res.isErr():
error "Failed writing accumulator to file", error "Failed writing HistoricalHashesAccumulator to file",
file = accumulatorFile, error = res.error file = accumulatorFile, error = res.error
quit 1 quit 1
else: else:
@ -480,16 +482,17 @@ when isMainModule:
let res = readAccumulator(file) let res = readAccumulator(file)
if res.isErr(): if res.isErr():
fatal "Failed reading accumulator from file", error = res.error, file fatal "Failed reading HistoricalHashesAccumulator from file",
error = res.error, file
quit 1 quit 1
let let
accumulator = res.get() accumulator = res.get()
accumulatorRoot = hash_tree_root(accumulator) accumulatorRoot = hash_tree_root(accumulator)
info "Accumulator decoded successfully", root = accumulatorRoot info "HistoricalHashesAccumulator decoded successfully", root = accumulatorRoot
echo "Master Accumulator:" echo "HistoricalHashesAccumulator:"
echo "-------------------" echo "-------------------"
echo &"Root: {accumulatorRoot}" echo &"Root: {accumulatorRoot}"
echo "" echo ""