2022-09-09 19:21:48 +00:00
|
|
|
# # Nimbus - Portal Network
|
2024-02-15 15:49:22 +00:00
|
|
|
# # Copyright (c) 2022-2024 Status Research & Development GmbH
|
2022-09-09 19:21:48 +00:00
|
|
|
# # Licensed and distributed under either of
|
|
|
|
# # * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# # * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
|
# # at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
2023-01-31 12:38:08 +00:00
|
|
|
{.push raises: [].}
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
import
|
2022-10-18 11:07:32 +00:00
|
|
|
std/[strformat, os],
|
2024-02-28 17:31:45 +00:00
|
|
|
results,
|
|
|
|
chronos,
|
|
|
|
chronicles,
|
|
|
|
eth/common/eth_types,
|
|
|
|
eth/rlp,
|
2022-09-09 19:21:48 +00:00
|
|
|
../network/wire/portal_protocol,
|
2023-02-16 11:40:07 +00:00
|
|
|
../network/history/[history_content, history_network, accumulator],
|
2024-02-15 15:49:22 +00:00
|
|
|
"."/[era1, history_data_json_store, history_data_ssz_e2s]
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
export results
|
|
|
|
|
|
|
|
### Helper calls to seed the local database and/or the network
|
|
|
|
|
|
|
|
proc historyStore*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, dataFile: string, verify = false
|
|
|
|
): Result[void, string] =
|
|
|
|
let blockData = ?readJsonType(dataFile, BlockDataTable)
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
for b in blocks(blockData, verify):
|
|
|
|
for value in b:
|
2022-11-08 17:31:45 +00:00
|
|
|
let encKey = history_content.encode(value[0])
|
2022-09-09 19:21:48 +00:00
|
|
|
# Note: This is the slowest part due to the hashing that takes place.
|
2022-11-08 17:31:45 +00:00
|
|
|
p.storeContent(encKey, history_content.toContentId(encKey), value[1])
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
ok()
|
|
|
|
|
|
|
|
proc propagateEpochAccumulator*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, file: string
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2022-09-09 19:21:48 +00:00
|
|
|
## Propagate a specific epoch accumulator into the network.
|
2022-10-18 11:07:32 +00:00
|
|
|
## file holds the SSZ serialized epoch accumulator.
|
|
|
|
let epochAccumulatorRes = readEpochAccumulator(file)
|
2022-09-09 19:21:48 +00:00
|
|
|
if epochAccumulatorRes.isErr():
|
|
|
|
return err(epochAccumulatorRes.error)
|
|
|
|
else:
|
|
|
|
let
|
|
|
|
accumulator = epochAccumulatorRes.get()
|
|
|
|
rootHash = accumulator.hash_tree_root()
|
|
|
|
key = ContentKey(
|
|
|
|
contentType: epochAccumulator,
|
2024-02-28 17:31:45 +00:00
|
|
|
epochAccumulatorKey: EpochAccumulatorKey(epochHash: rootHash),
|
|
|
|
)
|
2022-11-08 17:31:45 +00:00
|
|
|
encKey = history_content.encode(key)
|
2022-10-18 11:07:32 +00:00
|
|
|
# Note: The file actually holds the SSZ encoded accumulator, but we need
|
|
|
|
# to decode as we need the root for the content key.
|
|
|
|
encodedAccumulator = SSZ.encode(accumulator)
|
2022-12-16 16:47:52 +00:00
|
|
|
info "Gossiping epoch accumulator", rootHash, contentKey = encKey
|
2022-10-18 11:07:32 +00:00
|
|
|
|
2024-02-28 17:31:45 +00:00
|
|
|
p.storeContent(encKey, history_content.toContentId(encKey), encodedAccumulator)
|
2022-10-11 10:10:54 +00:00
|
|
|
discard await p.neighborhoodGossip(
|
2024-02-28 17:31:45 +00:00
|
|
|
Opt.none(NodeId), ContentKeysList(@[encKey]), @[encodedAccumulator]
|
|
|
|
)
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
return ok()
|
|
|
|
|
2022-10-18 11:07:32 +00:00
|
|
|
proc propagateEpochAccumulators*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, path: string
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2022-10-18 11:07:32 +00:00
|
|
|
## Propagate all epoch accumulators created when building the accumulator
|
|
|
|
## from the block headers.
|
|
|
|
## path is a directory that holds all SSZ encoded epoch accumulator files.
|
2024-02-28 17:31:45 +00:00
|
|
|
for i in 0 ..< preMergeEpochs:
|
2022-10-18 11:07:32 +00:00
|
|
|
let file =
|
2024-02-28 17:31:45 +00:00
|
|
|
try:
|
|
|
|
path / &"mainnet-epoch-accumulator-{i.uint64:05}.ssz"
|
|
|
|
except ValueError as e:
|
|
|
|
raiseAssert e.msg
|
2022-10-18 11:07:32 +00:00
|
|
|
|
|
|
|
let res = await p.propagateEpochAccumulator(file)
|
|
|
|
if res.isErr():
|
|
|
|
return err(res.error)
|
|
|
|
|
|
|
|
return ok()
|
|
|
|
|
2022-09-09 19:21:48 +00:00
|
|
|
proc historyPropagate*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, dataFile: string, verify = false
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2022-09-09 19:21:48 +00:00
|
|
|
const concurrentGossips = 20
|
|
|
|
|
|
|
|
var gossipQueue =
|
2023-09-04 10:21:01 +00:00
|
|
|
newAsyncQueue[(Opt[NodeId], ContentKeysList, seq[byte])](concurrentGossips)
|
2022-09-09 19:21:48 +00:00
|
|
|
var gossipWorkers: seq[Future[void]]
|
|
|
|
|
|
|
|
proc gossipWorker(p: PortalProtocol) {.async.} =
|
|
|
|
while true:
|
2023-09-04 10:21:01 +00:00
|
|
|
let (srcNodeId, keys, content) = await gossipQueue.popFirst()
|
2022-09-09 19:21:48 +00:00
|
|
|
|
2023-09-04 10:21:01 +00:00
|
|
|
discard await p.neighborhoodGossip(srcNodeId, keys, @[content])
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
for i in 0 ..< concurrentGossips:
|
|
|
|
gossipWorkers.add(gossipWorker(p))
|
|
|
|
|
|
|
|
let blockData = readJsonType(dataFile, BlockDataTable)
|
|
|
|
if blockData.isOk():
|
|
|
|
for b in blocks(blockData.get(), verify):
|
2022-09-29 06:42:54 +00:00
|
|
|
for i, value in b:
|
|
|
|
if i == 0:
|
|
|
|
# Note: Skipping propagation of headers here as they should be offered
|
|
|
|
# separately to be certain that bodies and receipts can be verified.
|
|
|
|
# TODO: Rename this chain of calls to be more clear about this and
|
|
|
|
# adjust the interator call.
|
|
|
|
continue
|
2022-09-09 19:21:48 +00:00
|
|
|
# Only sending non empty data, e.g. empty receipts are not send
|
|
|
|
# TODO: Could do a similar thing for a combination of empty
|
|
|
|
# txs and empty uncles, as then the serialization is always the same.
|
|
|
|
if value[1].len() > 0:
|
|
|
|
info "Seeding block content into the network", contentKey = value[0]
|
|
|
|
# Note: This is the slowest part due to the hashing that takes place.
|
2022-11-08 17:31:45 +00:00
|
|
|
let
|
|
|
|
encKey = history_content.encode(value[0])
|
|
|
|
contentId = history_content.toContentId(encKey)
|
|
|
|
p.storeContent(encKey, contentId, value[1])
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
await gossipQueue.addLast(
|
2024-02-28 17:31:45 +00:00
|
|
|
(Opt.none(NodeId), ContentKeysList(@[encode(value[0])]), value[1])
|
|
|
|
)
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
return ok()
|
|
|
|
else:
|
|
|
|
return err(blockData.error)
|
|
|
|
|
|
|
|
proc historyPropagateBlock*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, dataFile: string, blockHash: string, verify = false
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2022-09-09 19:21:48 +00:00
|
|
|
let blockDataTable = readJsonType(dataFile, BlockDataTable)
|
|
|
|
|
|
|
|
if blockDataTable.isOk():
|
|
|
|
let b =
|
|
|
|
try:
|
|
|
|
blockDataTable.get()[blockHash]
|
|
|
|
except KeyError:
|
|
|
|
return err("Block hash not found in block data file")
|
|
|
|
|
|
|
|
let blockDataRes = readBlockData(blockHash, b)
|
|
|
|
if blockDataRes.isErr:
|
|
|
|
return err(blockDataRes.error)
|
|
|
|
|
|
|
|
let blockData = blockDataRes.get()
|
|
|
|
|
|
|
|
for value in blockData:
|
|
|
|
info "Seeding block content into the network", contentKey = value[0]
|
2022-11-08 17:31:45 +00:00
|
|
|
let
|
|
|
|
encKey = history_content.encode(value[0])
|
|
|
|
contentId = history_content.toContentId(encKey)
|
|
|
|
p.storeContent(encKey, contentId, value[1])
|
2022-09-09 19:21:48 +00:00
|
|
|
|
2024-02-28 17:31:45 +00:00
|
|
|
discard await p.neighborhoodGossip(
|
|
|
|
Opt.none(NodeId), ContentKeysList(@[encode(value[0])]), @[value[1]]
|
|
|
|
)
|
2022-09-09 19:21:48 +00:00
|
|
|
|
|
|
|
return ok()
|
|
|
|
else:
|
|
|
|
return err(blockDataTable.error)
|
2022-09-29 06:42:54 +00:00
|
|
|
|
2022-11-04 08:27:01 +00:00
|
|
|
proc historyPropagateHeadersWithProof*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, epochHeadersFile: string, epochAccumulatorFile: string
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2022-11-04 08:27:01 +00:00
|
|
|
let res = readBlockHeaders(epochHeadersFile)
|
|
|
|
if res.isErr():
|
|
|
|
return err(res.error)
|
|
|
|
|
|
|
|
let blockHeaders = res.get()
|
|
|
|
|
|
|
|
let epochAccumulatorRes = readEpochAccumulatorCached(epochAccumulatorFile)
|
|
|
|
if epochAccumulatorRes.isErr():
|
|
|
|
return err(res.error)
|
|
|
|
|
|
|
|
let epochAccumulator = epochAccumulatorRes.get()
|
|
|
|
for header in blockHeaders:
|
|
|
|
if header.isPreMerge():
|
|
|
|
let headerWithProof = buildHeaderWithProof(header, epochAccumulator)
|
|
|
|
if headerWithProof.isErr:
|
|
|
|
return err(headerWithProof.error)
|
|
|
|
|
|
|
|
let
|
|
|
|
content = headerWithProof.get()
|
|
|
|
contentKey = ContentKey(
|
2023-01-20 22:04:58 +00:00
|
|
|
contentType: blockHeader,
|
2024-02-28 17:31:45 +00:00
|
|
|
blockHeaderKey: BlockKey(blockHash: header.blockHash()),
|
|
|
|
)
|
2022-11-08 17:31:45 +00:00
|
|
|
encKey = history_content.encode(contentKey)
|
|
|
|
contentId = history_content.toContentId(encKey)
|
2022-11-04 08:27:01 +00:00
|
|
|
encodedContent = SSZ.encode(content)
|
|
|
|
|
2022-11-08 17:31:45 +00:00
|
|
|
p.storeContent(encKey, contentId, encodedContent)
|
2022-11-04 08:27:01 +00:00
|
|
|
|
|
|
|
let keys = ContentKeysList(@[encode(contentKey)])
|
2023-09-04 10:21:01 +00:00
|
|
|
discard await p.neighborhoodGossip(Opt.none(NodeId), keys, @[encodedContent])
|
2022-11-04 08:27:01 +00:00
|
|
|
|
|
|
|
return ok()
|
|
|
|
|
|
|
|
proc historyPropagateHeadersWithProof*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, dataDir: string
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
|
|
|
for i in 0 ..< preMergeEpochs:
|
2022-11-04 08:27:01 +00:00
|
|
|
let
|
|
|
|
epochHeadersfile =
|
2024-02-28 17:31:45 +00:00
|
|
|
try:
|
|
|
|
dataDir / &"mainnet-headers-epoch-{i.uint64:05}.e2s"
|
|
|
|
except ValueError as e:
|
|
|
|
raiseAssert e.msg
|
2022-11-04 08:27:01 +00:00
|
|
|
epochAccumulatorFile =
|
2024-02-28 17:31:45 +00:00
|
|
|
try:
|
|
|
|
dataDir / &"mainnet-epoch-accumulator-{i.uint64:05}.ssz"
|
|
|
|
except ValueError as e:
|
|
|
|
raiseAssert e.msg
|
2022-11-04 08:27:01 +00:00
|
|
|
|
2024-02-28 17:31:45 +00:00
|
|
|
let res =
|
|
|
|
await p.historyPropagateHeadersWithProof(epochHeadersfile, epochAccumulatorFile)
|
2022-11-04 08:27:01 +00:00
|
|
|
if res.isOk():
|
|
|
|
info "Finished gossiping 1 epoch of headers with proof", i
|
|
|
|
else:
|
|
|
|
return err(res.error)
|
|
|
|
|
|
|
|
return ok()
|
|
|
|
|
2022-09-29 06:42:54 +00:00
|
|
|
proc historyPropagateHeaders*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, dataFile: string, verify = false
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2022-09-29 06:42:54 +00:00
|
|
|
# TODO: Should perhaps be integrated with `historyPropagate` call.
|
|
|
|
const concurrentGossips = 20
|
|
|
|
|
2024-02-28 17:31:45 +00:00
|
|
|
var gossipQueue = newAsyncQueue[(ContentKeysList, seq[byte])](concurrentGossips)
|
2022-09-29 06:42:54 +00:00
|
|
|
var gossipWorkers: seq[Future[void]]
|
|
|
|
|
|
|
|
proc gossipWorker(p: PortalProtocol) {.async.} =
|
|
|
|
while true:
|
|
|
|
let (keys, content) = await gossipQueue.popFirst()
|
|
|
|
|
2023-09-04 10:21:01 +00:00
|
|
|
discard await p.neighborhoodGossip(Opt.none(NodeId), keys, @[content])
|
2022-09-29 06:42:54 +00:00
|
|
|
|
|
|
|
for i in 0 ..< concurrentGossips:
|
|
|
|
gossipWorkers.add(gossipWorker(p))
|
|
|
|
|
|
|
|
let blockData = readJsonType(dataFile, BlockDataTable)
|
|
|
|
if blockData.isOk():
|
|
|
|
for header in headers(blockData.get(), verify):
|
|
|
|
info "Seeding header content into the network", contentKey = header[0]
|
2022-11-08 17:31:45 +00:00
|
|
|
let
|
|
|
|
encKey = history_content.encode(header[0])
|
|
|
|
contentId = history_content.toContentId(encKey)
|
|
|
|
p.storeContent(encKey, contentId, header[1])
|
2022-09-29 06:42:54 +00:00
|
|
|
|
2024-02-28 17:31:45 +00:00
|
|
|
await gossipQueue.addLast((ContentKeysList(@[encode(header[0])]), header[1]))
|
2022-09-29 06:42:54 +00:00
|
|
|
|
|
|
|
return ok()
|
|
|
|
else:
|
|
|
|
return err(blockData.error)
|
2024-02-15 15:49:22 +00:00
|
|
|
|
|
|
|
##
|
|
|
|
## Era1 based iterators that encode to Portal content
|
|
|
|
##
|
|
|
|
|
|
|
|
# Note: these iterators + the era1 iterators will assert on error. These asserts
|
|
|
|
# would indicate corrupt/invalid era1 files. We might want to instead break,
|
|
|
|
# raise an exception or return a Result type instead, but the latter does not
|
|
|
|
# have great support for usage in iterators.
|
|
|
|
|
|
|
|
iterator headersWithProof*(
|
|
|
|
f: Era1File, epochAccumulator: EpochAccumulatorCached
|
2024-02-28 17:31:45 +00:00
|
|
|
): (ByteList, seq[byte]) =
|
2024-02-15 15:49:22 +00:00
|
|
|
for blockHeader in f.era1BlockHeaders:
|
|
|
|
doAssert blockHeader.isPreMerge()
|
|
|
|
|
|
|
|
let
|
|
|
|
contentKey = ContentKey(
|
|
|
|
contentType: blockHeader,
|
2024-02-28 17:31:45 +00:00
|
|
|
blockHeaderKey: BlockKey(blockHash: blockHeader.blockHash()),
|
2024-02-15 15:49:22 +00:00
|
|
|
).encode()
|
|
|
|
|
|
|
|
headerWithProof = buildHeaderWithProof(blockHeader, epochAccumulator).valueOr:
|
2024-06-14 07:31:08 +00:00
|
|
|
raiseAssert "Failed to build header with proof: " & $blockHeader.number
|
2024-02-15 15:49:22 +00:00
|
|
|
|
|
|
|
contentValue = SSZ.encode(headerWithProof)
|
|
|
|
|
|
|
|
yield (contentKey, contentValue)
|
|
|
|
|
|
|
|
iterator blockContent*(f: Era1File): (ByteList, seq[byte]) =
|
|
|
|
for (header, body, receipts, _) in f.era1BlockTuples:
|
|
|
|
let blockHash = header.blockHash()
|
|
|
|
|
|
|
|
block: # block body
|
|
|
|
let
|
|
|
|
contentKey = ContentKey(
|
2024-02-28 17:31:45 +00:00
|
|
|
contentType: blockBody, blockBodyKey: BlockKey(blockHash: blockHash)
|
2024-02-15 15:49:22 +00:00
|
|
|
).encode()
|
|
|
|
|
|
|
|
contentValue = encode(body)
|
|
|
|
|
|
|
|
yield (contentKey, contentValue)
|
|
|
|
|
|
|
|
block: # receipts
|
|
|
|
let
|
|
|
|
contentKey = ContentKey(
|
2024-02-28 17:31:45 +00:00
|
|
|
contentType: receipts, receiptsKey: BlockKey(blockHash: blockHash)
|
2024-02-15 15:49:22 +00:00
|
|
|
).encode()
|
|
|
|
|
|
|
|
contentValue = encode(receipts)
|
|
|
|
|
|
|
|
yield (contentKey, contentValue)
|
|
|
|
|
|
|
|
##
|
|
|
|
## Era1 based Gossip calls
|
|
|
|
##
|
|
|
|
|
|
|
|
proc historyGossipHeadersWithProof*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol,
|
|
|
|
era1File: string,
|
|
|
|
epochAccumulatorFile: Opt[string],
|
|
|
|
verifyEra = false,
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2024-02-15 15:49:22 +00:00
|
|
|
let f = ?Era1File.open(era1File)
|
|
|
|
|
|
|
|
if verifyEra:
|
|
|
|
let _ = ?f.verify()
|
|
|
|
|
|
|
|
# Note: building the accumulator takes about 150ms vs 10ms for reading it,
|
|
|
|
# so it is probably not really worth using the read version considering the
|
|
|
|
# UX hassle it adds to provide the accumulator ssz files.
|
|
|
|
let epochAccumulator =
|
|
|
|
if epochAccumulatorFile.isNone:
|
|
|
|
?f.buildAccumulator()
|
|
|
|
else:
|
|
|
|
?readEpochAccumulatorCached(epochAccumulatorFile.get())
|
|
|
|
|
|
|
|
for (contentKey, contentValue) in f.headersWithProof(epochAccumulator):
|
|
|
|
let peers = await p.neighborhoodGossip(
|
2024-02-28 17:31:45 +00:00
|
|
|
Opt.none(NodeId), ContentKeysList(@[contentKey]), @[contentValue]
|
|
|
|
)
|
2024-02-15 15:49:22 +00:00
|
|
|
info "Gossiped block header", contentKey, peers
|
|
|
|
|
|
|
|
ok()
|
|
|
|
|
|
|
|
proc historyGossipBlockContent*(
|
2024-02-28 17:31:45 +00:00
|
|
|
p: PortalProtocol, era1File: string, verifyEra = false
|
|
|
|
): Future[Result[void, string]] {.async.} =
|
2024-02-15 15:49:22 +00:00
|
|
|
let f = ?Era1File.open(era1File)
|
|
|
|
|
|
|
|
if verifyEra:
|
|
|
|
let _ = ?f.verify()
|
|
|
|
|
|
|
|
for (contentKey, contentValue) in f.blockContent():
|
|
|
|
let peers = await p.neighborhoodGossip(
|
2024-02-28 17:31:45 +00:00
|
|
|
Opt.none(NodeId), ContentKeysList(@[contentKey]), @[contentValue]
|
|
|
|
)
|
2024-02-15 15:49:22 +00:00
|
|
|
info "Gossiped block content", contentKey, peers
|
|
|
|
|
|
|
|
ok()
|