2021-10-09 11:22:03 +00:00
|
|
|
# Nimbus
|
2022-04-13 05:56:01 +00:00
|
|
|
# Copyright (c) 2021-2022 Status Research & Development GmbH
|
2021-10-09 11:22:03 +00:00
|
|
|
# Licensed and distributed under either of
|
|
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
2022-03-02 14:29:12 +00:00
|
|
|
{.push raises: [Defect].}
|
|
|
|
|
2021-10-09 11:22:03 +00:00
|
|
|
import
|
2022-08-01 19:00:21 +00:00
|
|
|
std/[options, tables],
|
2022-09-03 18:15:35 +00:00
|
|
|
stew/results, chronos, chronicles,
|
|
|
|
eth/[common/eth_types_rlp, rlp, trie, trie/db],
|
2022-01-06 08:06:05 +00:00
|
|
|
eth/p2p/discoveryv5/[protocol, enr],
|
2021-10-09 11:22:03 +00:00
|
|
|
../../content_db,
|
2022-07-01 19:51:51 +00:00
|
|
|
../../../nimbus/constants,
|
2022-01-18 08:01:22 +00:00
|
|
|
../wire/[portal_protocol, portal_stream, portal_protocol_config],
|
2022-09-09 19:21:48 +00:00
|
|
|
"."/[history_content, accumulator]
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-04-13 05:56:01 +00:00
|
|
|
logScope:
|
|
|
|
topics = "portal_hist"
|
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
export accumulator
|
|
|
|
|
|
|
|
# TODO: To currently verify if content is from the canonical chain it is
|
|
|
|
# required to download the right epoch accunulator, which is ~0.5 MB. This is
|
|
|
|
# too much, at least for the local testnet tests. This needs to be improved
|
|
|
|
# by adding the proofs to the block header content. Another independent
|
|
|
|
# improvement would be to have a content cache (LRU or so). The latter would
|
|
|
|
# probably help mostly for the local testnet tests.
|
|
|
|
# For now, we disable this verification default until further improvements are
|
|
|
|
# made.
|
|
|
|
const canonicalVerify* {.booldefine.} = false
|
|
|
|
|
2021-10-09 11:22:03 +00:00
|
|
|
const
|
2021-12-08 10:54:22 +00:00
|
|
|
historyProtocolId* = [byte 0x50, 0x0B]
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-03-18 12:06:57 +00:00
|
|
|
type
|
2022-03-02 14:29:12 +00:00
|
|
|
HistoryNetwork* = ref object
|
|
|
|
portalProtocol*: PortalProtocol
|
|
|
|
contentDB*: ContentDB
|
2022-08-17 07:32:06 +00:00
|
|
|
contentQueue*: AsyncQueue[(ContentKeysList, seq[seq[byte]])]
|
2022-07-11 14:29:16 +00:00
|
|
|
processContentLoop: Future[void]
|
2022-03-02 14:29:12 +00:00
|
|
|
|
|
|
|
Block* = (BlockHeader, BlockBody)
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func toContentIdHandler(contentKey: ByteList): Option[ContentId] =
|
2022-01-06 08:06:05 +00:00
|
|
|
some(toContentId(contentKey))
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-03-02 14:29:12 +00:00
|
|
|
func encodeKey(k: ContentKey): (ByteList, ContentId) =
|
|
|
|
let keyEncoded = encode(k)
|
|
|
|
return (keyEncoded, toContentId(keyEncoded))
|
|
|
|
|
2022-04-01 16:01:50 +00:00
|
|
|
func getEncodedKeyForContent(
|
2022-09-28 07:09:38 +00:00
|
|
|
cType: ContentType, hash: BlockHash):
|
2022-04-01 16:01:50 +00:00
|
|
|
(ByteList, ContentId) =
|
2022-09-28 07:09:38 +00:00
|
|
|
let contentKeyType = BlockKey(blockHash: hash)
|
2022-03-18 12:06:57 +00:00
|
|
|
|
|
|
|
let contentKey =
|
2022-03-02 14:29:12 +00:00
|
|
|
case cType
|
|
|
|
of blockHeader:
|
|
|
|
ContentKey(contentType: cType, blockHeaderKey: contentKeyType)
|
|
|
|
of blockBody:
|
|
|
|
ContentKey(contentType: cType, blockBodyKey: contentKeyType)
|
|
|
|
of receipts:
|
|
|
|
ContentKey(contentType: cType, receiptsKey: contentKeyType)
|
2022-06-14 21:38:34 +00:00
|
|
|
of epochAccumulator:
|
|
|
|
raiseAssert("Not implemented")
|
|
|
|
of masterAccumulator:
|
|
|
|
raiseAssert("Not implemented")
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-03-02 14:29:12 +00:00
|
|
|
return encodeKey(contentKey)
|
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func decodeRlp*(bytes: openArray[byte], T: type): Result[T, string] =
|
2022-03-02 14:29:12 +00:00
|
|
|
try:
|
2022-07-01 19:51:51 +00:00
|
|
|
ok(rlp.decode(bytes, T))
|
2022-06-10 10:24:53 +00:00
|
|
|
except RlpError as e:
|
|
|
|
err(e.msg)
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
## Calls to go from SSZ decoded types to RLP fully decoded types
|
2022-06-14 21:38:34 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func fromPortalBlockBody(
|
|
|
|
T: type BlockBody, body: BlockBodySSZ): Result[T, string] =
|
|
|
|
## Get the full decoded BlockBody from the SSZ-decoded `PortalBlockBody`.
|
|
|
|
try:
|
|
|
|
var transactions: seq[Transaction]
|
|
|
|
for tx in body.transactions:
|
|
|
|
transactions.add(rlp.decode(tx.asSeq(), Transaction))
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
let uncles = rlp.decode(body.uncles.asSeq(), seq[BlockHeader])
|
2022-06-14 21:38:34 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
ok(BlockBody(transactions: transactions, uncles: uncles))
|
|
|
|
except RlpError as e:
|
|
|
|
err("RLP decoding failed: " & e.msg)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func fromReceipts(
|
|
|
|
T: type seq[Receipt], receipts: ReceiptsSSZ): Result[T, string] =
|
|
|
|
## Get the full decoded seq[Receipt] from the SSZ-decoded `Receipts`.
|
2022-03-02 14:29:12 +00:00
|
|
|
try:
|
2022-07-01 19:51:51 +00:00
|
|
|
var res: seq[Receipt]
|
|
|
|
for receipt in receipts:
|
|
|
|
res.add(rlp.decode(receipt.asSeq(), Receipt))
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
ok(res)
|
2022-06-10 10:24:53 +00:00
|
|
|
except RlpError as e:
|
2022-07-01 19:51:51 +00:00
|
|
|
err("RLP decoding failed: " & e.msg)
|
2022-06-14 21:38:34 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
## Calls to encode Block types to the SSZ types.
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func fromBlockBody(T: type BlockBodySSZ, body: BlockBody): T =
|
|
|
|
var transactions: Transactions
|
|
|
|
for tx in body.transactions:
|
|
|
|
discard transactions.add(TransactionByteList(rlp.encode(tx)))
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
let uncles = Uncles(rlp.encode(body.uncles))
|
2022-06-14 21:38:34 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
BlockBodySSZ(transactions: transactions, uncles: uncles)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func fromReceipts(T: type ReceiptsSSZ, receipts: seq[Receipt]): T =
|
|
|
|
var receiptsSSZ: ReceiptsSSZ
|
|
|
|
for receipt in receipts:
|
|
|
|
discard receiptsSSZ.add(ReceiptByteList(rlp.encode(receipt)))
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
receiptsSSZ
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
func encode*(blockBody: BlockBody): seq[byte] =
|
|
|
|
let portalBlockBody = BlockBodySSZ.fromBlockBody(blockBody)
|
2022-06-14 21:38:34 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
SSZ.encode(portalBlockBody)
|
|
|
|
|
|
|
|
func encode*(receipts: seq[Receipt]): seq[byte] =
|
|
|
|
let portalReceipts = ReceiptsSSZ.fromReceipts(receipts)
|
|
|
|
|
|
|
|
SSZ.encode(portalReceipts)
|
|
|
|
|
|
|
|
## Calls and helper calls to do validation of block header, body and receipts
|
|
|
|
# TODO: Failures on validation and perhaps deserialisation should be punished
|
|
|
|
# for if/when peer scoring/banning is added.
|
|
|
|
|
|
|
|
proc calcRootHash(items: Transactions | ReceiptsSSZ): Hash256 =
|
|
|
|
var tr = initHexaryTrie(newMemoryDB())
|
|
|
|
for i, t in items:
|
|
|
|
try:
|
|
|
|
tr.put(rlp.encode(i), t.asSeq())
|
|
|
|
except RlpError as e:
|
|
|
|
# TODO: Investigate this RlpError as it doesn't sound like this is
|
|
|
|
# something that can actually occur.
|
|
|
|
raiseAssert(e.msg)
|
|
|
|
|
|
|
|
return tr.rootHash
|
|
|
|
|
|
|
|
template calcTxsRoot*(transactions: Transactions): Hash256 =
|
|
|
|
calcRootHash(transactions)
|
|
|
|
|
|
|
|
template calcReceiptsRoot*(receipts: ReceiptsSSZ): Hash256 =
|
|
|
|
calcRootHash(receipts)
|
|
|
|
|
|
|
|
func validateBlockHeaderBytes*(
|
|
|
|
bytes: openArray[byte], hash: BlockHash): Result[BlockHeader, string] =
|
|
|
|
|
|
|
|
let header = ? decodeRlp(bytes, BlockHeader)
|
|
|
|
|
|
|
|
if not (header.blockHash() == hash):
|
|
|
|
err("Block header hash does not match")
|
|
|
|
else:
|
|
|
|
ok(header)
|
|
|
|
|
|
|
|
proc validateBlockBody(
|
|
|
|
body: BlockBodySSZ, txsRoot, ommersHash: KeccakHash):
|
|
|
|
Result[void, string] =
|
|
|
|
## Validate the block body against the txRoot amd ommersHash from the header.
|
2022-09-03 18:15:35 +00:00
|
|
|
let calculatedOmmersHash = keccakHash(body.uncles.asSeq())
|
2022-07-01 19:51:51 +00:00
|
|
|
if calculatedOmmersHash != ommersHash:
|
|
|
|
return err("Invalid ommers hash")
|
|
|
|
|
|
|
|
let calculatedTxsRoot = calcTxsRoot(body.transactions)
|
|
|
|
if calculatedTxsRoot != txsRoot:
|
|
|
|
return err("Invalid transactions root")
|
|
|
|
|
|
|
|
ok()
|
|
|
|
|
|
|
|
proc validateBlockBodyBytes*(
|
|
|
|
bytes: openArray[byte], txRoot, ommersHash: KeccakHash):
|
|
|
|
Result[BlockBody, string] =
|
|
|
|
## Fully decode the SSZ Block Body and validate it against the header.
|
|
|
|
let body =
|
|
|
|
try:
|
|
|
|
SSZ.decode(bytes, BlockBodySSZ)
|
|
|
|
except SszError as e:
|
2022-07-05 12:42:55 +00:00
|
|
|
return err("Failed to decode block body: " & e.msg)
|
2022-07-01 19:51:51 +00:00
|
|
|
|
|
|
|
? validateBlockBody(body, txRoot, ommersHash)
|
|
|
|
|
|
|
|
BlockBody.fromPortalBlockBody(body)
|
|
|
|
|
|
|
|
proc validateReceipts(
|
|
|
|
receipts: ReceiptsSSZ, receiptsRoot: KeccakHash): Result[void, string] =
|
|
|
|
let calculatedReceiptsRoot = calcReceiptsRoot(receipts)
|
|
|
|
|
|
|
|
if calculatedReceiptsRoot != receiptsRoot:
|
|
|
|
return err("Unexpected receipt root")
|
|
|
|
else:
|
|
|
|
return ok()
|
|
|
|
|
|
|
|
proc validateReceiptsBytes*(
|
|
|
|
bytes: openArray[byte],
|
|
|
|
receiptsRoot: KeccakHash): Result[seq[Receipt], string] =
|
|
|
|
## Fully decode the SSZ Block Body and validate it against the header.
|
|
|
|
let receipts =
|
|
|
|
try:
|
|
|
|
SSZ.decode(bytes, ReceiptsSSZ)
|
|
|
|
except SszError as e:
|
2022-07-05 12:42:55 +00:00
|
|
|
return err("Failed to decode receipts: " & e.msg)
|
2022-07-01 19:51:51 +00:00
|
|
|
|
|
|
|
? validateReceipts(receipts, receiptsRoot)
|
|
|
|
|
|
|
|
seq[Receipt].fromReceipts(receipts)
|
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
## ContentDB helper calls for specific history network types
|
2022-07-01 19:51:51 +00:00
|
|
|
|
2022-09-10 19:00:27 +00:00
|
|
|
proc get(db: ContentDB, T: type BlockHeader, contentId: ContentId): Option[T] =
|
2022-07-01 19:51:51 +00:00
|
|
|
let contentFromDB = db.get(contentId)
|
|
|
|
if contentFromDB.isSome():
|
|
|
|
let res = decodeRlp(contentFromDB.get(), T)
|
|
|
|
if res.isErr():
|
|
|
|
raiseAssert(res.error)
|
|
|
|
else:
|
|
|
|
some(res.get())
|
|
|
|
else:
|
|
|
|
none(T)
|
|
|
|
|
2022-09-10 19:00:27 +00:00
|
|
|
proc get(db: ContentDB, T: type BlockBody, contentId: ContentId): Option[T] =
|
2022-07-01 19:51:51 +00:00
|
|
|
let contentFromDB = db.getSszDecoded(contentId, BlockBodySSZ)
|
|
|
|
if contentFromDB.isSome():
|
|
|
|
let res = T.fromPortalBlockBody(contentFromDB.get())
|
|
|
|
if res.isErr():
|
|
|
|
raiseAssert(res.error)
|
|
|
|
else:
|
|
|
|
some(res.get())
|
|
|
|
else:
|
|
|
|
none(T)
|
|
|
|
|
2022-09-10 19:00:27 +00:00
|
|
|
proc get(db: ContentDB, T: type seq[Receipt], contentId: ContentId): Option[T] =
|
2022-07-01 19:51:51 +00:00
|
|
|
let contentFromDB = db.getSszDecoded(contentId, ReceiptsSSZ)
|
|
|
|
if contentFromDB.isSome():
|
|
|
|
let res = T.fromReceipts(contentFromDB.get())
|
|
|
|
if res.isErr():
|
|
|
|
raiseAssert(res.error)
|
|
|
|
else:
|
|
|
|
some(res.get())
|
|
|
|
else:
|
|
|
|
none(T)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
proc get(
|
2022-09-10 19:00:27 +00:00
|
|
|
db: ContentDB, T: type EpochAccumulator, contentId: ContentId): Option[T] =
|
2022-08-01 19:00:21 +00:00
|
|
|
db.getSszDecoded(contentId, T)
|
|
|
|
|
|
|
|
proc getAccumulator(db: ContentDB): Option[Accumulator] =
|
|
|
|
db.getPermanentSszDecoded(subkey(kLatestAccumulator), Accumulator)
|
|
|
|
|
|
|
|
proc putAccumulator*(db: ContentDB, value: openArray[byte]) =
|
|
|
|
db.putPermanent(subkey(kLatestAccumulator), value)
|
|
|
|
|
2022-04-01 16:01:50 +00:00
|
|
|
proc getContentFromDb(
|
2022-07-11 14:29:16 +00:00
|
|
|
n: HistoryNetwork, T: type, contentId: ContentId): Option[T] =
|
|
|
|
if n.portalProtocol.inRange(contentId):
|
|
|
|
n.contentDB.get(T, contentId)
|
2022-03-02 14:29:12 +00:00
|
|
|
else:
|
2022-07-01 19:51:51 +00:00
|
|
|
none(T)
|
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
proc dbGetHandler(db: ContentDB, contentKey: ByteList):
|
|
|
|
(Option[ContentId], Option[seq[byte]]) {.raises: [Defect], gcsafe.} =
|
|
|
|
let keyOpt = decode(contentKey)
|
|
|
|
if keyOpt.isNone():
|
|
|
|
return (none(ContentId), none(seq[byte]))
|
|
|
|
|
|
|
|
let key = keyOpt.get()
|
|
|
|
|
|
|
|
case key.contentType:
|
|
|
|
of masterAccumulator:
|
|
|
|
(none(ContentId), db.getPermanent(subkey(kLatestAccumulator)))
|
|
|
|
else:
|
|
|
|
let contentId = key.toContentId()
|
|
|
|
(some(contentId), db.get(contentId))
|
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
## Public API to get the history network specific types, either from database
|
|
|
|
## or through a lookup on the Portal Network
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
const requestRetries = 4
|
|
|
|
# TODO: Currently doing 4 retries on lookups but only when the validation fails.
|
|
|
|
# This is to avoid nodes that provide garbage from blocking us with getting the
|
|
|
|
# requested data. Might want to also do that on a failed lookup, as perhaps this
|
|
|
|
# could occur when being really unlucky with nodes timing out on requests.
|
|
|
|
# Additionally, more improvements could be done with the lookup, as currently
|
|
|
|
# ongoing requests are cancelled after the receival of the first response,
|
|
|
|
# however that response is not yet validated at that moment.
|
|
|
|
|
2022-04-01 16:01:50 +00:00
|
|
|
proc getBlockHeader*(
|
2022-09-28 07:09:38 +00:00
|
|
|
n: HistoryNetwork, hash: BlockHash):
|
2022-04-01 16:01:50 +00:00
|
|
|
Future[Option[BlockHeader]] {.async.} =
|
2022-07-01 19:51:51 +00:00
|
|
|
let (keyEncoded, contentId) =
|
2022-09-28 07:09:38 +00:00
|
|
|
getEncodedKeyForContent(blockHeader, hash)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
let headerFromDb = n.getContentFromDb(BlockHeader, contentId)
|
2022-07-01 19:51:51 +00:00
|
|
|
if headerFromDb.isSome():
|
2022-07-29 12:24:07 +00:00
|
|
|
info "Fetched block header from database", hash, contentKey = keyEncoded
|
2022-07-01 19:51:51 +00:00
|
|
|
return headerFromDb
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
for i in 0..<requestRetries:
|
|
|
|
let headerContentLookup =
|
2022-07-11 14:29:16 +00:00
|
|
|
await n.portalProtocol.contentLookup(keyEncoded, contentId)
|
2022-07-05 12:42:55 +00:00
|
|
|
if headerContentLookup.isNone():
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Failed fetching block header from the network", hash, contentKey = keyEncoded
|
2022-07-05 12:42:55 +00:00
|
|
|
return none(BlockHeader)
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let headerContent = headerContentLookup.unsafeGet()
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let res = validateBlockHeaderBytes(headerContent.content, hash)
|
|
|
|
if res.isOk():
|
2022-07-29 12:24:07 +00:00
|
|
|
info "Fetched block header from the network", hash, contentKey = keyEncoded
|
2022-07-05 12:42:55 +00:00
|
|
|
# Content is valid we can propagate it to interested peers
|
2022-07-11 14:29:16 +00:00
|
|
|
n.portalProtocol.triggerPoke(
|
2022-07-05 12:42:55 +00:00
|
|
|
headerContent.nodesInterestedInContent,
|
|
|
|
keyEncoded,
|
|
|
|
headerContent.content
|
|
|
|
)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
n.portalProtocol.storeContent(contentId, headerContent.content)
|
2022-04-11 09:25:36 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
return some(res.get())
|
|
|
|
else:
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Validation of block header failed", err = res.error, hash, contentKey = keyEncoded
|
2022-07-05 12:42:55 +00:00
|
|
|
|
|
|
|
# Headers were requested `requestRetries` times and all failed on validation
|
|
|
|
return none(BlockHeader)
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-06-29 15:44:08 +00:00
|
|
|
proc getBlockBody*(
|
2022-09-28 07:09:38 +00:00
|
|
|
n: HistoryNetwork, hash: BlockHash, header: BlockHeader):
|
2022-07-05 12:42:55 +00:00
|
|
|
Future[Option[BlockBody]] {.async.} =
|
2022-07-20 10:46:42 +00:00
|
|
|
|
|
|
|
# Got header with empty body, no need to make any db calls or network requests
|
2022-09-03 18:15:35 +00:00
|
|
|
if header.txRoot == EMPTY_ROOT_HASH and header.ommersHash == EMPTY_UNCLE_HASH:
|
2022-07-20 10:46:42 +00:00
|
|
|
return some(BlockBody(transactions: @[], uncles: @[]))
|
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
let
|
2022-09-28 07:09:38 +00:00
|
|
|
(keyEncoded, contentId) = getEncodedKeyForContent(blockBody, hash)
|
2022-07-11 14:29:16 +00:00
|
|
|
bodyFromDb = n.getContentFromDb(BlockBody, contentId)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
if bodyFromDb.isSome():
|
2022-07-29 12:24:07 +00:00
|
|
|
info "Fetched block body from database", hash, contentKey = keyEncoded
|
2022-07-05 12:42:55 +00:00
|
|
|
return bodyFromDb
|
2022-03-18 12:06:57 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
for i in 0..<requestRetries:
|
|
|
|
let bodyContentLookup =
|
2022-07-11 14:29:16 +00:00
|
|
|
await n.portalProtocol.contentLookup(keyEncoded, contentId)
|
2022-07-20 10:46:42 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
if bodyContentLookup.isNone():
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Failed fetching block body from the network", hash, contentKey = keyEncoded
|
2022-07-20 10:46:42 +00:00
|
|
|
return none(BlockBody)
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let bodyContent = bodyContentLookup.unsafeGet()
|
2022-03-02 14:29:12 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let res = validateBlockBodyBytes(
|
|
|
|
bodyContent.content, header.txRoot, header.ommersHash)
|
|
|
|
if res.isOk():
|
2022-07-29 12:24:07 +00:00
|
|
|
info "Fetched block body from the network", hash, contentKey = keyEncoded
|
2022-04-15 14:10:05 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
# body is valid, propagate it to interested peers
|
2022-07-11 14:29:16 +00:00
|
|
|
n.portalProtocol.triggerPoke(
|
2022-07-05 12:42:55 +00:00
|
|
|
bodyContent.nodesInterestedInContent,
|
|
|
|
keyEncoded,
|
|
|
|
bodyContent.content
|
|
|
|
)
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
n.portalProtocol.storeContent(contentId, bodyContent.content)
|
2022-04-11 09:25:36 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
return some(res.get())
|
|
|
|
else:
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Validation of block body failed", err = res.error, hash, contentKey = keyEncoded
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
return none(BlockBody)
|
2022-06-29 15:44:08 +00:00
|
|
|
|
|
|
|
proc getBlock*(
|
2022-09-28 07:09:38 +00:00
|
|
|
n: HistoryNetwork, hash: BlockHash):
|
2022-06-29 15:44:08 +00:00
|
|
|
Future[Option[Block]] {.async.} =
|
2022-07-29 12:24:07 +00:00
|
|
|
debug "Trying to retrieve block with hash", hash
|
|
|
|
|
2022-09-28 07:09:38 +00:00
|
|
|
let headerOpt = await n.getBlockHeader(hash)
|
2022-07-01 19:51:51 +00:00
|
|
|
if headerOpt.isNone():
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Failed to get header when getting block with hash", hash
|
2022-07-01 19:51:51 +00:00
|
|
|
# Cannot validate block without header.
|
2022-06-29 15:44:08 +00:00
|
|
|
return none(Block)
|
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
let header = headerOpt.unsafeGet()
|
2022-06-29 15:44:08 +00:00
|
|
|
|
2022-09-28 07:09:38 +00:00
|
|
|
let bodyOpt = await n.getBlockBody(hash, header)
|
2022-06-29 15:44:08 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
if bodyOpt.isNone():
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Failed to get body when gettin block with hash", hash
|
2022-06-29 15:44:08 +00:00
|
|
|
return none(Block)
|
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
let body = bodyOpt.unsafeGet()
|
2022-06-29 15:44:08 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
return some((header, body))
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-06-10 10:24:53 +00:00
|
|
|
proc getReceipts*(
|
2022-07-11 14:29:16 +00:00
|
|
|
n: HistoryNetwork,
|
2022-06-10 10:24:53 +00:00
|
|
|
hash: BlockHash,
|
2022-06-29 15:44:08 +00:00
|
|
|
header: BlockHeader): Future[Option[seq[Receipt]]] {.async.} =
|
2022-09-03 18:15:35 +00:00
|
|
|
if header.receiptRoot == EMPTY_ROOT_HASH:
|
2022-07-05 12:42:55 +00:00
|
|
|
# Short path for empty receipts indicated by receipts root
|
2022-06-10 10:24:53 +00:00
|
|
|
return some(newSeq[Receipt]())
|
|
|
|
|
2022-09-28 07:09:38 +00:00
|
|
|
let (keyEncoded, contentId) = getEncodedKeyForContent(receipts, hash)
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
let receiptsFromDb = n.getContentFromDb(seq[Receipt], contentId)
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-01 19:51:51 +00:00
|
|
|
if receiptsFromDb.isSome():
|
2022-06-10 10:24:53 +00:00
|
|
|
info "Fetched receipts from database", hash
|
2022-07-05 12:42:55 +00:00
|
|
|
return receiptsFromDb
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
for i in 0..<requestRetries:
|
|
|
|
let receiptsContentLookup =
|
2022-07-11 14:29:16 +00:00
|
|
|
await n.portalProtocol.contentLookup(keyEncoded, contentId)
|
2022-07-05 12:42:55 +00:00
|
|
|
if receiptsContentLookup.isNone():
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Failed fetching receipts from the network", hash, contentKey = keyEncoded
|
2022-07-05 12:42:55 +00:00
|
|
|
return none(seq[Receipt])
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let receiptsContent = receiptsContentLookup.unsafeGet()
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let res = validateReceiptsBytes(receiptsContent.content, header.receiptRoot)
|
|
|
|
if res.isOk():
|
2022-07-29 12:24:07 +00:00
|
|
|
info "Fetched receipts from the network", hash, contentKey = keyEncoded
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
let receipts = res.get()
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
# receipts are valid, propagate it to interested peers
|
2022-07-11 14:29:16 +00:00
|
|
|
n.portalProtocol.triggerPoke(
|
2022-07-05 12:42:55 +00:00
|
|
|
receiptsContent.nodesInterestedInContent,
|
|
|
|
keyEncoded,
|
|
|
|
receiptsContent.content
|
|
|
|
)
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
n.portalProtocol.storeContent(contentId, receiptsContent.content)
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
return some(res.get())
|
|
|
|
else:
|
2022-07-29 12:24:07 +00:00
|
|
|
warn "Validation of receipts failed", err = res.error, hash, contentKey = keyEncoded
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-07-05 12:42:55 +00:00
|
|
|
return none(seq[Receipt])
|
2022-06-10 10:24:53 +00:00
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
proc getEpochAccumulator(
|
|
|
|
n: HistoryNetwork, epochHash: Digest):
|
|
|
|
Future[Option[EpochAccumulator]] {.async.} =
|
|
|
|
let
|
|
|
|
contentKey = ContentKey(
|
|
|
|
contentType: epochAccumulator,
|
|
|
|
epochAccumulatorKey: EpochAccumulatorKey(epochHash: epochHash))
|
2022-07-01 19:51:51 +00:00
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
keyEncoded = encode(contentKey)
|
|
|
|
contentId = toContentId(keyEncoded)
|
|
|
|
|
|
|
|
accumulatorFromDb = n.getContentFromDb(EpochAccumulator, contentId)
|
|
|
|
|
|
|
|
if accumulatorFromDb.isSome():
|
|
|
|
info "Fetched epoch accumulator from database", epochHash
|
|
|
|
return accumulatorFromDb
|
|
|
|
|
|
|
|
for i in 0..<requestRetries:
|
|
|
|
let contentLookup =
|
|
|
|
await n.portalProtocol.contentLookup(keyEncoded, contentId)
|
|
|
|
if contentLookup.isNone():
|
|
|
|
warn "Failed fetching epoch accumulator from the network", epochHash
|
|
|
|
return none(EpochAccumulator)
|
|
|
|
|
|
|
|
let accumulatorContent = contentLookup.unsafeGet()
|
|
|
|
|
|
|
|
let epochAccumulator =
|
|
|
|
try:
|
|
|
|
SSZ.decode(accumulatorContent.content, EpochAccumulator)
|
|
|
|
except SszError:
|
|
|
|
continue
|
|
|
|
# return none(EpochAccumulator)
|
|
|
|
|
|
|
|
let hash = hash_tree_root(epochAccumulator)
|
|
|
|
if hash == epochHash:
|
|
|
|
info "Fetched epoch accumulator from the network", epochHash
|
|
|
|
|
|
|
|
n.portalProtocol.triggerPoke(
|
|
|
|
accumulatorContent.nodesInterestedInContent,
|
|
|
|
keyEncoded,
|
|
|
|
accumulatorContent.content
|
|
|
|
)
|
|
|
|
|
|
|
|
n.portalProtocol.storeContent(contentId, accumulatorContent.content)
|
|
|
|
|
|
|
|
return some(epochAccumulator)
|
|
|
|
else:
|
|
|
|
warn "Validation of epoch accumulator failed",
|
|
|
|
hash, expectedHash = epochHash
|
|
|
|
|
|
|
|
return none(EpochAccumulator)
|
|
|
|
|
2022-08-04 06:34:53 +00:00
|
|
|
proc getBlock*(
|
2022-09-28 07:09:38 +00:00
|
|
|
n: HistoryNetwork, bn: UInt256):
|
2022-08-04 06:34:53 +00:00
|
|
|
Future[Result[Option[Block], string]] {.async.} =
|
|
|
|
|
|
|
|
# TODO for now checking accumulator only in db, we could also ask our
|
|
|
|
# peers for it.
|
|
|
|
let accumulatorOpt = n.contentDB.getAccumulator()
|
|
|
|
|
|
|
|
if accumulatorOpt.isNone():
|
|
|
|
return err("Master accumulator not found in database")
|
|
|
|
|
|
|
|
let accumulator = accumulatorOpt.unsafeGet()
|
|
|
|
|
|
|
|
let hashResponse = accumulator.getHeaderHashForBlockNumber(bn)
|
|
|
|
|
|
|
|
case hashResponse.kind
|
|
|
|
of BHash:
|
|
|
|
# we got header hash in current epoch accumulator, try to retrieve it from network
|
2022-09-28 07:09:38 +00:00
|
|
|
let blockResponse = await n.getBlock(hashResponse.blockHash)
|
2022-08-04 06:34:53 +00:00
|
|
|
return ok(blockResponse)
|
|
|
|
of HEpoch:
|
|
|
|
let digest = Digest(data: hashResponse.epochHash)
|
|
|
|
|
|
|
|
let epochOpt = await n.getEpochAccumulator(digest)
|
|
|
|
|
|
|
|
if epochOpt.isNone():
|
|
|
|
return err("Cannot retrieve epoch accumulator for given block number")
|
|
|
|
|
|
|
|
let
|
|
|
|
epoch = epochOpt.unsafeGet()
|
|
|
|
blockHash = epoch[hashResponse.blockRelativeIndex].blockHash
|
|
|
|
|
2022-09-28 07:09:38 +00:00
|
|
|
let maybeBlock = await n.getBlock(blockHash)
|
2022-08-04 06:34:53 +00:00
|
|
|
|
|
|
|
return ok(maybeBlock)
|
|
|
|
of UnknownBlockNumber:
|
|
|
|
return err("Block number not included in master accumulator")
|
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
proc getInitialMasterAccumulator*(
|
|
|
|
n: HistoryNetwork):
|
|
|
|
Future[bool] {.async.} =
|
|
|
|
let
|
|
|
|
contentKey = ContentKey(
|
|
|
|
contentType: masterAccumulator,
|
|
|
|
masterAccumulatorKey: MasterAccumulatorKey(accumulaterKeyType: latest))
|
|
|
|
keyEncoded = encode(contentKey)
|
|
|
|
|
|
|
|
let nodes = await n.portalProtocol.queryRandom()
|
|
|
|
|
|
|
|
var hashes: CountTable[Accumulator]
|
|
|
|
|
|
|
|
for node in nodes:
|
|
|
|
# TODO: Could make concurrent
|
|
|
|
let foundContentRes = await n.portalProtocol.findContent(node, keyEncoded)
|
|
|
|
if foundContentRes.isOk():
|
|
|
|
let foundContent = foundContentRes.get()
|
|
|
|
if foundContent.kind == Content:
|
|
|
|
let masterAccumulator =
|
|
|
|
try:
|
|
|
|
SSZ.decode(foundContent.content, Accumulator)
|
|
|
|
except SszError:
|
|
|
|
continue
|
|
|
|
hashes.inc(masterAccumulator)
|
|
|
|
let (accumulator, count) = hashes.largest()
|
|
|
|
|
|
|
|
if count > 1: # Should be increased eventually
|
|
|
|
n.contentDB.putAccumulator(foundContent.content)
|
|
|
|
return true
|
|
|
|
|
|
|
|
# Could not find a common accumulator from all the queried nodes
|
|
|
|
return false
|
|
|
|
|
|
|
|
proc buildProof*(n: HistoryNetwork, header: BlockHeader):
|
|
|
|
Future[Result[seq[Digest], string]] {.async.} =
|
|
|
|
# Note: Temporarily needed proc until proofs are send over with headers.
|
|
|
|
let accumulatorOpt = n.contentDB.getAccumulator()
|
|
|
|
if accumulatorOpt.isNone():
|
|
|
|
return err("Master accumulator not found in database")
|
|
|
|
|
|
|
|
let
|
|
|
|
accumulator = accumulatorOpt.get()
|
|
|
|
epochIndex = getEpochIndex(header)
|
|
|
|
epochHash = Digest(data: accumulator.historicalEpochs[epochIndex])
|
|
|
|
|
|
|
|
epochAccumulatorOpt = await n.getEpochAccumulator(epochHash)
|
|
|
|
|
|
|
|
if epochAccumulatorOpt.isNone():
|
|
|
|
return err("Epoch accumulator not found")
|
|
|
|
|
|
|
|
let
|
|
|
|
epochAccumulator = epochAccumulatorOpt.get()
|
|
|
|
headerRecordIndex = getHeaderRecordIndex(header, epochIndex)
|
|
|
|
# TODO: Implement more generalized `get_generalized_index`
|
|
|
|
gIndex = GeneralizedIndex(epochSize*2*2 + (headerRecordIndex*2))
|
|
|
|
|
|
|
|
return epochAccumulator.build_proof(gIndex)
|
|
|
|
|
|
|
|
proc verifyCanonicalChain(
|
|
|
|
n: HistoryNetwork, header: BlockHeader):
|
|
|
|
Future[Result[void, string]] {.async.} =
|
|
|
|
when not canonicalVerify:
|
|
|
|
return ok()
|
|
|
|
|
|
|
|
let accumulatorOpt = n.contentDB.getAccumulator()
|
|
|
|
if accumulatorOpt.isNone():
|
|
|
|
# Should acquire a master accumulator first
|
|
|
|
return err("Cannot accept any data without a master accumulator")
|
|
|
|
|
|
|
|
let accumulator = accumulatorOpt.get()
|
|
|
|
|
|
|
|
# Note: It is a bit silly to build a proof, as we still need to request the
|
|
|
|
# epoch accumulators for it, and could just verify it with those. But the
|
|
|
|
# idea here is that eventually this gets changed so that the proof is send
|
|
|
|
# together with the header.
|
|
|
|
let proofOpt =
|
|
|
|
if header.inCurrentEpoch(accumulator):
|
|
|
|
none(seq[Digest])
|
|
|
|
else:
|
|
|
|
let proof = await n.buildProof(header)
|
|
|
|
if proof.isErr():
|
|
|
|
# Can't verify without master and epoch accumulators
|
|
|
|
return err("Cannot build proof: " & proof.error)
|
|
|
|
else:
|
|
|
|
some(proof.get())
|
|
|
|
|
|
|
|
return verifyHeader(accumulator, header, proofOpt)
|
2022-07-01 19:51:51 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
proc validateContent(
|
|
|
|
n: HistoryNetwork, content: seq[byte], contentKey: ByteList):
|
|
|
|
Future[bool] {.async.} =
|
2022-04-11 17:42:38 +00:00
|
|
|
let keyOpt = contentKey.decode()
|
|
|
|
|
|
|
|
if keyOpt.isNone():
|
|
|
|
return false
|
|
|
|
|
|
|
|
let key = keyOpt.get()
|
|
|
|
|
|
|
|
case key.contentType:
|
|
|
|
of blockHeader:
|
2022-08-01 19:00:21 +00:00
|
|
|
let validateResult =
|
|
|
|
validateBlockHeaderBytes(content, key.blockHeaderKey.blockHash)
|
|
|
|
if validateResult.isErr():
|
|
|
|
warn "Invalid block header offered", error = validateResult.error
|
|
|
|
return false
|
|
|
|
|
|
|
|
let header = validateResult.get()
|
|
|
|
|
|
|
|
let verifyResult = await n.verifyCanonicalChain(header)
|
|
|
|
if verifyResult.isErr():
|
|
|
|
warn "Failed on check if header is part of canonical chain",
|
|
|
|
error = verifyResult.error
|
|
|
|
return false
|
|
|
|
else:
|
|
|
|
return true
|
2022-04-11 17:42:38 +00:00
|
|
|
of blockBody:
|
2022-09-28 07:09:38 +00:00
|
|
|
let headerOpt = await n.getBlockHeader(key.blockBodyKey.blockHash)
|
2022-07-11 14:29:16 +00:00
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
if headerOpt.isNone():
|
|
|
|
warn "Cannot find the header, no way to validate the block body"
|
2022-07-11 14:29:16 +00:00
|
|
|
return false
|
2022-08-01 19:00:21 +00:00
|
|
|
|
|
|
|
let header = headerOpt.get()
|
|
|
|
let validationResult =
|
|
|
|
validateBlockBodyBytes(content, header.txRoot, header.ommersHash)
|
|
|
|
|
|
|
|
if validationResult.isErr():
|
|
|
|
warn "Failed validating block body", error = validationResult.error
|
|
|
|
return false
|
|
|
|
|
|
|
|
let verifyResult = await n.verifyCanonicalChain(header)
|
|
|
|
if verifyResult.isErr():
|
|
|
|
warn "Failed on check if header is part of canonical chain",
|
|
|
|
error = verifyResult.error
|
|
|
|
return false
|
|
|
|
else:
|
|
|
|
return true
|
2022-04-11 17:42:38 +00:00
|
|
|
of receipts:
|
2022-09-28 07:09:38 +00:00
|
|
|
let headerOpt = await n.getBlockHeader(key.receiptsKey.blockHash)
|
2022-07-11 14:29:16 +00:00
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
if headerOpt.isNone():
|
|
|
|
warn "Cannot find the header, no way to validate the receipts"
|
|
|
|
return false
|
|
|
|
|
|
|
|
let header = headerOpt.get()
|
|
|
|
let validationResult =
|
|
|
|
validateReceiptsBytes(content, header.receiptRoot)
|
|
|
|
|
|
|
|
if validationResult.isErr():
|
|
|
|
warn "Failed validating receipts", error = validationResult.error
|
2022-07-11 14:29:16 +00:00
|
|
|
return false
|
2022-08-01 19:00:21 +00:00
|
|
|
|
|
|
|
let verifyResult = await n.verifyCanonicalChain(header)
|
|
|
|
if verifyResult.isErr():
|
|
|
|
warn "Failed on check if header is part of canonical chain",
|
|
|
|
error = verifyResult.error
|
|
|
|
return false
|
|
|
|
else:
|
|
|
|
return true
|
2022-06-14 21:38:34 +00:00
|
|
|
of epochAccumulator:
|
2022-08-01 19:00:21 +00:00
|
|
|
# Check first if epochHash is part of master accumulator
|
|
|
|
let masterAccumulator = n.contentDB.getAccumulator()
|
|
|
|
if masterAccumulator.isNone():
|
|
|
|
error "Cannot accept any data without a master accumulator"
|
|
|
|
return false
|
|
|
|
|
|
|
|
let epochHash = key.epochAccumulatorKey.epochHash
|
|
|
|
|
|
|
|
if not masterAccumulator.get().historicalEpochs.contains(epochHash.data):
|
|
|
|
warn "Offered epoch accumulator is not part of master accumulator"
|
|
|
|
return false
|
|
|
|
|
|
|
|
let epochAccumulator =
|
|
|
|
try:
|
|
|
|
SSZ.decode(content, EpochAccumulator)
|
|
|
|
except SszError:
|
|
|
|
warn "Failed decoding epoch accumulator"
|
|
|
|
return false
|
|
|
|
# Next check the hash tree root, as this is probably more expensive
|
|
|
|
let hash = hash_tree_root(epochAccumulator)
|
|
|
|
if hash != epochHash:
|
|
|
|
warn "Epoch accumulator has invalid root hash"
|
|
|
|
return false
|
|
|
|
else:
|
|
|
|
return true
|
2022-06-14 21:38:34 +00:00
|
|
|
of masterAccumulator:
|
2022-08-01 19:00:21 +00:00
|
|
|
# Don't allow a master accumulator to be offered, we only request it.
|
|
|
|
warn "Node does not accept master accumulators through offer/accept"
|
|
|
|
return false
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-01-06 08:06:05 +00:00
|
|
|
proc new*(
|
|
|
|
T: type HistoryNetwork,
|
|
|
|
baseProtocol: protocol.Protocol,
|
|
|
|
contentDB: ContentDB,
|
2022-08-17 07:32:06 +00:00
|
|
|
streamManager: StreamManager,
|
2022-01-18 08:01:22 +00:00
|
|
|
bootstrapRecords: openArray[Record] = [],
|
|
|
|
portalConfig: PortalProtocolConfig = defaultPortalProtocolConfig): T =
|
2022-08-17 07:32:06 +00:00
|
|
|
|
|
|
|
let cq = newAsyncQueue[(ContentKeysList, seq[seq[byte]])](50)
|
|
|
|
|
|
|
|
let s = streamManager.registerNewStream(cq)
|
|
|
|
|
2021-10-09 11:22:03 +00:00
|
|
|
let portalProtocol = PortalProtocol.new(
|
2022-04-11 17:42:38 +00:00
|
|
|
baseProtocol, historyProtocolId, contentDB,
|
2022-08-17 07:32:06 +00:00
|
|
|
toContentIdHandler, dbGetHandler, s, bootstrapRecords,
|
2022-01-18 08:01:22 +00:00
|
|
|
config = portalConfig)
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-08-17 07:32:06 +00:00
|
|
|
return HistoryNetwork(
|
|
|
|
portalProtocol: portalProtocol,
|
|
|
|
contentDB: contentDB,
|
|
|
|
contentQueue: cq
|
|
|
|
)
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-09-29 06:42:54 +00:00
|
|
|
proc validateContent(
|
|
|
|
n: HistoryNetwork,
|
|
|
|
contentKeys: ContentKeysList,
|
|
|
|
contentItems: seq[seq[byte]]): Future[bool] {.async.} =
|
|
|
|
# content passed here can have less items then contentKeys, but not more.
|
|
|
|
for i, contentItem in contentItems:
|
|
|
|
let contentKey = contentKeys[i]
|
|
|
|
if await n.validateContent(contentItem, contentKey):
|
|
|
|
let contentIdOpt = n.portalProtocol.toContentId(contentKey)
|
|
|
|
if contentIdOpt.isNone():
|
|
|
|
error "Received offered content with invalid content key", contentKey
|
|
|
|
return false
|
|
|
|
|
|
|
|
let contentId = contentIdOpt.get()
|
|
|
|
|
|
|
|
n.portalProtocol.storeContent(contentId, contentItem)
|
|
|
|
|
|
|
|
info "Received offered content validated successfully", contentKey
|
|
|
|
|
|
|
|
else:
|
|
|
|
error "Received offered content failed validation", contentKey
|
|
|
|
return false
|
|
|
|
|
|
|
|
return true
|
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
proc processContentLoop(n: HistoryNetwork) {.async.} =
|
|
|
|
try:
|
|
|
|
while true:
|
|
|
|
let (contentKeys, contentItems) =
|
2022-08-17 07:32:06 +00:00
|
|
|
await n.contentQueue.popFirst()
|
2022-07-11 14:29:16 +00:00
|
|
|
|
2022-09-29 06:42:54 +00:00
|
|
|
# When there is one invalid content item, all other content items are
|
|
|
|
# dropped and not gossiped around.
|
|
|
|
# TODO: Differentiate between failures due to invalid data and failures
|
|
|
|
# due to missing network data for validation.
|
|
|
|
if await n.validateContent(contentKeys, contentItems):
|
|
|
|
asyncSpawn n.portalProtocol.neighborhoodGossip(contentKeys, contentItems)
|
2022-07-11 14:29:16 +00:00
|
|
|
|
|
|
|
except CancelledError:
|
|
|
|
trace "processContentLoop canceled"
|
|
|
|
|
|
|
|
proc start*(n: HistoryNetwork) =
|
2022-04-13 05:56:01 +00:00
|
|
|
info "Starting Portal execution history network",
|
2022-07-11 14:29:16 +00:00
|
|
|
protocolId = n.portalProtocol.protocolId
|
|
|
|
n.portalProtocol.start()
|
|
|
|
|
|
|
|
n.processContentLoop = processContentLoop(n)
|
|
|
|
|
2022-08-01 19:00:21 +00:00
|
|
|
proc initMasterAccumulator*(
|
|
|
|
n: HistoryNetwork,
|
|
|
|
accumulator: Option[Accumulator]) {.async.} =
|
|
|
|
if accumulator.isSome():
|
|
|
|
n.contentDB.putAccumulator(SSZ.encode(accumulator.get()))
|
|
|
|
info "Successfully retrieved master accumulator from local data"
|
|
|
|
else:
|
|
|
|
while true:
|
|
|
|
if await n.getInitialMasterAccumulator():
|
|
|
|
info "Successfully retrieved master accumulator from the network"
|
|
|
|
return
|
|
|
|
else:
|
|
|
|
warn "Could not retrieve initial master accumulator from the network"
|
|
|
|
when not canonicalVerify:
|
|
|
|
return
|
|
|
|
else:
|
|
|
|
await sleepAsync(2.seconds)
|
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
proc stop*(n: HistoryNetwork) =
|
|
|
|
n.portalProtocol.stop()
|
2021-10-09 11:22:03 +00:00
|
|
|
|
2022-07-11 14:29:16 +00:00
|
|
|
if not n.processContentLoop.isNil:
|
|
|
|
n.processContentLoop.cancel()
|