525 lines
16 KiB
Nim
525 lines
16 KiB
Nim
# fluffy
|
|
# Copyright (c) 2024 Status Research & Development GmbH
|
|
# Licensed and distributed under either of
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
{.push raises: [].}
|
|
|
|
import
|
|
std/[strformat, typetraits],
|
|
results,
|
|
stew/[endians2, io2, byteutils, arrayops],
|
|
stint,
|
|
snappy,
|
|
eth/common/[headers_rlp, blocks_rlp, receipts_rlp],
|
|
beacon_chain/spec/beacon_time,
|
|
ssz_serialization,
|
|
ncli/e2store,
|
|
../network/history/validation/historical_hashes_accumulator
|
|
|
|
from eth/common/eth_types_rlp import rlpHash
|
|
from nimcrypto/hash import fromHex
|
|
from ../../nimbus/utils/utils import calcTxRoot, calcReceiptsRoot
|
|
|
|
export e2store.readRecord
|
|
|
|
# Implementation of era1 file format as current described in:
|
|
# https://github.com/ethereum/go-ethereum/pull/26621
|
|
|
|
# era1 := Version | block-tuple* | other-entries* | Accumulator | BlockIndex
|
|
# block-tuple := CompressedHeader | CompressedBody | CompressedReceipts | TotalDifficulty
|
|
|
|
# block-index := starting-number | index | index | index ... | count
|
|
|
|
# CompressedHeader = { type: 0x03, data: snappyFramed(rlp(header)) }
|
|
# CompressedBody = { type: 0x04, data: snappyFramed(rlp(body)) }
|
|
# CompressedReceipts = { type: 0x05, data: snappyFramed(rlp(receipts)) }
|
|
# TotalDifficulty = { type: 0x06, data: uint256(header.total_difficulty) }
|
|
# AccumulatorRoot = { type: 0x07, data: hash_tree_root(List(HeaderRecord, 8192)) }
|
|
# BlockIndex = { type: 0x3266, data: block-index }
|
|
|
|
# Important note:
|
|
# Snappy does not give the same compression result as the implementation used
|
|
# by go-ethereum for some block headers and block bodies. This means that we
|
|
# cannot rely on the secondary verification mechanism that is based on doing the
|
|
# sha256sum of the full era1 files.
|
|
#
|
|
|
|
const
|
|
# Note: When specification is more official, these could go with the other
|
|
# E2S types.
|
|
CompressedHeader* = [byte 0x03, 0x00]
|
|
CompressedBody* = [byte 0x04, 0x00]
|
|
CompressedReceipts* = [byte 0x05, 0x00]
|
|
TotalDifficulty* = [byte 0x06, 0x00]
|
|
AccumulatorRoot* = [byte 0x07, 0x00]
|
|
E2BlockIndex* = [byte 0x66, 0x32]
|
|
|
|
MaxEra1Size* = 8192
|
|
|
|
type
|
|
BlockIndex* = object
|
|
startNumber*: uint64
|
|
offsets*: seq[int64] # Absolute positions in file
|
|
|
|
Era1* = distinct uint64 # Period of 8192 blocks (not an exact time unit)
|
|
|
|
Era1Group* = object
|
|
blockIndex*: BlockIndex
|
|
|
|
# As stated, not really a time unit but nevertheless, need the borrows
|
|
ethTimeUnit Era1
|
|
|
|
template lenu64(x: untyped): untyped =
|
|
uint64(len(x))
|
|
|
|
# Note: appendIndex, appendRecord and readIndex for BlockIndex are very similar
|
|
# to its consensus layer counter parts. The difference lies in the naming of
|
|
# slots vs block numbers and there is different behavior for the first era
|
|
# (first slot) and the last era (era1 ends at merge block).
|
|
|
|
proc appendIndex*(
|
|
f: IoHandle, startNumber: uint64, offsets: openArray[int64]
|
|
): Result[int64, string] =
|
|
let
|
|
len = offsets.len() * sizeof(int64) + 16
|
|
pos = ?f.appendHeader(E2BlockIndex, len)
|
|
|
|
?f.append(startNumber.uint64.toBytesLE())
|
|
|
|
for v in offsets:
|
|
?f.append(cast[uint64](v - pos).toBytesLE())
|
|
|
|
?f.append(offsets.lenu64().toBytesLE())
|
|
|
|
ok(pos)
|
|
|
|
proc appendRecord(f: IoHandle, index: BlockIndex): Result[int64, string] =
|
|
f.appendIndex(index.startNumber, index.offsets)
|
|
|
|
proc readBlockIndex*(f: IoHandle): Result[BlockIndex, string] =
|
|
var
|
|
buf: seq[byte]
|
|
pos: int
|
|
|
|
let
|
|
startPos = ?f.getFilePos().mapErr(toString)
|
|
fileSize = ?f.getFileSize().mapErr(toString)
|
|
header = ?f.readRecord(buf)
|
|
|
|
if header.typ != E2BlockIndex:
|
|
return err("not an index")
|
|
if buf.len < 16:
|
|
return err("index entry too small")
|
|
if buf.len mod 8 != 0:
|
|
return err("index length invalid")
|
|
|
|
let
|
|
blockNumber = uint64.fromBytesLE(buf.toOpenArray(pos, pos + 7))
|
|
count = buf.len div 8 - 2
|
|
pos += 8
|
|
|
|
# technically not an error, but we'll throw this sanity check in here..
|
|
if blockNumber > int32.high().uint64:
|
|
return err("fishy block number")
|
|
|
|
var offsets = newSeqUninitialized[int64](count)
|
|
for i in 0 ..< count:
|
|
let
|
|
offset = uint64.fromBytesLE(buf.toOpenArray(pos, pos + 7))
|
|
absolute =
|
|
if offset == 0:
|
|
0'i64
|
|
else:
|
|
# Wrapping math is actually convenient here
|
|
cast[int64](cast[uint64](startPos) + offset)
|
|
|
|
if absolute < 0 or absolute > fileSize:
|
|
return err("invalid offset")
|
|
offsets[i] = absolute
|
|
pos += 8
|
|
|
|
if uint64(count) != uint64.fromBytesLE(buf.toOpenArray(pos, pos + 7)):
|
|
return err("invalid count")
|
|
|
|
ok(BlockIndex(startNumber: blockNumber, offsets: offsets))
|
|
|
|
proc skipRecord*(f: IoHandle): Result[void, string] =
|
|
let header = ?readHeader(f)
|
|
if header.len > 0:
|
|
?f.setFilePos(header.len, SeekPosition.SeekCurrent).mapErr(ioErrorMsg)
|
|
|
|
ok()
|
|
|
|
func startNumber*(era: Era1): uint64 =
|
|
era * MaxEra1Size
|
|
|
|
func endNumber*(era: Era1): uint64 =
|
|
if (era + 1) * MaxEra1Size - 1'u64 >= mergeBlockNumber:
|
|
# The incomplete era just before the merge
|
|
mergeBlockNumber - 1'u64
|
|
else:
|
|
(era + 1) * MaxEra1Size - 1'u64
|
|
|
|
func endNumber*(blockIdx: BlockIndex): uint64 =
|
|
blockIdx.startNumber + blockIdx.offsets.lenu64() - 1
|
|
|
|
func era*(blockNumber: uint64): Era1 =
|
|
Era1(blockNumber div MaxEra1Size)
|
|
|
|
func offsetsLen(startNumber: uint64): int =
|
|
# For the era where the merge happens the era files only holds the blocks
|
|
# until the merge block so the offsets length needs to be adapted too.
|
|
if startNumber.era() >= mergeBlockNumber.era():
|
|
int((mergeBlockNumber) mod MaxEra1Size)
|
|
else:
|
|
MaxEra1Size
|
|
|
|
proc toCompressedRlpBytes(item: auto): seq[byte] =
|
|
snappy.encodeFramed(rlp.encode(item))
|
|
|
|
proc fromCompressedRlpBytes[T](bytes: openArray[byte], v: var T): Result[void, string] =
|
|
try:
|
|
v = rlp.decode(decodeFramed(bytes, checkIntegrity = false), T)
|
|
ok()
|
|
except RlpError as e:
|
|
err("Invalid compressed RLP data for " & $T & ": " & e.msg)
|
|
|
|
proc init*(T: type Era1Group, f: IoHandle, startNumber: uint64): Result[T, string] =
|
|
discard ?f.appendHeader(E2Version, 0)
|
|
|
|
ok(
|
|
Era1Group(
|
|
blockIndex: BlockIndex(
|
|
startNumber: startNumber, offsets: newSeq[int64](startNumber.offsetsLen())
|
|
)
|
|
)
|
|
)
|
|
|
|
proc update*(
|
|
g: var Era1Group,
|
|
f: IoHandle,
|
|
blockNumber: uint64,
|
|
header, body, receipts, totalDifficulty: openArray[byte],
|
|
): Result[void, string] =
|
|
doAssert blockNumber >= g.blockIndex.startNumber
|
|
|
|
g.blockIndex.offsets[int(blockNumber - g.blockIndex.startNumber)] =
|
|
?f.appendRecord(CompressedHeader, header)
|
|
discard ?f.appendRecord(CompressedBody, body)
|
|
discard ?f.appendRecord(CompressedReceipts, receipts)
|
|
discard ?f.appendRecord(TotalDifficulty, totalDifficulty)
|
|
|
|
ok()
|
|
|
|
proc update*(
|
|
g: var Era1Group,
|
|
f: IoHandle,
|
|
blockNumber: uint64,
|
|
header: headers.Header,
|
|
body: BlockBody,
|
|
receipts: seq[Receipt],
|
|
totalDifficulty: UInt256,
|
|
): Result[void, string] =
|
|
g.update(
|
|
f,
|
|
blockNumber,
|
|
toCompressedRlpBytes(header),
|
|
toCompressedRlpBytes(body),
|
|
toCompressedRlpBytes(receipts),
|
|
totalDifficulty.toBytesLE(),
|
|
)
|
|
|
|
proc finish*(
|
|
g: var Era1Group, f: IoHandle, accumulatorRoot: Digest, lastBlockNumber: uint64
|
|
): Result[void, string] =
|
|
let accumulatorRootPos = ?f.appendRecord(AccumulatorRoot, accumulatorRoot.data)
|
|
|
|
if lastBlockNumber > 0:
|
|
discard ?f.appendRecord(g.blockIndex)
|
|
|
|
# TODO:
|
|
# This is not something added in current specification of era1.
|
|
# But perhaps we want to be able to quickly jump to acummulator root.
|
|
# discard ? f.appendIndex(lastBlockNumber, [accumulatorRootPos])
|
|
discard accumulatorRootPos
|
|
|
|
ok()
|
|
|
|
func shortLog*(x: Digest): string =
|
|
x.data.toOpenArray(0, 3).toHex()
|
|
|
|
func era1FileName*(network: string, era: Era1, eraRoot: Digest): string =
|
|
try:
|
|
&"{network}-{era.uint64:05}-{shortLog(eraRoot)}.era1"
|
|
except ValueError as exc:
|
|
raiseAssert exc.msg
|
|
|
|
# Helpers to directly read objects from era1 files
|
|
# TODO: Might want to var parameters to avoid copying as is done for era files.
|
|
|
|
type
|
|
Era1File* = ref object
|
|
handle: Opt[IoHandle]
|
|
blockIdx*: BlockIndex
|
|
|
|
BlockTuple* =
|
|
tuple[header: headers.Header, body: BlockBody, receipts: seq[Receipt], td: UInt256]
|
|
|
|
proc open*(_: type Era1File, name: string): Result[Era1File, string] =
|
|
var f = Opt[IoHandle].ok(?openFile(name, {OpenFlags.Read}).mapErr(ioErrorMsg))
|
|
|
|
defer:
|
|
if f.isSome():
|
|
discard closeFile(f[])
|
|
|
|
# Indices can be found at the end of each era file - we only support
|
|
# single-era files for now
|
|
?f[].setFilePos(0, SeekPosition.SeekEnd).mapErr(ioErrorMsg)
|
|
|
|
# Last in the file is the block index
|
|
let blockIdxPos = ?f[].findIndexStartOffset()
|
|
?f[].setFilePos(blockIdxPos, SeekPosition.SeekCurrent).mapErr(ioErrorMsg)
|
|
|
|
let blockIdx = ?f[].readBlockIndex()
|
|
if blockIdx.offsets.len() != blockIdx.startNumber.offsetsLen():
|
|
return err("Block index length invalid")
|
|
|
|
let res = Era1File(handle: f, blockIdx: blockIdx)
|
|
reset(f)
|
|
ok res
|
|
|
|
proc close*(f: Era1File) =
|
|
if f.handle.isSome():
|
|
discard closeFile(f.handle.get())
|
|
reset(f.handle)
|
|
|
|
proc skipRecord*(f: Era1File): Result[void, string] =
|
|
doAssert f[].handle.isSome()
|
|
|
|
f[].handle.get().skipRecord()
|
|
|
|
proc getBlockHeader(f: Era1File, res: var headers.Header): Result[void, string] =
|
|
var bytes: seq[byte]
|
|
|
|
let header = ?f[].handle.get().readRecord(bytes)
|
|
if header.typ != CompressedHeader:
|
|
return err("Invalid era file: didn't find block header at index position")
|
|
|
|
fromCompressedRlpBytes(bytes, res)
|
|
|
|
proc getBlockBody(f: Era1File, res: var BlockBody): Result[void, string] =
|
|
var bytes: seq[byte]
|
|
|
|
let header = ?f[].handle.get().readRecord(bytes)
|
|
if header.typ != CompressedBody:
|
|
return err("Invalid era file: didn't find block body at index position")
|
|
|
|
fromCompressedRlpBytes(bytes, res)
|
|
|
|
proc getReceipts(f: Era1File, res: var seq[Receipt]): Result[void, string] =
|
|
var bytes: seq[byte]
|
|
|
|
let header = ?f[].handle.get().readRecord(bytes)
|
|
if header.typ != CompressedReceipts:
|
|
return err("Invalid era file: didn't find receipts at index position")
|
|
|
|
fromCompressedRlpBytes(bytes, res)
|
|
|
|
proc getTotalDifficulty(f: Era1File): Result[UInt256, string] =
|
|
var bytes: seq[byte]
|
|
|
|
let header = ?f[].handle.get().readRecord(bytes)
|
|
if header.typ != TotalDifficulty:
|
|
return err("Invalid era file: didn't find total difficulty at index position")
|
|
|
|
if bytes.len != 32:
|
|
return err("Invalid total difficulty length")
|
|
|
|
ok(UInt256.fromBytesLE(bytes))
|
|
|
|
proc getNextEthBlock*(f: Era1File, res: var Block): Result[void, string] =
|
|
doAssert not isNil(f) and f[].handle.isSome
|
|
|
|
var body: BlockBody
|
|
?getBlockHeader(f, res.header)
|
|
?getBlockBody(f, body)
|
|
|
|
?skipRecord(f) # receipts
|
|
?skipRecord(f) # totalDifficulty
|
|
|
|
res.transactions = move(body.transactions)
|
|
res.uncles = move(body.uncles)
|
|
res.withdrawals = move(body.withdrawals)
|
|
|
|
ok()
|
|
|
|
proc getEthBlock*(
|
|
f: Era1File, blockNumber: uint64, res: var Block
|
|
): Result[void, string] =
|
|
doAssert not isNil(f) and f[].handle.isSome
|
|
doAssert(
|
|
blockNumber >= f[].blockIdx.startNumber and blockNumber <= f[].blockIdx.endNumber,
|
|
"Wrong era1 file for selected block number",
|
|
)
|
|
|
|
let pos = f[].blockIdx.offsets[blockNumber - f[].blockIdx.startNumber]
|
|
|
|
?f[].handle.get().setFilePos(pos, SeekPosition.SeekBegin).mapErr(ioErrorMsg)
|
|
|
|
getNextEthBlock(f, res)
|
|
|
|
proc getNextBlockTuple*(f: Era1File, res: var BlockTuple): Result[void, string] =
|
|
doAssert not isNil(f) and f[].handle.isSome
|
|
|
|
?getBlockHeader(f, res.header)
|
|
?getBlockBody(f, res.body)
|
|
?getReceipts(f, res.receipts)
|
|
res.td = ?getTotalDifficulty(f)
|
|
|
|
ok()
|
|
|
|
proc getBlockTuple*(
|
|
f: Era1File, blockNumber: uint64, res: var BlockTuple
|
|
): Result[void, string] =
|
|
doAssert not isNil(f) and f[].handle.isSome
|
|
doAssert(
|
|
blockNumber >= f[].blockIdx.startNumber and blockNumber <= f[].blockIdx.endNumber,
|
|
"Wrong era1 file for selected block number",
|
|
)
|
|
|
|
let pos = f[].blockIdx.offsets[blockNumber - f[].blockIdx.startNumber]
|
|
|
|
?f[].handle.get().setFilePos(pos, SeekPosition.SeekBegin).mapErr(ioErrorMsg)
|
|
|
|
getNextBlockTuple(f, res)
|
|
|
|
proc getBlockHeader*(
|
|
f: Era1File, blockNumber: uint64, res: var headers.Header
|
|
): Result[void, string] =
|
|
doAssert not isNil(f) and f[].handle.isSome
|
|
doAssert(
|
|
blockNumber >= f[].blockIdx.startNumber and blockNumber <= f[].blockIdx.endNumber,
|
|
"Wrong era1 file for selected block number",
|
|
)
|
|
|
|
let pos = f[].blockIdx.offsets[blockNumber - f[].blockIdx.startNumber]
|
|
|
|
?f[].handle.get().setFilePos(pos, SeekPosition.SeekBegin).mapErr(ioErrorMsg)
|
|
|
|
getBlockHeader(f, res)
|
|
|
|
proc getTotalDifficulty*(f: Era1File, blockNumber: uint64): Result[UInt256, string] =
|
|
doAssert not isNil(f) and f[].handle.isSome
|
|
doAssert(
|
|
blockNumber >= f[].blockIdx.startNumber and blockNumber <= f[].blockIdx.endNumber,
|
|
"Wrong era1 file for selected block number",
|
|
)
|
|
|
|
let pos = f[].blockIdx.offsets[blockNumber - f[].blockIdx.startNumber]
|
|
|
|
?f[].handle.get().setFilePos(pos, SeekPosition.SeekBegin).mapErr(ioErrorMsg)
|
|
|
|
?skipRecord(f) # Header
|
|
?skipRecord(f) # BlockBody
|
|
?skipRecord(f) # Receipts
|
|
getTotalDifficulty(f)
|
|
|
|
# TODO: Should we add this perhaps in the Era1File object and grab it in open()?
|
|
proc getAccumulatorRoot*(f: Era1File): Result[Digest, string] =
|
|
# Get position of BlockIndex
|
|
?f[].handle.get().setFilePos(0, SeekPosition.SeekEnd).mapErr(ioErrorMsg)
|
|
let blockIdxPos = ?f[].handle.get().findIndexStartOffset()
|
|
|
|
# Accumulator root is 40 bytes before the BlockIndex
|
|
let accumulatorRootPos = blockIdxPos - 40 # 8 + 32
|
|
?f[].handle.get().setFilePos(accumulatorRootPos, SeekPosition.SeekCurrent).mapErr(
|
|
ioErrorMsg
|
|
)
|
|
|
|
var bytes: seq[byte]
|
|
let header = ?f[].handle.get().readRecord(bytes)
|
|
|
|
if header.typ != AccumulatorRoot:
|
|
return err("Invalid era file: didn't find accumulator root at index position")
|
|
|
|
if bytes.len != 32:
|
|
return err("invalid accumulator root")
|
|
|
|
ok(Digest(data: array[32, byte].initCopyFrom(bytes)))
|
|
|
|
proc buildAccumulator*(f: Era1File): Result[EpochRecordCached, string] =
|
|
let
|
|
startNumber = f.blockIdx.startNumber
|
|
endNumber = f.blockIdx.endNumber()
|
|
|
|
var headerRecords: seq[HeaderRecord]
|
|
var header: headers.Header
|
|
for blockNumber in startNumber .. endNumber:
|
|
?f.getBlockHeader(blockNumber, header)
|
|
let totalDifficulty = ?f.getTotalDifficulty(blockNumber)
|
|
|
|
headerRecords.add(
|
|
HeaderRecord(blockHash: header.rlpHash(), totalDifficulty: totalDifficulty)
|
|
)
|
|
|
|
ok(EpochRecordCached.init(headerRecords))
|
|
|
|
proc verify*(f: Era1File): Result[Digest, string] =
|
|
let
|
|
startNumber = f.blockIdx.startNumber
|
|
endNumber = f.blockIdx.endNumber()
|
|
|
|
var headerRecords: seq[HeaderRecord]
|
|
var blockTuple: BlockTuple
|
|
for blockNumber in startNumber .. endNumber:
|
|
?f.getBlockTuple(blockNumber, blockTuple)
|
|
let
|
|
txRoot = calcTxRoot(blockTuple.body.transactions)
|
|
ommershHash = rlpHash(blockTuple.body.uncles)
|
|
|
|
if blockTuple.header.txRoot != txRoot:
|
|
return err("Invalid transactions root")
|
|
|
|
if blockTuple.header.ommersHash != ommershHash:
|
|
return err("Invalid ommers hash")
|
|
|
|
if blockTuple.header.receiptsRoot != calcReceiptsRoot(blockTuple.receipts):
|
|
return err("Invalid receipts root")
|
|
|
|
headerRecords.add(
|
|
HeaderRecord(
|
|
blockHash: blockTuple.header.rlpHash(), totalDifficulty: blockTuple.td
|
|
)
|
|
)
|
|
|
|
let expectedRoot = ?f.getAccumulatorRoot()
|
|
let accumulatorRoot = getEpochRecordRoot(headerRecords)
|
|
|
|
if accumulatorRoot != expectedRoot:
|
|
err("Invalid accumulator root")
|
|
else:
|
|
ok(accumulatorRoot)
|
|
|
|
iterator era1BlockHeaders*(f: Era1File): headers.Header =
|
|
let
|
|
startNumber = f.blockIdx.startNumber
|
|
endNumber = f.blockIdx.endNumber()
|
|
|
|
var header: headers.Header
|
|
for blockNumber in startNumber .. endNumber:
|
|
f.getBlockHeader(blockNumber, header).expect("Header can be read")
|
|
yield header
|
|
|
|
iterator era1BlockTuples*(f: Era1File): BlockTuple =
|
|
let
|
|
startNumber = f.blockIdx.startNumber
|
|
endNumber = f.blockIdx.endNumber()
|
|
|
|
var blockTuple: BlockTuple
|
|
for blockNumber in startNumber .. endNumber:
|
|
f.getBlockTuple(blockNumber, blockTuple).expect("Block tuple can be read")
|
|
yield blockTuple
|