314 lines
12 KiB
Nim
314 lines
12 KiB
Nim
# beacon_chain
|
|
# Copyright (c) 2022-2024 Status Research & Development GmbH
|
|
# Licensed and distributed under either of
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
{.push raises: [].}
|
|
|
|
import std/macros,
|
|
results, stew/byteutils, presto/route,
|
|
../spec/[forks],
|
|
../spec/eth2_apis/[rest_types, eth2_rest_serialization, rest_common],
|
|
../validators/beacon_validators,
|
|
../consensus_object_pools/blockchain_dag,
|
|
../beacon_node,
|
|
"."/[rest_constants, state_ttl_cache]
|
|
|
|
export
|
|
results, eth2_rest_serialization, blockchain_dag, rest_types,
|
|
rest_constants, rest_common, route
|
|
|
|
proc getSyncedHead*(
|
|
node: BeaconNode,
|
|
slot: Slot
|
|
): Result[BlockRef, cstring] =
|
|
let head = node.dag.head
|
|
|
|
if not node.isSynced(head):
|
|
return err("Beacon node not fully and non-optimistically synced")
|
|
|
|
# Enough ahead not to know the shuffling
|
|
if slot > head.slot + SLOTS_PER_EPOCH * 2:
|
|
return err("Requesting far ahead of the current head")
|
|
|
|
ok(head)
|
|
|
|
func getCurrentSlot*(node: BeaconNode, slot: Slot):
|
|
Result[Slot, cstring] =
|
|
if slot <= (node.dag.head.slot + (SLOTS_PER_EPOCH * 2)):
|
|
ok(slot)
|
|
else:
|
|
err("Requesting slot too far ahead of the current head")
|
|
|
|
proc getSyncedHead*(
|
|
node: BeaconNode,
|
|
epoch: Epoch,
|
|
): Result[BlockRef, cstring] =
|
|
if epoch > MaxEpoch:
|
|
return err("Requesting epoch for which slot would overflow")
|
|
node.getSyncedHead(epoch.start_slot())
|
|
|
|
func getBlockSlotId*(node: BeaconNode,
|
|
stateIdent: StateIdent): Result[BlockSlotId, cstring] =
|
|
case stateIdent.kind
|
|
of StateQueryKind.Slot:
|
|
# Limit requests by state id to the next epoch with respect to the current
|
|
# head to avoid long empty slot replays (in particular a second epoch
|
|
# transition)
|
|
if stateIdent.slot.epoch > (node.dag.head.slot.epoch + 1):
|
|
return err("Requesting state too far ahead of current head")
|
|
|
|
let bsi = node.dag.getBlockIdAtSlot(stateIdent.slot).valueOr:
|
|
return err("History for given slot not available")
|
|
|
|
ok(bsi)
|
|
|
|
of StateQueryKind.Root:
|
|
if stateIdent.root == getStateRoot(node.dag.headState):
|
|
ok(node.dag.head.bid.atSlot())
|
|
else:
|
|
# The `state_roots` field holds 8k historical state roots but not the
|
|
# one of the current state - this trick allows us to lookup states without
|
|
# keeping an on-disk index.
|
|
let headSlot = getStateField(node.dag.headState, slot)
|
|
for i in 0'u64..<SLOTS_PER_HISTORICAL_ROOT:
|
|
if i >= headSlot:
|
|
break
|
|
if getStateField(node.dag.headState, state_roots).item(
|
|
(headSlot - i - 1) mod SLOTS_PER_HISTORICAL_ROOT) ==
|
|
stateIdent.root:
|
|
return node.dag.getBlockIdAtSlot(headSlot - i - 1).orErr(
|
|
cstring("History for for given root not available"))
|
|
|
|
# We don't have a state root -> BlockSlot mapping
|
|
err("State root not found - use by-slot lookup to query deep state history")
|
|
of StateQueryKind.Named:
|
|
case stateIdent.value
|
|
of StateIdentType.Head:
|
|
ok(node.dag.head.bid.atSlot())
|
|
of StateIdentType.Genesis:
|
|
let bid = node.dag.getBlockIdAtSlot(GENESIS_SLOT).valueOr:
|
|
return err("Genesis state not available / pruned")
|
|
ok bid
|
|
of StateIdentType.Finalized:
|
|
ok(node.dag.finalizedHead.toBlockSlotId().expect("not nil"))
|
|
of StateIdentType.Justified:
|
|
# Take checkpoint-synced nodes into account
|
|
let justifiedEpoch =
|
|
max(
|
|
getStateField(node.dag.headState, current_justified_checkpoint).epoch,
|
|
node.dag.finalizedHead.slot.epoch)
|
|
ok(node.dag.head.atEpochStart(justifiedEpoch).toBlockSlotId().expect("not nil"))
|
|
|
|
proc getBlockId*(node: BeaconNode, id: BlockIdent): Opt[BlockId] =
|
|
case id.kind
|
|
of BlockQueryKind.Named:
|
|
case id.value
|
|
of BlockIdentType.Head:
|
|
ok(node.dag.head.bid)
|
|
of BlockIdentType.Genesis:
|
|
node.dag.getBlockIdAtSlot(GENESIS_SLOT).map(proc(x: auto): auto = x.bid)
|
|
of BlockIdentType.Finalized:
|
|
ok(node.dag.finalizedHead.blck.bid)
|
|
of BlockQueryKind.Root:
|
|
node.dag.getBlockId(id.root)
|
|
of BlockQueryKind.Slot:
|
|
let bsid = node.dag.getBlockIdAtSlot(id.slot)
|
|
if bsid.isSome and bsid.get().isProposed():
|
|
ok bsid.get().bid
|
|
else:
|
|
err()
|
|
|
|
proc getForkedBlock*(node: BeaconNode, id: BlockIdent):
|
|
Opt[ForkedTrustedSignedBeaconBlock] =
|
|
let bid = ? node.getBlockId(id)
|
|
|
|
node.dag.getForkedBlock(bid)
|
|
|
|
func disallowInterruptionsAux(body: NimNode) =
|
|
for n in body:
|
|
const because =
|
|
"because the `state` variable may be mutated (and thus invalidated) " &
|
|
"before the function resumes execution."
|
|
|
|
if n.kind == nnkYieldStmt:
|
|
macros.error "You cannot use yield in this block " & because, n
|
|
|
|
if (n.kind in {nnkCall, nnkCommand} and
|
|
n[0].kind in {nnkIdent, nnkSym} and
|
|
$n[0] == "await"):
|
|
macros.error "You cannot use await in this block " & because, n
|
|
|
|
disallowInterruptionsAux(n)
|
|
|
|
macro disallowInterruptions(body: untyped) =
|
|
disallowInterruptionsAux(body)
|
|
|
|
template withStateForBlockSlotId*(nodeParam: BeaconNode,
|
|
blockSlotIdParam: BlockSlotId,
|
|
body: untyped): untyped =
|
|
|
|
block:
|
|
let
|
|
node = nodeParam
|
|
blockSlotId = blockSlotIdParam
|
|
|
|
template isState(state: ForkedHashedBeaconState): bool =
|
|
state.matches_block_slot(blockSlotId.bid.root, blockSlotId.slot)
|
|
|
|
var cache {.inject, used.}: StateCache
|
|
|
|
# If we have a cache hit, there is a concern that the REST request
|
|
# handler may continue executing asynchronously while we hit the same
|
|
# advanced state is another request. We don't want the two requests
|
|
# to work over the same state object because mutations to it will be
|
|
# visible in both, so we must outlaw yielding within the `body` block.
|
|
# Please note that the problem is not limited to the situations where
|
|
# we have a cache hit. Working with the `headState` will result in the
|
|
# same problem as it may change while the request is executing.
|
|
#
|
|
# TODO
|
|
# The solution below is only partion, because it theory yields or awaits
|
|
# can still be hidden in the body through the use of helper templates
|
|
disallowInterruptions(body)
|
|
|
|
# TODO view-types
|
|
# Avoid the code bloat produced by the double `body` reference through a lent var
|
|
if isState(node.dag.headState):
|
|
template state: untyped {.inject, used.} = node.dag.headState
|
|
template stateRoot: untyped {.inject, used.} =
|
|
getStateRoot(node.dag.headState)
|
|
body
|
|
else:
|
|
let cachedState = if node.stateTtlCache != nil:
|
|
node.stateTtlCache.getClosestState(node.dag, blockSlotId)
|
|
else:
|
|
nil
|
|
|
|
let stateToAdvance = if cachedState != nil:
|
|
cachedState
|
|
else:
|
|
assignClone(node.dag.headState)
|
|
|
|
if node.dag.updateState(stateToAdvance[], blockSlotId, false, cache,
|
|
node.dag.updateFlags):
|
|
if cachedState == nil and node.stateTtlCache != nil:
|
|
# This was not a cached state, we can cache it now
|
|
node.stateTtlCache.add(stateToAdvance)
|
|
|
|
template state: untyped {.inject, used.} = stateToAdvance[]
|
|
template stateRoot: untyped {.inject, used.} = getStateRoot(stateToAdvance[])
|
|
|
|
body
|
|
|
|
template strData*(body: ContentBody): string =
|
|
bind fromBytes
|
|
string.fromBytes(body.data)
|
|
|
|
func syncCommitteeParticipants*(forkedState: ForkedHashedBeaconState,
|
|
epoch: Epoch
|
|
): Result[seq[ValidatorPubKey], cstring] =
|
|
withState(forkedState):
|
|
when consensusFork >= ConsensusFork.Altair:
|
|
let
|
|
epochPeriod = sync_committee_period(epoch)
|
|
curPeriod = sync_committee_period(forkyState.data.slot)
|
|
if epochPeriod == curPeriod:
|
|
ok(@(forkyState.data.current_sync_committee.pubkeys.data))
|
|
elif epochPeriod == curPeriod + 1:
|
|
ok(@(forkyState.data.next_sync_committee.pubkeys.data))
|
|
else:
|
|
err("Epoch is outside the sync committee period of the state")
|
|
else:
|
|
err("State's fork do not support sync committees")
|
|
|
|
func keysToIndices*(cacheTable: var Table[ValidatorPubKey, ValidatorIndex],
|
|
forkedState: ForkedHashedBeaconState,
|
|
keys: openArray[ValidatorPubKey]
|
|
): seq[Opt[ValidatorIndex]] =
|
|
var indices = newSeq[Opt[ValidatorIndex]](len(keys))
|
|
let totalValidatorsInState = getStateField(forkedState, validators).lenu64
|
|
var keyset =
|
|
block:
|
|
var res: Table[ValidatorPubKey, int]
|
|
for inputIndex, pubkey in keys:
|
|
# Try to search in cache first.
|
|
cacheTable.withValue(pubkey, vindex):
|
|
if uint64(vindex[]) < totalValidatorsInState:
|
|
indices[inputIndex] = Opt.some(vindex[])
|
|
do:
|
|
res[pubkey] = inputIndex
|
|
res
|
|
if len(keyset) > 0:
|
|
for validatorIndex, validator in getStateField(forkedState, validators):
|
|
keyset.withValue(validator.pubkey, listIndex):
|
|
# Store pair (pubkey, index) into cache table.
|
|
cacheTable[validator.pubkey] = ValidatorIndex(validatorIndex)
|
|
# Fill result sequence.
|
|
indices[listIndex[]] = Opt.some(ValidatorIndex(validatorIndex))
|
|
indices
|
|
|
|
proc getBidOptimistic*(node: BeaconNode, bid: BlockId): Opt[bool] =
|
|
if node.currentSlot().epoch() >= node.dag.cfg.BELLATRIX_FORK_EPOCH:
|
|
Opt.some(node.dag.is_optimistic(bid))
|
|
else:
|
|
Opt.none(bool)
|
|
|
|
proc getShufflingOptimistic*(node: BeaconNode,
|
|
dependentSlot: Slot,
|
|
dependentRoot: Eth2Digest): Opt[bool] =
|
|
if node.currentSlot().epoch() >= node.dag.cfg.BELLATRIX_FORK_EPOCH:
|
|
# `slot` in this `BlockId` may be higher than block's actual slot,
|
|
# this is alright for the purpose of calling `is_optimistic`.
|
|
let bid = BlockId(slot: dependentSlot, root: dependentRoot)
|
|
Opt.some(node.dag.is_optimistic(bid))
|
|
else:
|
|
Opt.none(bool)
|
|
|
|
proc getStateOptimistic*(node: BeaconNode,
|
|
state: ForkedHashedBeaconState): Opt[bool] =
|
|
if node.currentSlot().epoch() >= node.dag.cfg.BELLATRIX_FORK_EPOCH:
|
|
if state.kind >= ConsensusFork.Bellatrix:
|
|
# A state is optimistic iff the block which created it is
|
|
let stateBid = withState(state): forkyState.latest_block_id
|
|
Opt.some(node.dag.is_optimistic(stateBid))
|
|
else:
|
|
Opt.some(false)
|
|
else:
|
|
Opt.none(bool)
|
|
|
|
proc getBlockOptimistic*(node: BeaconNode,
|
|
blck: ForkedTrustedSignedBeaconBlock |
|
|
ForkedSignedBeaconBlock): Opt[bool] =
|
|
if node.currentSlot().epoch() >= node.dag.cfg.BELLATRIX_FORK_EPOCH:
|
|
if blck.kind >= ConsensusFork.Bellatrix:
|
|
Opt.some(node.dag.is_optimistic(blck.toBlockId()))
|
|
else:
|
|
Opt.some(false)
|
|
else:
|
|
Opt.none(bool)
|
|
|
|
const
|
|
jsonMediaType* = MediaType.init("application/json")
|
|
sszMediaType* = MediaType.init("application/octet-stream")
|
|
textEventStreamMediaType* = MediaType.init("text/event-stream")
|
|
|
|
proc verifyRandao*(
|
|
node: BeaconNode, slot: Slot, proposer: ValidatorIndex,
|
|
randao: ValidatorSig, skip_randao_verification: bool): bool =
|
|
let
|
|
proposer_pubkey = node.dag.validatorKey(proposer).valueOr:
|
|
return false
|
|
|
|
if skip_randao_verification:
|
|
randao == ValidatorSig.infinity()
|
|
else:
|
|
let
|
|
fork = node.dag.forkAtEpoch(slot.epoch)
|
|
genesis_validators_root = node.dag.genesis_validators_root
|
|
|
|
verify_epoch_signature(
|
|
fork, genesis_validators_root, slot.epoch, proposer_pubkey, randao)
|