2020-05-19 14:18:07 +00:00
|
|
|
# beacon_chain
|
2021-02-25 13:37:22 +00:00
|
|
|
# Copyright (c) 2018-2021 Status Research & Development GmbH
|
2020-05-19 14:18:07 +00:00
|
|
|
# Licensed and distributed under either of
|
|
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
2020-06-16 05:45:04 +00:00
|
|
|
{.push raises: [Defect].}
|
|
|
|
|
2020-05-19 14:18:07 +00:00
|
|
|
import
|
2020-10-20 12:31:20 +00:00
|
|
|
std/tables,
|
2020-07-28 13:54:32 +00:00
|
|
|
chronicles,
|
2020-12-16 08:37:22 +00:00
|
|
|
stew/[assign2, results],
|
2021-01-25 18:45:48 +00:00
|
|
|
eth/keys,
|
2021-03-05 13:12:00 +00:00
|
|
|
../extras, ../beacon_clock,
|
2021-06-11 17:51:46 +00:00
|
|
|
../spec/[
|
2021-06-21 08:35:24 +00:00
|
|
|
crypto, digest, forkedbeaconstate_helpers, helpers, signatures,
|
2021-06-11 17:51:46 +00:00
|
|
|
signatures_batch, state_transition],
|
2021-06-21 08:35:24 +00:00
|
|
|
../spec/datatypes/[phase0, altair],
|
2021-03-04 09:13:44 +00:00
|
|
|
./block_pools_types, ./blockchain_dag, ./block_quarantine
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
from libp2p/protocols/pubsub/pubsub import ValidationResult
|
|
|
|
|
|
|
|
export results, ValidationResult
|
2020-05-21 17:08:31 +00:00
|
|
|
|
2020-05-19 14:18:07 +00:00
|
|
|
# Clearance
|
|
|
|
# ---------------------------------------------
|
|
|
|
#
|
|
|
|
# This module is in charge of making the
|
|
|
|
# "quarantined" network blocks
|
2020-07-30 19:18:17 +00:00
|
|
|
# pass the firewall and be stored in the chain DAG
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2020-06-16 05:45:04 +00:00
|
|
|
logScope:
|
|
|
|
topics = "clearance"
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-06-21 08:35:24 +00:00
|
|
|
## At the GC-level, the GC is type-agnostic; it's all type-erased so
|
|
|
|
## casting between seq[Attestation] and seq[TrustedAttestation] will
|
|
|
|
## not disrupt GC operations.
|
|
|
|
##
|
|
|
|
## These SHOULD be used in function calls to avoid expensive temporary.
|
|
|
|
## see https://github.com/status-im/nimbus-eth2/pull/2250#discussion_r562010679
|
|
|
|
template asSigVerified(x: phase0.SignedBeaconBlock):
|
|
|
|
phase0.SigVerifiedSignedBeaconBlock =
|
|
|
|
## This converts a signed beacon block to a sig verified beacon clock.
|
|
|
|
## This assumes that their bytes representation is the same.
|
|
|
|
static: # TODO See isomorphicCast
|
|
|
|
doAssert sizeof(phase0.SignedBeaconBlock) ==
|
|
|
|
sizeof(phase0.SigVerifiedSignedBeaconBlock)
|
|
|
|
|
|
|
|
cast[ptr phase0.SigVerifiedSignedBeaconBlock](signedBlock.unsafeAddr)[]
|
|
|
|
|
|
|
|
template asSigVerified(x: altair.SignedBeaconBlock):
|
|
|
|
altair.SigVerifiedSignedBeaconBlock =
|
2021-01-25 18:45:48 +00:00
|
|
|
## This converts a signed beacon block to a sig verified beacon clock.
|
|
|
|
## This assumes that their bytes representation is the same.
|
2021-05-29 18:56:30 +00:00
|
|
|
static: # TODO See isomorphicCast
|
2021-06-21 08:35:24 +00:00
|
|
|
doAssert sizeof(altair.SignedBeaconBlock) ==
|
|
|
|
sizeof(altair.SigVerifiedSignedBeaconBlock)
|
2021-05-29 18:56:30 +00:00
|
|
|
|
2021-06-21 08:35:24 +00:00
|
|
|
cast[ptr altair.SigVerifiedSignedBeaconBlock](signedBlock.unsafeAddr)[]
|
2021-01-25 18:45:48 +00:00
|
|
|
|
2021-06-21 08:35:24 +00:00
|
|
|
template asTrusted(x: phase0.SignedBeaconBlock or phase0.SigVerifiedBeaconBlock):
|
|
|
|
phase0.TrustedSignedBeaconBlock =
|
|
|
|
## This converts a sigverified beacon block to a trusted beacon clock.
|
|
|
|
## This assumes that their bytes representation is the same.
|
|
|
|
static: # TODO See isomorphicCast
|
|
|
|
doAssert sizeof(x) == sizeof(phase0.TrustedSignedBeaconBlock)
|
|
|
|
|
|
|
|
cast[ptr phase0.TrustedSignedBeaconBlock](signedBlock.unsafeAddr)[]
|
|
|
|
|
|
|
|
template asTrusted(x: altair.SignedBeaconBlock or altair.SigVerifiedBeaconBlock):
|
|
|
|
altair.TrustedSignedBeaconBlock =
|
2021-01-25 18:45:48 +00:00
|
|
|
## This converts a sigverified beacon block to a trusted beacon clock.
|
|
|
|
## This assumes that their bytes representation is the same.
|
2021-05-29 18:56:30 +00:00
|
|
|
static: # TODO See isomorphicCast
|
2021-06-21 08:35:24 +00:00
|
|
|
doAssert sizeof(x) == sizeof(altair.TrustedSignedBeaconBlock)
|
2021-05-29 18:56:30 +00:00
|
|
|
|
2021-06-21 08:35:24 +00:00
|
|
|
cast[ptr altair.TrustedSignedBeaconBlock](signedBlock.unsafeAddr)[]
|
2021-01-25 18:45:48 +00:00
|
|
|
|
2021-06-03 09:42:25 +00:00
|
|
|
func batchVerify(quarantine: QuarantineRef, sigs: openArray[SignatureSet]): bool =
|
2021-01-25 18:45:48 +00:00
|
|
|
var secureRandomBytes: array[32, byte]
|
|
|
|
quarantine.rng[].brHmacDrbgGenerate(secureRandomBytes)
|
|
|
|
|
|
|
|
# TODO: For now only enable serial batch verification
|
|
|
|
return batchVerifySerial(quarantine.sigVerifCache, sigs, secureRandomBytes)
|
|
|
|
|
2020-07-09 09:29:32 +00:00
|
|
|
proc addRawBlock*(
|
2021-06-01 11:13:40 +00:00
|
|
|
dag: ChainDAGRef, quarantine: QuarantineRef,
|
2021-06-23 14:43:18 +00:00
|
|
|
signedBlock: phase0.SignedBeaconBlock | altair.SignedBeaconBlock,
|
|
|
|
onBlockAdded: OnPhase0BlockAdded | OnAltairBlockAdded
|
2020-09-18 11:53:09 +00:00
|
|
|
): Result[BlockRef, (ValidationResult, BlockError)] {.gcsafe.}
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-06-23 14:43:18 +00:00
|
|
|
# Now that we have the new block, we should see if any of the previously
|
|
|
|
# unresolved blocks magically become resolved
|
|
|
|
# TODO This code is convoluted because when there are more than ~1.5k
|
|
|
|
# blocks being synced, there's a stack overflow as `add` gets called
|
|
|
|
# for the whole chain of blocks. Instead we use this ugly field in `dag`
|
|
|
|
# which could be avoided by refactoring the code
|
|
|
|
# TODO unit test the logic, in particular interaction with fork choice block parents
|
|
|
|
proc resolveQuarantinedBlocks(
|
|
|
|
dag: ChainDAGRef, quarantine: QuarantineRef,
|
|
|
|
onBlockAdded: OnPhase0BlockAdded) =
|
|
|
|
if not quarantine.inAdd:
|
|
|
|
quarantine.inAdd = true
|
|
|
|
defer: quarantine.inAdd = false
|
|
|
|
var entries = 0
|
|
|
|
while entries != quarantine.orphansPhase0.len:
|
|
|
|
entries = quarantine.orphansPhase0.len # keep going while quarantine is shrinking
|
|
|
|
var resolved: seq[phase0.SignedBeaconBlock]
|
|
|
|
for _, v in quarantine.orphansPhase0:
|
|
|
|
if v.message.parent_root in dag:
|
|
|
|
resolved.add(v)
|
|
|
|
|
|
|
|
for v in resolved:
|
|
|
|
discard addRawBlock(dag, quarantine, v, onBlockAdded)
|
|
|
|
|
|
|
|
proc resolveQuarantinedBlocks(
|
|
|
|
dag: ChainDAGRef, quarantine: QuarantineRef,
|
|
|
|
onBlockAdded: OnAltairBlockAdded) =
|
|
|
|
if not quarantine.inAdd:
|
|
|
|
quarantine.inAdd = true
|
|
|
|
defer: quarantine.inAdd = false
|
|
|
|
var entries = 0
|
|
|
|
while entries != quarantine.orphansAltair.len:
|
|
|
|
entries = quarantine.orphansAltair.len # keep going while quarantine is shrinking
|
|
|
|
var resolved: seq[altair.SignedBeaconBlock]
|
|
|
|
for _, v in quarantine.orphansAltair:
|
|
|
|
if v.message.parent_root in dag:
|
|
|
|
resolved.add(v)
|
|
|
|
|
|
|
|
for v in resolved:
|
|
|
|
discard addRawBlock(dag, quarantine, v, onBlockAdded)
|
|
|
|
|
2020-05-19 14:18:07 +00:00
|
|
|
proc addResolvedBlock(
|
2021-06-01 11:13:40 +00:00
|
|
|
dag: ChainDAGRef, quarantine: QuarantineRef,
|
2021-06-23 14:43:18 +00:00
|
|
|
state: var StateData,
|
|
|
|
trustedBlock: phase0.TrustedSignedBeaconBlock | altair.TrustedSignedBeaconBlock,
|
2020-08-05 06:28:43 +00:00
|
|
|
parent: BlockRef, cache: var StateCache,
|
2021-06-23 14:43:18 +00:00
|
|
|
onBlockAdded: OnPhase0BlockAdded | OnAltairBlockAdded,
|
|
|
|
stateDataDur, sigVerifyDur,
|
2021-05-28 19:03:20 +00:00
|
|
|
stateVerifyDur: Duration
|
2020-08-18 20:29:33 +00:00
|
|
|
) =
|
2021-06-11 17:51:46 +00:00
|
|
|
doAssert getStateField(state.data, slot) == trustedBlock.message.slot,
|
2020-08-18 20:29:33 +00:00
|
|
|
"state must match block"
|
2021-01-25 18:45:48 +00:00
|
|
|
doAssert state.blck.root == trustedBlock.message.parent_root,
|
2020-08-18 20:29:33 +00:00
|
|
|
"the StateData passed into the addResolved function not yet updated!"
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2020-07-16 13:16:51 +00:00
|
|
|
let
|
2021-01-25 18:45:48 +00:00
|
|
|
blockRoot = trustedBlock.root
|
|
|
|
blockRef = BlockRef.init(blockRoot, trustedBlock.message)
|
2021-05-28 16:34:00 +00:00
|
|
|
startTick = Moment.now()
|
2020-08-11 19:39:53 +00:00
|
|
|
|
|
|
|
link(parent, blockRef)
|
|
|
|
|
2021-03-17 10:17:15 +00:00
|
|
|
dag.blocks.incl(KeyedBlockRef.init(blockRef))
|
2020-05-19 14:18:07 +00:00
|
|
|
|
|
|
|
# Resolved blocks should be stored in database
|
2021-06-24 18:34:08 +00:00
|
|
|
dag.putBlock(trustedBlock)
|
2021-05-28 19:03:20 +00:00
|
|
|
let putBlockTick = Moment.now()
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-05-29 18:56:30 +00:00
|
|
|
var foundHead: bool
|
2020-05-19 14:18:07 +00:00
|
|
|
for head in dag.heads.mitems():
|
2020-07-28 13:54:32 +00:00
|
|
|
if head.isAncestorOf(blockRef):
|
|
|
|
head = blockRef
|
2021-05-29 18:56:30 +00:00
|
|
|
foundHead = true
|
2020-05-19 14:18:07 +00:00
|
|
|
break
|
|
|
|
|
2021-05-29 18:56:30 +00:00
|
|
|
if not foundHead:
|
|
|
|
dag.heads.add(blockRef)
|
|
|
|
|
|
|
|
# Up to here, state.data was referring to the new state after the block had
|
|
|
|
# been applied but the `blck` field was still set to the parent
|
|
|
|
state.blck = blockRef
|
|
|
|
|
2021-06-01 11:13:40 +00:00
|
|
|
# Regardless of the chain we're on, the deposits come in the same order so
|
|
|
|
# as soon as we import a block, we'll also update the shared public key
|
|
|
|
# cache
|
|
|
|
|
2021-06-11 17:51:46 +00:00
|
|
|
dag.updateValidatorKeys(getStateField(state.data, validators).asSeq())
|
2021-06-01 11:13:40 +00:00
|
|
|
|
2021-05-29 18:56:30 +00:00
|
|
|
# Getting epochRef with the state will potentially create a new EpochRef
|
|
|
|
let
|
|
|
|
epochRef = dag.getEpochRef(state, cache)
|
|
|
|
epochRefTick = Moment.now()
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2020-10-01 18:56:42 +00:00
|
|
|
debug "Block resolved",
|
2021-01-25 18:45:48 +00:00
|
|
|
blck = shortLog(trustedBlock.message),
|
2020-05-19 14:18:07 +00:00
|
|
|
blockRoot = shortLog(blockRoot),
|
2021-05-28 16:34:00 +00:00
|
|
|
heads = dag.heads.len(),
|
2021-05-28 19:03:20 +00:00
|
|
|
stateDataDur, sigVerifyDur, stateVerifyDur,
|
2021-05-29 18:56:30 +00:00
|
|
|
putBlockDur = putBlockTick - startTick,
|
|
|
|
epochRefDur = epochRefTick - putBlockTick
|
2020-08-18 20:29:33 +00:00
|
|
|
|
2020-07-30 15:48:25 +00:00
|
|
|
# Notify others of the new block before processing the quarantine, such that
|
|
|
|
# notifications for parents happens before those of the children
|
2020-07-22 09:42:55 +00:00
|
|
|
if onBlockAdded != nil:
|
2021-06-21 08:35:24 +00:00
|
|
|
onBlockAdded(blockRef, trustedBlock, epochRef)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2021-06-23 14:43:18 +00:00
|
|
|
resolveQuarantinedBlocks(dag, quarantine, onBlockAdded)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2021-06-21 08:35:24 +00:00
|
|
|
# TODO workaround for https://github.com/nim-lang/Nim/issues/18095
|
2021-06-23 14:43:18 +00:00
|
|
|
type SomeSignedBlock =
|
2021-06-21 08:35:24 +00:00
|
|
|
phase0.SignedBeaconBlock | phase0.SigVerifiedSignedBeaconBlock |
|
2021-06-23 14:43:18 +00:00
|
|
|
phase0.TrustedSignedBeaconBlock |
|
|
|
|
altair.SignedBeaconBlock | altair.SigVerifiedSignedBeaconBlock |
|
|
|
|
altair.TrustedSignedBeaconBlock
|
2021-06-01 11:13:40 +00:00
|
|
|
proc checkStateTransition(
|
2021-06-23 14:43:18 +00:00
|
|
|
dag: ChainDAGRef, signedBlock: SomeSignedBlock,
|
2021-06-01 11:13:40 +00:00
|
|
|
cache: var StateCache): (ValidationResult, BlockError) =
|
|
|
|
## Ensure block can be applied on a state
|
2021-06-11 17:51:46 +00:00
|
|
|
func restore(v: var ForkedHashedBeaconState) =
|
2021-01-25 18:45:48 +00:00
|
|
|
# TODO address this ugly workaround - there should probably be a
|
|
|
|
# `state_transition` that takes a `StateData` instead and updates
|
|
|
|
# the block as well
|
2021-05-05 06:54:21 +00:00
|
|
|
doAssert v.addr == addr dag.clearanceState.data
|
|
|
|
assign(dag.clearanceState, dag.headState)
|
2021-01-25 18:45:48 +00:00
|
|
|
|
2021-05-17 16:37:26 +00:00
|
|
|
logScope:
|
|
|
|
blck = shortLog(signedBlock.message)
|
|
|
|
blockRoot = shortLog(signedBlock.root)
|
|
|
|
|
2021-06-03 09:42:25 +00:00
|
|
|
if not state_transition_block(
|
|
|
|
dag.runtimePreset, dag.clearanceState.data, signedBlock,
|
2021-06-23 14:43:18 +00:00
|
|
|
cache, dag.updateFlags, restore, dag.altairTransitionSlot):
|
2021-01-25 18:45:48 +00:00
|
|
|
info "Invalid block"
|
|
|
|
|
|
|
|
return (ValidationResult.Reject, Invalid)
|
|
|
|
return (ValidationResult.Accept, default(BlockError))
|
|
|
|
|
2021-06-01 11:13:40 +00:00
|
|
|
proc advanceClearanceState*(dag: ChainDagRef) =
|
2021-05-29 18:56:30 +00:00
|
|
|
# When the chain is synced, the most likely block to be produced is the block
|
|
|
|
# right after head - we can exploit this assumption and advance the state
|
|
|
|
# to that slot before the block arrives, thus allowing us to do the expensive
|
|
|
|
# epoch transition ahead of time.
|
|
|
|
# Notably, we use the clearance state here because that's where the block will
|
|
|
|
# first be seen - later, this state will be copied to the head state!
|
2021-06-11 17:51:46 +00:00
|
|
|
if dag.clearanceState.blck.slot == getStateField(dag.clearanceState.data, slot):
|
2021-06-01 11:13:40 +00:00
|
|
|
let next =
|
|
|
|
dag.clearanceState.blck.atSlot(dag.clearanceState.blck.slot + 1)
|
|
|
|
|
2021-06-01 15:33:00 +00:00
|
|
|
let startTick = Moment.now()
|
2021-05-29 18:56:30 +00:00
|
|
|
var cache = StateCache()
|
2021-06-01 11:13:40 +00:00
|
|
|
updateStateData(dag, dag.clearanceState, next, true, cache)
|
2021-05-29 18:56:30 +00:00
|
|
|
|
2021-06-01 15:33:00 +00:00
|
|
|
debug "Prepared clearance state for next block",
|
|
|
|
next, updateStateDur = Moment.now() - startTick
|
|
|
|
|
2021-01-25 18:45:48 +00:00
|
|
|
proc addRawBlockKnownParent(
|
2021-06-01 11:13:40 +00:00
|
|
|
dag: ChainDAGRef, quarantine: QuarantineRef,
|
2021-06-23 14:43:18 +00:00
|
|
|
signedBlock: phase0.SignedBeaconBlock | altair.SignedBeaconBlock,
|
2021-01-25 18:45:48 +00:00
|
|
|
parent: BlockRef,
|
2021-06-23 14:43:18 +00:00
|
|
|
onBlockAdded: OnPhase0BlockAdded | OnAltairBlockAdded
|
2020-09-18 11:53:09 +00:00
|
|
|
): Result[BlockRef, (ValidationResult, BlockError)] =
|
2021-06-01 11:13:40 +00:00
|
|
|
## Add a block whose parent is known, after performing validity checks
|
2021-01-25 18:45:48 +00:00
|
|
|
|
|
|
|
if parent.slot >= signedBlock.message.slot:
|
|
|
|
# A block whose parent is newer than the block itself is clearly invalid -
|
|
|
|
# discard it immediately
|
|
|
|
debug "Invalid block slot",
|
|
|
|
parentBlock = shortLog(parent)
|
|
|
|
|
|
|
|
return err((ValidationResult.Reject, Invalid))
|
|
|
|
|
|
|
|
if (parent.slot < dag.finalizedHead.slot) or
|
|
|
|
(parent.slot == dag.finalizedHead.slot and
|
|
|
|
parent != dag.finalizedHead.blck):
|
|
|
|
# We finalized a block that's newer than the parent of this block - this
|
|
|
|
# block, although recent, is thus building on a history we're no longer
|
|
|
|
# interested in pursuing. This can happen if a client produces a block
|
|
|
|
# while syncing - ie it's own head block will be old, but it'll create
|
|
|
|
# a block according to the wall clock, in its own little world - this is
|
|
|
|
# correct - from their point of view, the head block they have is the
|
|
|
|
# latest thing that happened on the chain and they're performing their
|
|
|
|
# duty correctly.
|
|
|
|
debug "Unviable block, dropping",
|
2020-05-19 14:18:07 +00:00
|
|
|
finalizedHead = shortLog(dag.finalizedHead),
|
2020-07-16 13:16:51 +00:00
|
|
|
tail = shortLog(dag.tail)
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2020-10-20 12:31:20 +00:00
|
|
|
return err((ValidationResult.Ignore, Unviable))
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-01-25 18:45:48 +00:00
|
|
|
# The block might have been in either of `orphans` or `missing` - we don't
|
|
|
|
# want any more work done on its behalf
|
|
|
|
quarantine.removeOrphan(signedBlock)
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-01-25 18:45:48 +00:00
|
|
|
# The block is resolved, now it's time to validate it to ensure that the
|
|
|
|
# blocks we add to the database are clean for the given state
|
2021-05-28 19:03:20 +00:00
|
|
|
let startTick = Moment.now()
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-01-25 18:45:48 +00:00
|
|
|
var cache = StateCache()
|
|
|
|
updateStateData(
|
|
|
|
dag, dag.clearanceState, parent.atSlot(signedBlock.message.slot), true, cache)
|
2021-05-28 19:03:20 +00:00
|
|
|
let stateDataTick = Moment.now()
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-06-01 11:13:40 +00:00
|
|
|
# First, batch-verify all signatures in block
|
2021-01-25 18:45:48 +00:00
|
|
|
if skipBLSValidation notin dag.updateFlags:
|
|
|
|
# TODO: remove skipBLSValidation
|
|
|
|
var sigs: seq[SignatureSet]
|
2021-06-01 11:13:40 +00:00
|
|
|
if sigs.collectSignatureSets(
|
2021-06-11 17:51:46 +00:00
|
|
|
signedBlock, dag.db.immutableValidators, dag.clearanceState.data, cache).isErr():
|
2021-01-25 18:45:48 +00:00
|
|
|
# A PublicKey or Signature isn't on the BLS12-381 curve
|
|
|
|
return err((ValidationResult.Reject, Invalid))
|
|
|
|
if not quarantine.batchVerify(sigs):
|
|
|
|
return err((ValidationResult.Reject, Invalid))
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-05-28 19:03:20 +00:00
|
|
|
let sigVerifyTick = Moment.now()
|
2021-06-01 11:13:40 +00:00
|
|
|
let (valRes, blockErr) = checkStateTransition(
|
|
|
|
dag, signedBlock.asSigVerified(), cache)
|
2021-01-25 18:45:48 +00:00
|
|
|
if valRes != ValidationResult.Accept:
|
|
|
|
return err((valRes, blockErr))
|
2021-05-29 18:56:30 +00:00
|
|
|
|
2021-05-28 19:03:20 +00:00
|
|
|
let stateVerifyTick = Moment.now()
|
2021-01-25 18:45:48 +00:00
|
|
|
# Careful, clearanceState.data has been updated but not blck - we need to
|
|
|
|
# create the BlockRef first!
|
|
|
|
addResolvedBlock(
|
|
|
|
dag, quarantine, dag.clearanceState,
|
|
|
|
signedBlock.asTrusted(),
|
|
|
|
parent, cache,
|
2021-05-28 19:03:20 +00:00
|
|
|
onBlockAdded,
|
|
|
|
stateDataDur = stateDataTick - startTick,
|
|
|
|
sigVerifyDur = sigVerifyTick - stateDataTick,
|
|
|
|
stateVerifyDur = stateVerifyTick - sigVerifyTick)
|
2021-01-25 18:45:48 +00:00
|
|
|
|
|
|
|
return ok dag.clearanceState.blck
|
|
|
|
|
|
|
|
proc addRawBlockUnresolved(
|
2021-06-01 11:13:40 +00:00
|
|
|
dag: ChainDAGRef,
|
|
|
|
quarantine: QuarantineRef,
|
2021-06-23 14:43:18 +00:00
|
|
|
signedBlock: phase0.SignedBeaconBlock | altair.SignedBeaconBlock,
|
2021-01-25 18:45:48 +00:00
|
|
|
): Result[BlockRef, (ValidationResult, BlockError)] =
|
|
|
|
## addRawBlock - Block is unresolved / has no parent
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-05-17 16:37:26 +00:00
|
|
|
logScope:
|
|
|
|
blck = shortLog(signedBlock.message)
|
|
|
|
blockRoot = shortLog(signedBlock.root)
|
|
|
|
|
2020-11-16 19:15:43 +00:00
|
|
|
# This is an unresolved block - add it to the quarantine, which will cause its
|
|
|
|
# parent to be scheduled for downloading
|
2020-08-31 09:00:38 +00:00
|
|
|
if not quarantine.add(dag, signedBlock):
|
|
|
|
debug "Block quarantine full"
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2021-01-25 18:45:48 +00:00
|
|
|
if signedBlock.message.parent_root in quarantine.missing or
|
2020-09-04 06:39:46 +00:00
|
|
|
containsOrphan(quarantine, signedBlock):
|
2020-08-04 19:28:40 +00:00
|
|
|
debug "Unresolved block (parent missing or orphaned)",
|
2021-06-23 14:43:18 +00:00
|
|
|
orphansPhase0 = quarantine.orphansPhase0.len,
|
|
|
|
orphansAltair = quarantine.orphansAltair.len,
|
2020-08-04 19:28:40 +00:00
|
|
|
missing = quarantine.missing.len
|
|
|
|
|
2020-10-20 12:31:20 +00:00
|
|
|
return err((ValidationResult.Ignore, MissingParent))
|
2020-05-19 14:18:07 +00:00
|
|
|
|
|
|
|
# TODO if we receive spam blocks, one heurestic to implement might be to wait
|
|
|
|
# for a couple of attestations to appear before fetching parents - this
|
|
|
|
# would help prevent using up network resources for spam - this serves
|
|
|
|
# two purposes: one is that attestations are likely to appear for the
|
|
|
|
# block only if it's valid / not spam - the other is that malicious
|
|
|
|
# validators that are not proposers can sign invalid blocks and send
|
|
|
|
# them out without penalty - but signing invalid attestations carries
|
|
|
|
# a risk of being slashed, making attestations a more valuable spam
|
|
|
|
# filter.
|
|
|
|
debug "Unresolved block (parent missing)",
|
2021-06-23 14:43:18 +00:00
|
|
|
orphansPhase0 = quarantine.orphansPhase0.len,
|
|
|
|
orphansAltair = quarantine.orphansAltair.len,
|
2020-07-16 13:16:51 +00:00
|
|
|
missing = quarantine.missing.len
|
2020-05-19 14:18:07 +00:00
|
|
|
|
2020-10-20 12:31:20 +00:00
|
|
|
return err((ValidationResult.Ignore, MissingParent))
|
2020-05-21 17:08:31 +00:00
|
|
|
|
2021-06-03 09:42:25 +00:00
|
|
|
proc addRawBlock(
|
2021-06-01 11:13:40 +00:00
|
|
|
dag: ChainDAGRef, quarantine: QuarantineRef,
|
2021-06-23 14:43:18 +00:00
|
|
|
signedBlock: phase0.SignedBeaconBlock | altair.SignedBeaconBlock,
|
|
|
|
onBlockAdded: OnPhase0BlockAdded | OnAltairBlockAdded
|
2021-01-25 18:45:48 +00:00
|
|
|
): Result[BlockRef, (ValidationResult, BlockError)] =
|
|
|
|
## Try adding a block to the chain, verifying first that it passes the state
|
|
|
|
## transition function and contains correct cryptographic signature.
|
|
|
|
##
|
|
|
|
## Cryptographic checks can be skipped by adding skipBLSValidation to dag.updateFlags
|
|
|
|
|
|
|
|
logScope:
|
|
|
|
blck = shortLog(signedBlock.message)
|
|
|
|
blockRoot = shortLog(signedBlock.root)
|
|
|
|
|
|
|
|
template blck(): untyped = signedBlock.message # shortcuts without copy
|
|
|
|
template blockRoot(): untyped = signedBlock.root
|
|
|
|
|
2021-03-17 10:17:15 +00:00
|
|
|
if blockRoot in dag:
|
2021-01-25 18:45:48 +00:00
|
|
|
debug "Block already exists"
|
|
|
|
|
|
|
|
# We should not call the block added callback for blocks that already
|
|
|
|
# existed in the pool, as that may confuse consumers such as the fork
|
|
|
|
# choice. While the validation result won't be accessed, it's IGNORE,
|
|
|
|
# according to the spec.
|
|
|
|
return err((ValidationResult.Ignore, Duplicate))
|
|
|
|
|
|
|
|
quarantine.missing.del(blockRoot)
|
|
|
|
|
|
|
|
# If the block we get is older than what we finalized already, we drop it.
|
|
|
|
# One way this can happen is that we start resolving a block and finalization
|
|
|
|
# happens in the meantime - the block we requested will then be stale
|
|
|
|
# by the time it gets here.
|
|
|
|
if blck.slot <= dag.finalizedHead.slot:
|
|
|
|
debug "Old block, dropping",
|
|
|
|
finalizedHead = shortLog(dag.finalizedHead),
|
|
|
|
tail = shortLog(dag.tail)
|
|
|
|
|
|
|
|
# Doesn't correspond to any specific validation condition, and still won't
|
|
|
|
# be used, but certainly would be IGNORE.
|
|
|
|
return err((ValidationResult.Ignore, Unviable))
|
|
|
|
|
2021-03-17 10:17:15 +00:00
|
|
|
let parent = dag.getRef(blck.parent_root)
|
2021-01-25 18:45:48 +00:00
|
|
|
|
|
|
|
if parent != nil:
|
|
|
|
return addRawBlockKnownParent(dag, quarantine, signedBlock, parent, onBlockAdded)
|
|
|
|
return addRawBlockUnresolved(dag, quarantine, signedBlock)
|