mirror of
https://github.com/status-im/nimbus-eth2.git
synced 2025-02-04 10:43:40 +00:00
do not cache zero block hash if block unavailable (#5865)
With checkpoint sync, the checkpoint block is typically unavailable at the start, and only backfilled later. To avoid treating it as having zero hash, execution disabled in some contexts, wrap the result of `loadExecutionBlockHash` in `Opt` and handle block hash being unknown. --------- Co-authored-by: Jacek Sieka <jacek@status.im>
This commit is contained in:
parent
7c53841cd8
commit
9593ef74b8
@ -768,6 +768,7 @@ proc getBeaconHead*(
|
|||||||
let
|
let
|
||||||
finalizedExecutionBlockHash =
|
finalizedExecutionBlockHash =
|
||||||
pool.dag.loadExecutionBlockHash(pool.dag.finalizedHead.blck)
|
pool.dag.loadExecutionBlockHash(pool.dag.finalizedHead.blck)
|
||||||
|
.get(ZERO_HASH)
|
||||||
|
|
||||||
# https://github.com/ethereum/consensus-specs/blob/v1.4.0-beta.6/fork_choice/safe-block.md#get_safe_execution_payload_hash
|
# https://github.com/ethereum/consensus-specs/blob/v1.4.0-beta.6/fork_choice/safe-block.md#get_safe_execution_payload_hash
|
||||||
safeBlockRoot = pool.forkChoice.get_safe_beacon_block_root()
|
safeBlockRoot = pool.forkChoice.get_safe_beacon_block_root()
|
||||||
@ -783,6 +784,7 @@ proc getBeaconHead*(
|
|||||||
finalizedExecutionBlockHash
|
finalizedExecutionBlockHash
|
||||||
else:
|
else:
|
||||||
pool.dag.loadExecutionBlockHash(safeBlock.get)
|
pool.dag.loadExecutionBlockHash(safeBlock.get)
|
||||||
|
.get(finalizedExecutionBlockHash)
|
||||||
|
|
||||||
BeaconHead(
|
BeaconHead(
|
||||||
blck: headBlock,
|
blck: headBlock,
|
||||||
|
@ -2278,20 +2278,25 @@ proc pruneHistory*(dag: ChainDAGRef, startup = false) =
|
|||||||
if dag.db.clearBlocks(fork):
|
if dag.db.clearBlocks(fork):
|
||||||
break
|
break
|
||||||
|
|
||||||
proc loadExecutionBlockHash*(dag: ChainDAGRef, bid: BlockId): Eth2Digest =
|
proc loadExecutionBlockHash*(
|
||||||
|
dag: ChainDAGRef, bid: BlockId): Opt[Eth2Digest] =
|
||||||
let blockData = dag.getForkedBlock(bid).valueOr:
|
let blockData = dag.getForkedBlock(bid).valueOr:
|
||||||
return ZERO_HASH
|
# Besides database inconsistency issues, this is hit with checkpoint sync.
|
||||||
|
# The initial `BlockRef` is creted before the checkpoint block is loaded.
|
||||||
|
# It is backfilled later, so return `none` and keep retrying.
|
||||||
|
return Opt.none(Eth2Digest)
|
||||||
|
|
||||||
withBlck(blockData):
|
withBlck(blockData):
|
||||||
when consensusFork >= ConsensusFork.Bellatrix:
|
when consensusFork >= ConsensusFork.Bellatrix:
|
||||||
forkyBlck.message.body.execution_payload.block_hash
|
Opt.some forkyBlck.message.body.execution_payload.block_hash
|
||||||
else:
|
else:
|
||||||
ZERO_HASH
|
Opt.some ZERO_HASH
|
||||||
|
|
||||||
proc loadExecutionBlockHash*(dag: ChainDAGRef, blck: BlockRef): Eth2Digest =
|
proc loadExecutionBlockHash*(
|
||||||
|
dag: ChainDAGRef, blck: BlockRef): Opt[Eth2Digest] =
|
||||||
if blck.executionBlockHash.isNone:
|
if blck.executionBlockHash.isNone:
|
||||||
blck.executionBlockHash = Opt.some dag.loadExecutionBlockHash(blck.bid)
|
blck.executionBlockHash = dag.loadExecutionBlockHash(blck.bid)
|
||||||
blck.executionBlockHash.unsafeGet
|
blck.executionBlockHash
|
||||||
|
|
||||||
from std/packedsets import PackedSet, incl, items
|
from std/packedsets import PackedSet, incl, items
|
||||||
|
|
||||||
@ -2509,10 +2514,12 @@ proc updateHead*(
|
|||||||
|
|
||||||
dag.db.updateFinalizedBlocks(newFinalized)
|
dag.db.updateFinalizedBlocks(newFinalized)
|
||||||
|
|
||||||
if dag.loadExecutionBlockHash(oldFinalizedHead.blck).isZero and
|
let oldBlockHash = dag.loadExecutionBlockHash(oldFinalizedHead.blck)
|
||||||
not dag.loadExecutionBlockHash(dag.finalizedHead.blck).isZero and
|
if oldBlockHash.isSome and oldBlockHash.unsafeGet.isZero:
|
||||||
dag.vanityLogs.onFinalizedMergeTransitionBlock != nil:
|
let newBlockHash = dag.loadExecutionBlockHash(dag.finalizedHead.blck)
|
||||||
dag.vanityLogs.onFinalizedMergeTransitionBlock()
|
if newBlockHash.isSome and not newBlockHash.unsafeGet.isZero:
|
||||||
|
if dag.vanityLogs.onFinalizedMergeTransitionBlock != nil:
|
||||||
|
dag.vanityLogs.onFinalizedMergeTransitionBlock()
|
||||||
|
|
||||||
# Pruning the block dag is required every time the finalized head changes
|
# Pruning the block dag is required every time the finalized head changes
|
||||||
# in order to clear out blocks that are no longer viable and should
|
# in order to clear out blocks that are no longer viable and should
|
||||||
|
@ -25,6 +25,8 @@ from ../validators/keystore_management import
|
|||||||
getSuggestedGasLimit
|
getSuggestedGasLimit
|
||||||
from ../validators/action_tracker import ActionTracker, getNextProposalSlot
|
from ../validators/action_tracker import ActionTracker, getNextProposalSlot
|
||||||
|
|
||||||
|
logScope: topics = "cman"
|
||||||
|
|
||||||
type
|
type
|
||||||
ConsensusManager* = object
|
ConsensusManager* = object
|
||||||
expectedSlot: Slot
|
expectedSlot: Slot
|
||||||
@ -155,7 +157,20 @@ func setOptimisticHead*(
|
|||||||
|
|
||||||
proc updateExecutionClientHead(self: ref ConsensusManager,
|
proc updateExecutionClientHead(self: ref ConsensusManager,
|
||||||
newHead: BeaconHead): Future[Opt[void]] {.async: (raises: [CancelledError]).} =
|
newHead: BeaconHead): Future[Opt[void]] {.async: (raises: [CancelledError]).} =
|
||||||
let headExecutionBlockHash = self.dag.loadExecutionBlockHash(newHead.blck)
|
let headExecutionBlockHash =
|
||||||
|
self.dag.loadExecutionBlockHash(newHead.blck).valueOr:
|
||||||
|
# `BlockRef` are only created for blocks that have passed
|
||||||
|
# execution block hash validation, either explicitly in
|
||||||
|
# `block_processor.storeBlock`, or implicitly, e.g., through
|
||||||
|
# checkpoint sync. With checkpoint sync, the checkpoint block
|
||||||
|
# is initially not available, so if there is a reorg to it,
|
||||||
|
# this may be triggered. Such a reorg could happen if the first
|
||||||
|
# imported chain is completely invalid (after the checkpoint block)
|
||||||
|
# and is subsequently pruned, in which case checkpoint block is head.
|
||||||
|
# Because execution block hash validation has already passed,
|
||||||
|
# we can treat this as `SYNCING`.
|
||||||
|
warn "Failed to load head execution block hash", head = newHead.blck
|
||||||
|
return Opt[void].ok()
|
||||||
|
|
||||||
if headExecutionBlockHash.isZero:
|
if headExecutionBlockHash.isZero:
|
||||||
# Blocks without execution payloads can't be optimistic.
|
# Blocks without execution payloads can't be optimistic.
|
||||||
@ -240,13 +255,15 @@ proc updateHead*(self: var ConsensusManager, wallSlot: Slot) =
|
|||||||
## `pruneFinalized` must be called for pruning.
|
## `pruneFinalized` must be called for pruning.
|
||||||
|
|
||||||
# Grab the new head according to our latest attestation data
|
# Grab the new head according to our latest attestation data
|
||||||
let newHead = self.attestationPool[].selectOptimisticHead(
|
let
|
||||||
wallSlot.start_beacon_time).valueOr:
|
newHead = self.attestationPool[].selectOptimisticHead(
|
||||||
warn "Head selection failed, using previous head",
|
wallSlot.start_beacon_time).valueOr:
|
||||||
head = shortLog(self.dag.head), wallSlot
|
warn "Head selection failed, using previous head",
|
||||||
return
|
head = shortLog(self.dag.head), wallSlot
|
||||||
|
return
|
||||||
|
executionBlockHash = self.dag.loadExecutionBlockHash(newHead.blck)
|
||||||
|
|
||||||
if self.dag.loadExecutionBlockHash(newHead.blck).isZero:
|
if executionBlockHash.isSome and executionBlockHash.unsafeGet.isZero:
|
||||||
# Blocks without execution payloads can't be optimistic.
|
# Blocks without execution payloads can't be optimistic.
|
||||||
self.dag.markBlockVerified(newHead.blck)
|
self.dag.markBlockVerified(newHead.blck)
|
||||||
|
|
||||||
@ -347,7 +364,7 @@ proc runProposalForkchoiceUpdated*(
|
|||||||
feeRecipient = self[].getFeeRecipient(
|
feeRecipient = self[].getFeeRecipient(
|
||||||
nextProposer, Opt.some(validatorIndex), nextWallSlot.epoch)
|
nextProposer, Opt.some(validatorIndex), nextWallSlot.epoch)
|
||||||
beaconHead = self.attestationPool[].getBeaconHead(self.dag.head)
|
beaconHead = self.attestationPool[].getBeaconHead(self.dag.head)
|
||||||
headBlockHash = self.dag.loadExecutionBlockHash(beaconHead.blck)
|
headBlockHash = ? self.dag.loadExecutionBlockHash(beaconHead.blck)
|
||||||
|
|
||||||
if headBlockHash.isZero:
|
if headBlockHash.isZero:
|
||||||
return err()
|
return err()
|
||||||
|
@ -652,7 +652,7 @@ proc storeBlock(
|
|||||||
else:
|
else:
|
||||||
let
|
let
|
||||||
headExecutionBlockHash =
|
headExecutionBlockHash =
|
||||||
dag.loadExecutionBlockHash(newHead.get.blck)
|
dag.loadExecutionBlockHash(newHead.get.blck).get(ZERO_HASH)
|
||||||
wallSlot = self.getBeaconTime().slotOrZero
|
wallSlot = self.getBeaconTime().slotOrZero
|
||||||
if headExecutionBlockHash.isZero or
|
if headExecutionBlockHash.isZero or
|
||||||
NewPayloadStatus.noResponse == payloadStatus:
|
NewPayloadStatus.noResponse == payloadStatus:
|
||||||
|
@ -285,8 +285,18 @@ template validateBeaconBlockBellatrix(
|
|||||||
# shows that `state.latest_execution_payload_header` being default or not is
|
# shows that `state.latest_execution_payload_header` being default or not is
|
||||||
# exactly equivalent to whether that block's execution payload is default or
|
# exactly equivalent to whether that block's execution payload is default or
|
||||||
# not, so test cached block information rather than reconstructing a state.
|
# not, so test cached block information rather than reconstructing a state.
|
||||||
if signed_beacon_block.message.is_execution_block or
|
let isExecutionEnabled =
|
||||||
not dag.loadExecutionBlockHash(parent).isZero:
|
if signed_beacon_block.message.is_execution_block:
|
||||||
|
true
|
||||||
|
else:
|
||||||
|
# If we don't know whether the parent block had execution enabled,
|
||||||
|
# assume it didn't. This way, we don't reject here if the timestamp
|
||||||
|
# is invalid, and let state transition check the timestamp.
|
||||||
|
# This is an edge case, and may be hit in a pathological scenario with
|
||||||
|
# checkpoint sync, because the checkpoint block may be unavailable
|
||||||
|
# and it could already be the parent of the new block before backfill.
|
||||||
|
not dag.loadExecutionBlockHash(parent).get(ZERO_HASH).isZero
|
||||||
|
if isExecutionEnabled:
|
||||||
# [REJECT] The block's execution payload timestamp is correct with respect
|
# [REJECT] The block's execution payload timestamp is correct with respect
|
||||||
# to the slot -- i.e. execution_payload.timestamp ==
|
# to the slot -- i.e. execution_payload.timestamp ==
|
||||||
# compute_timestamp_at_slot(state, block.slot).
|
# compute_timestamp_at_slot(state, block.slot).
|
||||||
@ -523,15 +533,34 @@ proc validateBeaconBlock*(
|
|||||||
# https://github.com/ethereum/consensus-specs/blob/v1.3.0/specs/bellatrix/p2p-interface.md#beacon_block
|
# https://github.com/ethereum/consensus-specs/blob/v1.3.0/specs/bellatrix/p2p-interface.md#beacon_block
|
||||||
# `is_execution_enabled(state, block.body)` check, but unlike in
|
# `is_execution_enabled(state, block.body)` check, but unlike in
|
||||||
# validateBeaconBlockBellatrix() don't have parent BlockRef.
|
# validateBeaconBlockBellatrix() don't have parent BlockRef.
|
||||||
if signed_beacon_block.message.is_execution_block or
|
if signed_beacon_block.message.is_execution_block:
|
||||||
not dag.loadExecutionBlockHash(dag.finalizedHead.blck).isZero:
|
|
||||||
# Blocks with execution enabled will be permitted to propagate
|
# Blocks with execution enabled will be permitted to propagate
|
||||||
# regardless of the validity of the execution payload. This prevents
|
# regardless of the validity of the execution payload. This prevents
|
||||||
# network segregation between optimistic and non-optimistic nodes.
|
# network segregation between optimistic and non-optimistic nodes.
|
||||||
#
|
#
|
||||||
# [IGNORE] The block's parent (defined by `block.parent_root`)
|
# If execution_payload verification of block's parent by an execution
|
||||||
# passes all validation (including execution node verification of the
|
# node is not complete:
|
||||||
# `block.body.execution_payload`).
|
#
|
||||||
|
# - [REJECT] The block's parent (defined by `block.parent_root`) passes
|
||||||
|
# all validation (excluding execution node verification of the
|
||||||
|
# `block.body.execution_payload`).
|
||||||
|
#
|
||||||
|
# otherwise:
|
||||||
|
#
|
||||||
|
# - [IGNORE] The block's parent (defined by `block.parent_root`) passes
|
||||||
|
# all validation (including execution node verification of the
|
||||||
|
# `block.body.execution_payload`).
|
||||||
|
|
||||||
|
# Implementation restrictions:
|
||||||
|
#
|
||||||
|
# - We don't know if the parent state had execution enabled.
|
||||||
|
# If it had, and the block doesn't have it enabled anymore,
|
||||||
|
# we end up in the pre-Merge path below (`else`) and REJECT.
|
||||||
|
# Such a block is clearly invalid, though, without asking the EL.
|
||||||
|
#
|
||||||
|
# - We know that the parent was marked unviable, but don't know
|
||||||
|
# whether it was marked unviable due to consensus (REJECT) or
|
||||||
|
# execution (IGNORE) verification failure. We err on the IGNORE side.
|
||||||
return errIgnore("BeaconBlock: ignored, parent from unviable fork")
|
return errIgnore("BeaconBlock: ignored, parent from unviable fork")
|
||||||
else:
|
else:
|
||||||
# [REJECT] The block's parent (defined by `block.parent_root`) passes
|
# [REJECT] The block's parent (defined by `block.parent_root`) passes
|
||||||
|
@ -126,7 +126,8 @@ proc installDebugApiHandlers*(router: var RestRouter, node: BeaconNode) =
|
|||||||
RestNodeValidity.optimistic
|
RestNodeValidity.optimistic
|
||||||
else:
|
else:
|
||||||
RestNodeValidity.valid,
|
RestNodeValidity.valid,
|
||||||
execution_block_hash: node.dag.loadExecutionBlockHash(item.bid),
|
execution_block_hash:
|
||||||
|
node.dag.loadExecutionBlockHash(item.bid).get(ZERO_HASH),
|
||||||
extra_data: some RestNodeExtraData(
|
extra_data: some RestNodeExtraData(
|
||||||
justified_root: item.checkpoints.justified.root,
|
justified_root: item.checkpoints.justified.root,
|
||||||
finalized_root: item.checkpoints.finalized.root,
|
finalized_root: item.checkpoints.finalized.root,
|
||||||
|
@ -767,7 +767,14 @@ proc getBlindedBlockParts[
|
|||||||
Future[Result[(EPH, UInt256, ForkedBeaconBlock), string]]
|
Future[Result[(EPH, UInt256, ForkedBeaconBlock), string]]
|
||||||
{.async: (raises: [CancelledError]).} =
|
{.async: (raises: [CancelledError]).} =
|
||||||
let
|
let
|
||||||
executionBlockHash = node.dag.loadExecutionBlockHash(head)
|
executionBlockHash = node.dag.loadExecutionBlockHash(head).valueOr:
|
||||||
|
# With checkpoint sync, the checkpoint block may be unavailable,
|
||||||
|
# and it could already be the parent of the new block before backfill.
|
||||||
|
# Fallback to EL, hopefully the block is available on the local path.
|
||||||
|
warn "Failed to load parent execution block hash, skipping block builder",
|
||||||
|
slot, validator_index, head = shortLog(head)
|
||||||
|
return err("loadExecutionBlockHash failed")
|
||||||
|
|
||||||
executionPayloadHeader =
|
executionPayloadHeader =
|
||||||
try:
|
try:
|
||||||
awaitWithTimeout(
|
awaitWithTimeout(
|
||||||
|
Loading…
x
Reference in New Issue
Block a user