# beacon_chain # Copyright (c) 2018-2022 Status Research & Development GmbH # Licensed and distributed under either of # * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT). # * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0). # at your option. This file may not be copied, modified, or distributed except according to those terms. when (NimMajor, NimMinor) < (1, 4): {.push raises: [Defect].} else: {.push raises: [].} import chronicles, chronos, ../spec/datatypes/base, ../consensus_object_pools/[blockchain_dag, block_quarantine, attestation_pool], ../eth1/eth1_monitor from ../spec/eth2_apis/dynamic_fee_recipients import DynamicFeeRecipientsStore, getDynamicFeeRecipient from ../validators/keystore_management import KeymanagerHost, getSuggestedFeeRecipient type ForkChoiceUpdatedInformation* = object payloadId*: PayloadID headBlockRoot*: Eth2Digest safeBlockRoot*: Eth2Digest finalizedBlockRoot*: Eth2Digest timestamp*: uint64 feeRecipient*: Eth1Address ConsensusManager* = object expectedSlot: Slot expectedBlockReceived: Future[bool] # Validated & Verified # ---------------------------------------------------------------- dag*: ChainDAGRef attestationPool*: ref AttestationPool # Missing info # ---------------------------------------------------------------- quarantine*: ref Quarantine # Execution layer integration # ---------------------------------------------------------------- eth1Monitor*: Eth1Monitor # Allow determination of preferred fee recipient during proposals # ---------------------------------------------------------------- dynamicFeeRecipientsStore: ref DynamicFeeRecipientsStore keymanagerHost: ref KeymanagerHost defaultFeeRecipient: Eth1Address # Tracking last proposal forkchoiceUpdated payload information # ---------------------------------------------------------------- forkchoiceUpdatedInfo*: Opt[ForkchoiceUpdatedInformation] optimisticHead: tuple[bid: BlockId, execution_block_hash: Eth2Digest] # Initialization # ------------------------------------------------------------------------------ func new*(T: type ConsensusManager, dag: ChainDAGRef, attestationPool: ref AttestationPool, quarantine: ref Quarantine, eth1Monitor: Eth1Monitor, dynamicFeeRecipientsStore: ref DynamicFeeRecipientsStore, keymanagerHost: ref KeymanagerHost, defaultFeeRecipient: Eth1Address ): ref ConsensusManager = (ref ConsensusManager)( dag: dag, attestationPool: attestationPool, quarantine: quarantine, eth1Monitor: eth1Monitor, dynamicFeeRecipientsStore: dynamicFeeRecipientsStore, keymanagerHost: keymanagerHost, forkchoiceUpdatedInfo: Opt.none ForkchoiceUpdatedInformation, defaultFeeRecipient: defaultFeeRecipient ) # Consensus Management # ----------------------------------------------------------------------------------- proc checkExpectedBlock(self: var ConsensusManager) = if self.expectedBlockReceived == nil: return if self.dag.head.slot < self.expectedSlot: return self.expectedBlockReceived.complete(true) self.expectedBlockReceived = nil # Don't keep completed futures around! proc expectBlock*(self: var ConsensusManager, expectedSlot: Slot): Future[bool] = ## Return a future that will complete when a head is selected whose slot is ## equal or greater than the given slot, or a new expectation is created if self.expectedBlockReceived != nil: # Reset the old future to not leave it hanging.. an alternative would be to # cancel it, but it doesn't make any practical difference for now self.expectedBlockReceived.complete(false) let fut = newFuture[bool]("ConsensusManager.expectBlock") self.expectedSlot = expectedSlot self.expectedBlockReceived = fut # It might happen that by the time we're expecting a block, it might have # already been processed! self.checkExpectedBlock() return fut from eth/async_utils import awaitWithTimeout from web3/engine_api_types import ForkchoiceUpdatedResponse, PayloadExecutionStatus, PayloadStatusV1 func `$`(h: BlockHash): string = $h.asEth2Digest func shouldSyncOptimistically*( optimisticSlot, dagSlot, wallSlot: Slot): bool = ## Determine whether an optimistic execution block hash should be reported ## to the EL client instead of the current head as determined by fork choice. # Check whether optimistic head is sufficiently ahead of DAG const minProgress = 8 * SLOTS_PER_EPOCH # Set arbitrarily if optimisticSlot < dagSlot or optimisticSlot - dagSlot < minProgress: return false # Check whether optimistic head has synced sufficiently close to wall slot const maxAge = 2 * SLOTS_PER_EPOCH # Set arbitrarily if optimisticSlot < max(wallSlot, maxAge.Slot) - maxAge: return false true func shouldSyncOptimistically*(self: ConsensusManager, wallSlot: Slot): bool = if self.eth1Monitor == nil: return false if self.optimisticHead.execution_block_hash.isZero: return false shouldSyncOptimistically( optimisticSlot = self.optimisticHead.bid.slot, dagSlot = getStateField(self.dag.headState, slot), wallSlot = wallSlot) func optimisticHead*(self: ConsensusManager): BlockId = self.optimisticHead.bid func optimisticExecutionPayloadHash*(self: ConsensusManager): Eth2Digest = self.optimisticHead.execution_block_hash func setOptimisticHead*( self: var ConsensusManager, bid: BlockId, execution_block_hash: Eth2Digest) = self.optimisticHead = (bid: bid, execution_block_hash: execution_block_hash) proc runForkchoiceUpdated*( eth1Monitor: Eth1Monitor, headBlockRoot, safeBlockRoot, finalizedBlockRoot: Eth2Digest): Future[PayloadExecutionStatus] {.async.} = # Allow finalizedBlockRoot to be 0 to avoid sync deadlocks. # # https://github.com/ethereum/EIPs/blob/master/EIPS/eip-3675.md#pos-events # has "Before the first finalized block occurs in the system the finalized # block hash provided by this event is stubbed with # `0x0000000000000000000000000000000000000000000000000000000000000000`." # and # https://github.com/ethereum/consensus-specs/blob/v1.2.0-rc.3/specs/bellatrix/validator.md#executionpayload # notes "`finalized_block_hash` is the hash of the latest finalized execution # payload (`Hash32()` if none yet finalized)" doAssert not headBlockRoot.isZero try: # Minimize window for Eth1 monitor to shut down connection await eth1Monitor.ensureDataProvider() let fcuR = awaitWithTimeout( forkchoiceUpdated( eth1Monitor, headBlockRoot, safeBlockRoot, finalizedBlockRoot), FORKCHOICEUPDATED_TIMEOUT): debug "runForkchoiceUpdated: forkchoiceUpdated timed out" ForkchoiceUpdatedResponse( payloadStatus: PayloadStatusV1( status: PayloadExecutionStatus.syncing)) debug "runForkchoiceUpdated: ran forkchoiceUpdated", headBlockRoot, safeBlockRoot, finalizedBlockRoot, payloadStatus = $fcuR.payloadStatus.status, latestValidHash = $fcuR.payloadStatus.latestValidHash, validationError = $fcuR.payloadStatus.validationError return fcuR.payloadStatus.status except CatchableError as err: error "runForkchoiceUpdated: forkchoiceUpdated failed", err = err.msg return PayloadExecutionStatus.syncing proc runForkchoiceUpdatedDiscardResult*( eth1Monitor: Eth1Monitor, headBlockRoot, safeBlockRoot, finalizedBlockRoot: Eth2Digest) {.async.} = discard await eth1Monitor.runForkchoiceUpdated( headBlockRoot, safeBlockRoot, finalizedBlockRoot) proc updateExecutionClientHead(self: ref ConsensusManager, newHead: BeaconHead) {.async.} = if self.eth1Monitor.isNil: return let headExecutionPayloadHash = self.dag.loadExecutionBlockRoot(newHead.blck) if headExecutionPayloadHash.isZero: # Blocks without execution payloads can't be optimistic. self.dag.markBlockVerified(self.quarantine[], newHead.blck.root) return # Can't use dag.head here because it hasn't been updated yet let payloadExecutionStatus = await self.eth1Monitor.runForkchoiceUpdated( headExecutionPayloadHash, newHead.safeExecutionPayloadHash, newHead.finalizedExecutionPayloadHash) case payloadExecutionStatus of PayloadExecutionStatus.valid: self.dag.markBlockVerified(self.quarantine[], newHead.blck.root) of PayloadExecutionStatus.invalid, PayloadExecutionStatus.invalid_block_hash: self.dag.markBlockInvalid(newHead.blck.root) self.quarantine[].addUnviable(newHead.blck.root) of PayloadExecutionStatus.accepted, PayloadExecutionStatus.syncing: self.dag.optimisticRoots.incl newHead.blck.root proc updateHead*(self: var ConsensusManager, newHead: BlockRef) = ## Trigger fork choice and update the DAG with the new head block ## This does not automatically prune the DAG after finalization ## `pruneFinalized` must be called for pruning. # Store the new head in the chain DAG - this may cause epochs to be # justified and finalized self.dag.updateHead(newHead, self.quarantine[]) self.checkExpectedBlock() proc updateHead*(self: var ConsensusManager, wallSlot: Slot) = ## Trigger fork choice and update the DAG with the new head block ## This does not automatically prune the DAG after finalization ## `pruneFinalized` must be called for pruning. # Grab the new head according to our latest attestation data let newHead = self.attestationPool[].selectOptimisticHead( wallSlot.start_beacon_time).valueOr: warn "Head selection failed, using previous head", head = shortLog(self.dag.head), wallSlot return if self.dag.loadExecutionBlockRoot(newHead.blck).isZero: # Blocks without execution payloads can't be optimistic. self.dag.markBlockVerified(self.quarantine[], newHead.blck.root) self.updateHead(newHead.blck) proc checkNextProposer(dag: ChainDAGRef, slot: Slot): Opt[(ValidatorIndex, ValidatorPubKey)] = let proposer = dag.getProposer(dag.head, slot + 1) if proposer.isNone(): return Opt.none((ValidatorIndex, ValidatorPubKey)) Opt.some((proposer.get, dag.validatorKey(proposer.get).get().toPubKey)) proc getFeeRecipient*( self: ref ConsensusManager, pubkey: ValidatorPubKey, validatorIdx: ValidatorIndex, epoch: Epoch): Eth1Address = self.dynamicFeeRecipientsStore[].getDynamicFeeRecipient(validatorIdx, epoch).valueOr: if self.keymanagerHost != nil: self.keymanagerHost[].getSuggestedFeeRecipient(pubkey).valueOr: self.defaultFeeRecipient else: self.defaultFeeRecipient from ../spec/datatypes/bellatrix import PayloadID proc runProposalForkchoiceUpdated*(self: ref ConsensusManager) {.async.} = withState(self.dag.headState): let nextSlot = state.data.slot + 1 (validatorIndex, nextProposer) = self.dag.checkNextProposer(nextSlot).valueOr: return # Approximately lines up with validator_duties version. Used optimistcally/ # opportunistically, so mismatches are fine if not too frequent. let timestamp = compute_timestamp_at_slot(state.data, nextSlot) randomData = get_randao_mix(state.data, get_current_epoch(state.data)).data feeRecipient = self.getFeeRecipient( nextProposer, validatorIndex, nextSlot.epoch) beaconHead = self.attestationPool[].getBeaconHead(self.dag.head) headBlockRoot = self.dag.loadExecutionBlockRoot(beaconHead.blck) if headBlockRoot.isZero: return try: let fcResult = awaitWithTimeout( forkchoiceUpdated( self.eth1Monitor, headBlockRoot, beaconHead.safeExecutionPayloadHash, beaconHead.finalizedExecutionPayloadHash, timestamp, randomData, feeRecipient), FORKCHOICEUPDATED_TIMEOUT): debug "runProposalForkchoiceUpdated: forkchoiceUpdated timed out" ForkchoiceUpdatedResponse( payloadStatus: PayloadStatusV1(status: PayloadExecutionStatus.syncing)) if fcResult.payloadStatus.status != PayloadExecutionStatus.valid or fcResult.payloadId.isNone: return self.forkchoiceUpdatedInfo = Opt.some ForkchoiceUpdatedInformation( payloadId: bellatrix.PayloadID(fcResult.payloadId.get), headBlockRoot: headBlockRoot, safeBlockRoot: beaconHead.safeExecutionPayloadHash, finalizedBlockRoot: beaconHead.finalizedExecutionPayloadHash, timestamp: timestamp, feeRecipient: feeRecipient) except CatchableError as err: error "Engine API fork-choice update failed", err = err.msg proc updateHeadWithExecution*(self: ref ConsensusManager, newHead: BeaconHead) {.async.} = ## Trigger fork choice and update the DAG with the new head block ## This does not automatically prune the DAG after finalization ## `pruneFinalized` must be called for pruning. # Grab the new head according to our latest attestation data try: # Ensure dag.updateHead has most current information await self.updateExecutionClientHead(newHead) # Store the new head in the chain DAG - this may cause epochs to be # justified and finalized self.dag.updateHead(newHead.blck, self.quarantine[]) # TODO after things stabilize with this, check for upcoming proposal and # don't bother sending first fcU, but initially, keep both in place asyncSpawn self.runProposalForkchoiceUpdated() self[].checkExpectedBlock() except CatchableError as exc: debug "updateHeadWithExecution error", error = exc.msg proc pruneStateCachesAndForkChoice*(self: var ConsensusManager) = ## Prune unneeded and invalidated data after finalization ## - the DAG state checkpoints ## - the DAG EpochRef ## - the attestation pool/fork choice # Cleanup DAG & fork choice if we have a finalized head if self.dag.needStateCachesAndForkChoicePruning(): self.dag.pruneStateCachesDAG() self.attestationPool[].prune()