2022-07-29 08:36:20 +00:00
|
|
|
# beacon_chain
|
2024-01-06 06:18:28 +00:00
|
|
|
# Copyright (c) 2021-2024 Status Research & Development GmbH
|
2022-07-29 08:36:20 +00:00
|
|
|
# Licensed and distributed under either of
|
|
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
2024-02-18 01:16:49 +00:00
|
|
|
{.push raises: [].}
|
|
|
|
|
2022-08-19 10:30:07 +00:00
|
|
|
import
|
|
|
|
chronicles,
|
|
|
|
".."/validators/activity_metrics,
|
|
|
|
".."/spec/forks,
|
2024-01-06 06:18:28 +00:00
|
|
|
"."/[common, api, fallback_service]
|
2021-07-13 11:15:07 +00:00
|
|
|
|
2023-04-17 21:31:54 +00:00
|
|
|
const
|
|
|
|
ServiceName = "block_service"
|
2023-06-08 08:44:32 +00:00
|
|
|
BlockPollInterval = attestationSlotOffset.nanoseconds div 4
|
|
|
|
BlockPollOffset1 = TimeDiff(nanoseconds: BlockPollInterval)
|
|
|
|
BlockPollOffset2 = TimeDiff(nanoseconds: BlockPollInterval * 2)
|
|
|
|
BlockPollOffset3 = TimeDiff(nanoseconds: BlockPollInterval * 3)
|
2023-04-17 21:31:54 +00:00
|
|
|
|
|
|
|
logScope: service = ServiceName
|
2021-07-13 11:15:07 +00:00
|
|
|
|
2024-03-11 14:18:50 +00:00
|
|
|
func shortLog(v: Opt[UInt256]): auto =
|
|
|
|
if v.isNone(): "<not available>" else: toString(v.get, 10)
|
|
|
|
|
|
|
|
func shortLog(v: ForkedMaybeBlindedBeaconBlock): auto =
|
|
|
|
withForkyMaybeBlindedBlck(v):
|
2024-04-10 08:54:00 +00:00
|
|
|
when consensusFork < ConsensusFork.Deneb:
|
2024-03-11 14:18:50 +00:00
|
|
|
shortLog(forkyMaybeBlindedBlck)
|
|
|
|
else:
|
|
|
|
when isBlinded:
|
|
|
|
shortLog(forkyMaybeBlindedBlck)
|
|
|
|
else:
|
|
|
|
shortLog(forkyMaybeBlindedBlck.`block`)
|
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc proposeBlock(
|
|
|
|
vc: ValidatorClientRef,
|
|
|
|
slot: Slot,
|
|
|
|
proposerKey: ValidatorPubKey
|
|
|
|
) {.async: (raises: [CancelledError]).}
|
|
|
|
|
|
|
|
proc prepareRandao(
|
|
|
|
vc: ValidatorClientRef,
|
|
|
|
slot: Slot,
|
|
|
|
proposerKey: ValidatorPubKey
|
|
|
|
) {.async: (raises: [CancelledError]).} =
|
|
|
|
if slot == vc.beaconClock.now().slotOrZero():
|
|
|
|
# Its impossible to prepare RANDAO in the beginning of the epoch. Epoch
|
|
|
|
# signature will be requested by block proposer.
|
2023-11-04 07:14:14 +00:00
|
|
|
return
|
|
|
|
|
|
|
|
let
|
|
|
|
destSlot = slot - 1'u64
|
|
|
|
destOffset = TimeDiff(nanoseconds: NANOSECONDS_PER_SLOT.int64 div 2)
|
|
|
|
deadline = destSlot.start_beacon_time() + destOffset
|
|
|
|
epoch = slot.epoch()
|
|
|
|
# We going to wait to T - (T / 4 * 2), where T is proposer's
|
|
|
|
# duty slot.
|
|
|
|
currentSlot = (await vc.checkedWaitForSlot(destSlot, destOffset,
|
|
|
|
false)).valueOr:
|
|
|
|
debug "Unable to perform RANDAO signature preparation because of " &
|
|
|
|
"system time failure"
|
|
|
|
return
|
|
|
|
validator =
|
|
|
|
vc.getValidatorForDuties(proposerKey, slot, true).valueOr: return
|
|
|
|
|
|
|
|
if currentSlot <= destSlot:
|
|
|
|
# We do not need result, because we want it to be cached.
|
|
|
|
let
|
|
|
|
start = Moment.now()
|
|
|
|
genesisRoot = vc.beaconGenesis.genesis_validators_root
|
|
|
|
fork = vc.forkAtEpoch(epoch)
|
|
|
|
rsig = await validator.getEpochSignature(fork, genesisRoot, epoch)
|
|
|
|
timeElapsed = Moment.now() - start
|
|
|
|
if rsig.isErr():
|
|
|
|
debug "Unable to prepare RANDAO signature", epoch = epoch,
|
2024-03-22 02:37:44 +00:00
|
|
|
validator = validatorLog(validator), elapsed_time = timeElapsed,
|
2023-11-04 07:14:14 +00:00
|
|
|
current_slot = currentSlot, destination_slot = destSlot,
|
|
|
|
delay = vc.getDelay(deadline)
|
|
|
|
else:
|
|
|
|
debug "RANDAO signature has been prepared", epoch = epoch,
|
2024-03-22 02:37:44 +00:00
|
|
|
validator = validatorLog(validator), elapsed_time = timeElapsed,
|
2023-11-04 07:14:14 +00:00
|
|
|
current_slot = currentSlot, destination_slot = destSlot,
|
|
|
|
delay = vc.getDelay(deadline)
|
|
|
|
else:
|
|
|
|
debug "RANDAO signature preparation timed out", epoch = epoch,
|
2024-03-22 02:37:44 +00:00
|
|
|
validator = validatorLog(validator),
|
2023-11-04 07:14:14 +00:00
|
|
|
current_slot = currentSlot, destination_slot = destSlot,
|
|
|
|
delay = vc.getDelay(deadline)
|
|
|
|
|
|
|
|
proc spawnProposalTask(vc: ValidatorClientRef,
|
|
|
|
duty: RestProposerDuty): ProposerTask =
|
|
|
|
ProposerTask(
|
|
|
|
randaoFut: prepareRandao(vc, duty.slot, duty.pubkey),
|
|
|
|
proposeFut: proposeBlock(vc, duty.slot, duty.pubkey),
|
|
|
|
duty: duty
|
|
|
|
)
|
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc publishBlockV3(
|
|
|
|
vc: ValidatorClientRef,
|
|
|
|
currentSlot, slot: Slot,
|
|
|
|
fork: Fork,
|
|
|
|
randaoReveal: ValidatorSig,
|
|
|
|
validator: AttachedValidator
|
|
|
|
) {.async: (raises: [CancelledError]).} =
|
2021-07-13 11:15:07 +00:00
|
|
|
let
|
|
|
|
genesisRoot = vc.beaconGenesis.genesis_validators_root
|
|
|
|
graffiti =
|
|
|
|
if vc.config.graffiti.isSome():
|
|
|
|
vc.config.graffiti.get()
|
|
|
|
else:
|
|
|
|
defaultGraffitiBytes()
|
2022-07-21 16:54:07 +00:00
|
|
|
vindex = validator.index.get()
|
|
|
|
|
2022-11-24 09:14:05 +00:00
|
|
|
logScope:
|
2024-03-22 02:37:44 +00:00
|
|
|
validator = validatorLog(validator)
|
2022-11-24 09:14:05 +00:00
|
|
|
validator_index = vindex
|
|
|
|
slot = slot
|
|
|
|
wall_slot = currentSlot
|
|
|
|
|
2023-11-04 07:14:14 +00:00
|
|
|
let
|
2024-02-02 15:24:40 +00:00
|
|
|
maybeBlock =
|
2023-11-04 07:14:14 +00:00
|
|
|
try:
|
2024-09-12 09:11:08 +00:00
|
|
|
await vc.produceBlockV3(slot, randaoReveal, graffiti,
|
2024-05-19 01:49:43 +00:00
|
|
|
vc.config.builderBoostFactor,
|
2024-02-02 15:24:40 +00:00
|
|
|
ApiStrategyKind.Best)
|
|
|
|
except ValidatorApiError as exc:
|
|
|
|
warn "Unable to retrieve block data", reason = exc.getFailureReason()
|
2024-05-21 22:03:38 +00:00
|
|
|
return
|
2023-11-04 07:14:14 +00:00
|
|
|
except CancelledError as exc:
|
2024-02-02 15:24:40 +00:00
|
|
|
debug "Block data production has been interrupted"
|
2023-11-04 07:14:14 +00:00
|
|
|
raise exc
|
2024-02-02 15:24:40 +00:00
|
|
|
|
|
|
|
withForkyMaybeBlindedBlck(maybeBlock):
|
|
|
|
when isBlinded:
|
|
|
|
let
|
|
|
|
blockRoot = hash_tree_root(forkyMaybeBlindedBlck)
|
2024-03-11 14:18:50 +00:00
|
|
|
|
|
|
|
debug "Block produced",
|
|
|
|
block_type = "blinded",
|
|
|
|
block_root = shortLog(blockRoot),
|
|
|
|
blck = shortLog(maybeBlock),
|
|
|
|
execution_value = shortLog(maybeBlock.executionValue),
|
|
|
|
consensus_value = shortLog(maybeBlock.consensusValue)
|
|
|
|
|
|
|
|
let
|
2024-02-02 15:24:40 +00:00
|
|
|
signingRoot =
|
|
|
|
compute_block_signing_root(fork, genesisRoot, slot, blockRoot)
|
|
|
|
notSlashable = vc.attachedValidators[]
|
|
|
|
.slashingProtection
|
|
|
|
.registerBlock(vindex, validator.pubkey, slot, signingRoot)
|
|
|
|
|
|
|
|
logScope:
|
|
|
|
blck = shortLog(forkyMaybeBlindedBlck)
|
|
|
|
block_root = shortLog(blockRoot)
|
|
|
|
signing_root = shortLog(signingRoot)
|
|
|
|
|
|
|
|
if notSlashable.isErr():
|
|
|
|
warn "Slashing protection activated for blinded block proposal"
|
2024-05-21 22:03:38 +00:00
|
|
|
return
|
2024-02-02 15:24:40 +00:00
|
|
|
|
|
|
|
let signature =
|
|
|
|
try:
|
|
|
|
let res = await validator.getBlockSignature(fork, genesisRoot,
|
|
|
|
slot, blockRoot,
|
|
|
|
maybeBlock)
|
|
|
|
if res.isErr():
|
|
|
|
warn "Unable to sign blinded block proposal using remote signer",
|
|
|
|
reason = res.error()
|
2024-05-21 22:03:38 +00:00
|
|
|
return
|
2024-02-02 15:24:40 +00:00
|
|
|
res.get()
|
|
|
|
except CancelledError as exc:
|
|
|
|
debug "Blinded block signature process has been interrupted"
|
|
|
|
raise exc
|
|
|
|
|
|
|
|
let
|
|
|
|
signedBlock =
|
|
|
|
ForkedSignedBlindedBeaconBlock.init(forkyMaybeBlindedBlck,
|
|
|
|
blockRoot, signature)
|
|
|
|
res =
|
|
|
|
try:
|
|
|
|
debug "Sending blinded block"
|
2024-10-16 17:20:39 +00:00
|
|
|
if vc.isPastElectraFork(slot.epoch()):
|
|
|
|
await vc.publishBlindedBlockV2(
|
|
|
|
signedBlock, BroadcastValidationType.Gossip,
|
|
|
|
ApiStrategyKind.First)
|
|
|
|
else:
|
|
|
|
await vc.publishBlindedBlock(
|
|
|
|
signedBlock, ApiStrategyKind.First)
|
2024-02-02 15:24:40 +00:00
|
|
|
except ValidatorApiError as exc:
|
|
|
|
warn "Unable to publish blinded block",
|
|
|
|
reason = exc.getFailureReason()
|
2024-05-21 22:03:38 +00:00
|
|
|
return
|
2024-02-02 15:24:40 +00:00
|
|
|
except CancelledError as exc:
|
|
|
|
debug "Blinded block publication has been interrupted"
|
|
|
|
raise exc
|
|
|
|
|
|
|
|
if res:
|
|
|
|
let delay = vc.getDelay(slot.block_deadline())
|
|
|
|
beacon_blocks_sent.inc()
|
|
|
|
beacon_blocks_sent_delay.observe(delay.toFloatSeconds())
|
|
|
|
notice "Blinded block published", delay = delay
|
|
|
|
else:
|
|
|
|
warn "Blinded block was not accepted by beacon node"
|
|
|
|
else:
|
|
|
|
let
|
|
|
|
blockRoot = hash_tree_root(
|
2024-04-10 08:54:00 +00:00
|
|
|
when consensusFork < ConsensusFork.Deneb:
|
2024-02-02 15:24:40 +00:00
|
|
|
forkyMaybeBlindedBlck
|
2024-02-24 13:44:15 +00:00
|
|
|
else:
|
2024-02-25 19:42:44 +00:00
|
|
|
forkyMaybeBlindedBlck.`block`
|
2024-02-02 15:24:40 +00:00
|
|
|
)
|
2024-03-11 14:18:50 +00:00
|
|
|
|
|
|
|
debug "Block produced",
|
|
|
|
block_type = "non-blinded",
|
|
|
|
block_root = shortLog(blockRoot),
|
|
|
|
blck = shortLog(maybeBlock),
|
|
|
|
execution_value = shortLog(maybeBlock.executionValue),
|
|
|
|
consensus_value = shortLog(maybeBlock.consensusValue)
|
|
|
|
|
|
|
|
let
|
2024-02-02 15:24:40 +00:00
|
|
|
signingRoot =
|
|
|
|
compute_block_signing_root(fork, genesisRoot, slot, blockRoot)
|
|
|
|
notSlashable = vc.attachedValidators[]
|
|
|
|
.slashingProtection
|
|
|
|
.registerBlock(vindex, validator.pubkey, slot, signingRoot)
|
|
|
|
|
|
|
|
logScope:
|
|
|
|
blck = shortLog(
|
2024-04-10 08:54:00 +00:00
|
|
|
when consensusFork < ConsensusFork.Deneb:
|
2024-02-02 15:24:40 +00:00
|
|
|
forkyMaybeBlindedBlck
|
2024-02-24 13:44:15 +00:00
|
|
|
else:
|
2024-02-25 19:42:44 +00:00
|
|
|
forkyMaybeBlindedBlck.`block`
|
2024-02-02 15:24:40 +00:00
|
|
|
)
|
|
|
|
block_root = shortLog(blockRoot)
|
|
|
|
signing_root = shortLog(signingRoot)
|
|
|
|
|
|
|
|
if notSlashable.isErr():
|
|
|
|
warn "Slashing protection activated for block proposal"
|
2024-05-21 22:03:38 +00:00
|
|
|
return
|
2024-02-02 15:24:40 +00:00
|
|
|
|
|
|
|
let
|
|
|
|
signature =
|
|
|
|
try:
|
|
|
|
let res = await validator.getBlockSignature(
|
|
|
|
fork, genesisRoot, slot, blockRoot, maybeBlock)
|
|
|
|
if res.isErr():
|
|
|
|
warn "Unable to sign block proposal using remote signer",
|
|
|
|
reason = res.error()
|
2024-05-21 22:03:38 +00:00
|
|
|
return
|
2024-02-02 15:24:40 +00:00
|
|
|
res.get()
|
|
|
|
except CancelledError as exc:
|
|
|
|
debug "Block signature process has been interrupted"
|
|
|
|
raise exc
|
|
|
|
|
|
|
|
signedBlockContents =
|
|
|
|
RestPublishedSignedBlockContents.init(
|
|
|
|
forkyMaybeBlindedBlck, blockRoot, signature)
|
|
|
|
|
2022-11-24 09:14:05 +00:00
|
|
|
res =
|
|
|
|
try:
|
|
|
|
debug "Sending block"
|
2024-10-16 17:20:39 +00:00
|
|
|
if vc.isPastElectraFork(slot.epoch()):
|
|
|
|
await vc.publishBlockV2(
|
|
|
|
signedBlockContents, BroadcastValidationType.Gossip,
|
|
|
|
ApiStrategyKind.First)
|
|
|
|
else:
|
|
|
|
await vc.publishBlock(
|
|
|
|
signedBlockContents, ApiStrategyKind.First)
|
2023-02-23 00:11:00 +00:00
|
|
|
except ValidatorApiError as exc:
|
2023-03-03 20:20:01 +00:00
|
|
|
warn "Unable to publish block", reason = exc.getFailureReason()
|
2022-11-24 09:14:05 +00:00
|
|
|
return
|
|
|
|
except CancelledError as exc:
|
|
|
|
debug "Block publication has been interrupted"
|
|
|
|
raise exc
|
|
|
|
|
|
|
|
if res:
|
|
|
|
let delay = vc.getDelay(slot.block_deadline())
|
|
|
|
beacon_blocks_sent.inc()
|
|
|
|
beacon_blocks_sent_delay.observe(delay.toFloatSeconds())
|
|
|
|
notice "Block published", delay = delay
|
|
|
|
else:
|
|
|
|
warn "Block was not accepted by beacon node"
|
2021-07-13 11:15:07 +00:00
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc publishBlock(
|
|
|
|
vc: ValidatorClientRef,
|
|
|
|
currentSlot, slot: Slot,
|
|
|
|
validator: AttachedValidator
|
|
|
|
) {.async: (raises: [CancelledError]).} =
|
2024-02-02 15:24:40 +00:00
|
|
|
let
|
|
|
|
genesisRoot = vc.beaconGenesis.genesis_validators_root
|
2024-03-14 03:44:00 +00:00
|
|
|
graffiti = vc.getGraffitiBytes(validator)
|
2024-02-02 15:24:40 +00:00
|
|
|
fork = vc.forkAtEpoch(slot.epoch)
|
|
|
|
vindex = validator.index.get()
|
|
|
|
|
|
|
|
logScope:
|
2024-03-22 02:37:44 +00:00
|
|
|
validator = validatorLog(validator)
|
2024-02-02 15:24:40 +00:00
|
|
|
validator_index = vindex
|
|
|
|
slot = slot
|
|
|
|
wall_slot = currentSlot
|
|
|
|
|
|
|
|
debug "Publishing block", delay = vc.getDelay(slot.block_deadline()),
|
|
|
|
genesis_root = genesisRoot,
|
|
|
|
graffiti = graffiti, fork = fork
|
|
|
|
let
|
|
|
|
randaoReveal =
|
|
|
|
try:
|
|
|
|
(await validator.getEpochSignature(fork, genesisRoot,
|
|
|
|
slot.epoch())).valueOr:
|
|
|
|
warn "Unable to generate RANDAO reveal using remote signer",
|
|
|
|
reason = error
|
|
|
|
return
|
|
|
|
except CancelledError as exc:
|
|
|
|
debug "RANDAO reveal production has been interrupted"
|
|
|
|
raise exc
|
|
|
|
|
2024-05-21 22:03:38 +00:00
|
|
|
await vc.publishBlockV3(currentSlot, slot, fork, randaoReveal, validator)
|
2024-02-02 15:24:40 +00:00
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc proposeBlock(
|
|
|
|
vc: ValidatorClientRef,
|
|
|
|
slot: Slot,
|
|
|
|
proposerKey: ValidatorPubKey
|
|
|
|
) {.async: (raises: [CancelledError]).} =
|
2023-04-17 21:31:54 +00:00
|
|
|
let
|
|
|
|
currentSlot = (await vc.checkedWaitForSlot(slot, ZeroTimeDiff,
|
|
|
|
false)).valueOr:
|
|
|
|
error "Unable to perform block production because of system time"
|
|
|
|
return
|
2021-07-13 11:15:07 +00:00
|
|
|
|
2023-04-17 21:31:54 +00:00
|
|
|
if currentSlot > slot:
|
|
|
|
warn "Skip block production for expired slot",
|
|
|
|
current_slot = currentSlot, duties_slot = slot
|
|
|
|
return
|
|
|
|
|
|
|
|
let validator = vc.getValidatorForDuties(proposerKey, slot).valueOr: return
|
|
|
|
|
|
|
|
try:
|
|
|
|
await vc.publishBlock(currentSlot, slot, validator)
|
2022-07-14 21:11:25 +00:00
|
|
|
except CancelledError as exc:
|
2023-04-17 21:31:54 +00:00
|
|
|
debug "Block proposing process was interrupted",
|
2024-03-22 02:37:44 +00:00
|
|
|
slot = slot, validator = validatorLog(validator)
|
2022-07-14 21:11:25 +00:00
|
|
|
raise exc
|
2021-07-13 11:15:07 +00:00
|
|
|
|
|
|
|
proc contains(data: openArray[RestProposerDuty], task: ProposerTask): bool =
|
|
|
|
for item in data:
|
|
|
|
if (item.pubkey == task.duty.pubkey) and (item.slot == task.duty.slot):
|
|
|
|
return true
|
|
|
|
false
|
|
|
|
|
|
|
|
proc contains(data: openArray[ProposerTask], duty: RestProposerDuty): bool =
|
|
|
|
for item in data:
|
|
|
|
if (item.duty.pubkey == duty.pubkey) and (item.duty.slot == duty.slot):
|
|
|
|
return true
|
|
|
|
false
|
|
|
|
|
2021-07-19 14:31:02 +00:00
|
|
|
proc checkDuty(duty: RestProposerDuty, epoch: Epoch, slot: Slot): bool =
|
2022-01-11 10:01:54 +00:00
|
|
|
let lastSlot = start_slot(epoch + 1'u64)
|
2021-07-19 14:31:02 +00:00
|
|
|
if duty.slot >= slot:
|
|
|
|
if duty.slot < lastSlot:
|
|
|
|
true
|
|
|
|
else:
|
|
|
|
warn "Block proposal duty is in the far future, ignoring",
|
2024-03-22 02:37:44 +00:00
|
|
|
duty_slot = duty.slot, pubkey = shortLog(duty.pubkey),
|
2021-07-19 14:31:02 +00:00
|
|
|
wall_slot = slot, last_slot_in_epoch = (lastSlot - 1'u64)
|
|
|
|
false
|
|
|
|
else:
|
|
|
|
warn "Block proposal duty is in the past, ignoring", duty_slot = duty.slot,
|
2024-03-22 02:37:44 +00:00
|
|
|
pubkey = shortLog(duty.pubkey), wall_slot = slot
|
2021-07-19 14:31:02 +00:00
|
|
|
false
|
|
|
|
|
2021-07-13 11:15:07 +00:00
|
|
|
proc addOrReplaceProposers*(vc: ValidatorClientRef, epoch: Epoch,
|
|
|
|
dependentRoot: Eth2Digest,
|
|
|
|
duties: openArray[RestProposerDuty]) =
|
2023-04-17 21:31:54 +00:00
|
|
|
let
|
|
|
|
default = ProposedData(epoch: FAR_FUTURE_EPOCH)
|
|
|
|
currentSlot = vc.getCurrentSlot().get(Slot(0))
|
|
|
|
epochDuties = vc.proposers.getOrDefault(epoch, default)
|
|
|
|
|
|
|
|
if not(epochDuties.isDefault()):
|
|
|
|
if epochDuties.dependentRoot != dependentRoot:
|
|
|
|
warn "Proposer duties re-organization", duties_count = len(duties),
|
|
|
|
wall_slot = currentSlot, epoch = epoch,
|
|
|
|
prior_dependent_root = epochDuties.dependentRoot,
|
|
|
|
dependent_root = dependentRoot
|
2021-07-13 11:15:07 +00:00
|
|
|
let tasks =
|
|
|
|
block:
|
2021-07-19 14:31:02 +00:00
|
|
|
var res: seq[ProposerTask]
|
2023-04-17 21:31:54 +00:00
|
|
|
var hashset = initHashSet[Slot]()
|
|
|
|
|
|
|
|
for task in epochDuties.duties:
|
|
|
|
if task notin duties:
|
|
|
|
# Task is no more relevant, so cancel it.
|
2023-11-04 07:14:14 +00:00
|
|
|
debug "Cancelling running proposal duty tasks",
|
2023-04-17 21:31:54 +00:00
|
|
|
slot = task.duty.slot,
|
2024-03-22 02:37:44 +00:00
|
|
|
pubkey = shortLog(task.duty.pubkey)
|
2023-11-04 07:14:14 +00:00
|
|
|
task.proposeFut.cancelSoon()
|
|
|
|
task.randaoFut.cancelSoon()
|
2023-04-17 21:31:54 +00:00
|
|
|
else:
|
|
|
|
# If task is already running for proper slot, we keep it alive.
|
2023-11-04 07:14:14 +00:00
|
|
|
debug "Keep running previous proposal duty tasks",
|
2023-04-17 21:31:54 +00:00
|
|
|
slot = task.duty.slot,
|
2024-03-22 02:37:44 +00:00
|
|
|
pubkey = shortLog(task.duty.pubkey)
|
2021-07-13 11:15:07 +00:00
|
|
|
res.add(task)
|
2023-04-17 21:31:54 +00:00
|
|
|
|
|
|
|
for duty in duties:
|
|
|
|
if duty notin res:
|
2024-03-22 02:37:44 +00:00
|
|
|
info "Received new proposer duty", slot = duty.slot,
|
|
|
|
pubkey = shortLog(duty.pubkey)
|
2023-04-17 21:31:54 +00:00
|
|
|
if checkDuty(duty, epoch, currentSlot):
|
|
|
|
let task = vc.spawnProposalTask(duty)
|
|
|
|
if duty.slot in hashset:
|
|
|
|
error "Multiple block proposers for this slot, " &
|
|
|
|
"producing blocks for all proposers", slot = duty.slot
|
|
|
|
else:
|
|
|
|
hashset.incl(duty.slot)
|
|
|
|
res.add(task)
|
2021-07-13 11:15:07 +00:00
|
|
|
res
|
|
|
|
vc.proposers[epoch] = ProposedData.init(epoch, dependentRoot, tasks)
|
2023-04-17 21:31:54 +00:00
|
|
|
else:
|
|
|
|
debug "New block proposal duties received",
|
|
|
|
dependent_root = dependentRoot, duties_count = len(duties),
|
|
|
|
wall_slot = currentSlot, epoch = epoch
|
|
|
|
# Spawn new proposer tasks and modify proposers map.
|
|
|
|
let tasks =
|
|
|
|
block:
|
|
|
|
var hashset = initHashSet[Slot]()
|
|
|
|
var res: seq[ProposerTask]
|
|
|
|
for duty in duties:
|
2024-03-22 02:37:44 +00:00
|
|
|
info "Received new proposer duty", slot = duty.slot,
|
|
|
|
pubkey = shortLog(duty.pubkey)
|
2023-04-17 21:31:54 +00:00
|
|
|
if checkDuty(duty, epoch, currentSlot):
|
|
|
|
let task = vc.spawnProposalTask(duty)
|
|
|
|
if duty.slot in hashset:
|
|
|
|
error "Multiple block proposers for this slot, " &
|
|
|
|
"producing blocks for all proposers", slot = duty.slot
|
|
|
|
else:
|
|
|
|
hashset.incl(duty.slot)
|
|
|
|
res.add(task)
|
|
|
|
res
|
|
|
|
vc.proposers[epoch] = ProposedData.init(epoch, dependentRoot, tasks)
|
2021-07-13 11:15:07 +00:00
|
|
|
|
2023-06-08 08:44:32 +00:00
|
|
|
proc pollForEvents(service: BlockServiceRef, node: BeaconNodeServerRef,
|
2024-10-16 17:20:39 +00:00
|
|
|
response: RestHttpResponseRef) {.
|
|
|
|
async: (raises: [CancelledError]).} =
|
2023-06-08 08:44:32 +00:00
|
|
|
let vc = service.client
|
2023-04-17 21:31:54 +00:00
|
|
|
|
2022-11-29 10:52:21 +00:00
|
|
|
logScope:
|
2023-06-08 08:44:32 +00:00
|
|
|
node = node
|
|
|
|
|
|
|
|
while true:
|
|
|
|
let events =
|
2022-11-29 10:52:21 +00:00
|
|
|
try:
|
2023-06-08 08:44:32 +00:00
|
|
|
await response.getServerSentEvents()
|
2024-10-16 17:20:39 +00:00
|
|
|
except HttpError as exc:
|
|
|
|
debug "Unable to receive server-sent event", reason = $exc.msg
|
|
|
|
return
|
2023-06-08 08:44:32 +00:00
|
|
|
except RestError as exc:
|
|
|
|
debug "Unable to receive server-sent event", reason = $exc.msg
|
|
|
|
return
|
2022-11-29 10:52:21 +00:00
|
|
|
except CancelledError as exc:
|
|
|
|
raise exc
|
2023-04-17 21:31:54 +00:00
|
|
|
|
2023-06-08 08:44:32 +00:00
|
|
|
for event in events:
|
|
|
|
case event.name
|
|
|
|
of "data":
|
|
|
|
let blck = EventBeaconBlockObject.decodeString(event.data).valueOr:
|
|
|
|
debug "Got invalid block event format", reason = error
|
|
|
|
return
|
2023-06-28 13:33:21 +00:00
|
|
|
vc.registerBlock(blck, node)
|
2023-06-08 08:44:32 +00:00
|
|
|
of "event":
|
|
|
|
if event.data != "block":
|
|
|
|
debug "Got unexpected event name field", event_name = event.name,
|
|
|
|
event_data = event.data
|
|
|
|
else:
|
|
|
|
debug "Got some unexpected event field", event_name = event.name
|
|
|
|
|
|
|
|
if len(events) == 0:
|
|
|
|
break
|
|
|
|
|
|
|
|
proc runBlockEventMonitor(service: BlockServiceRef,
|
2024-10-16 17:20:39 +00:00
|
|
|
node: BeaconNodeServerRef) {.
|
|
|
|
async: (raises: [CancelledError]).} =
|
2023-06-08 08:44:32 +00:00
|
|
|
let
|
|
|
|
vc = service.client
|
|
|
|
roles = {BeaconNodeRole.BlockProposalData}
|
|
|
|
statuses = {RestBeaconNodeStatus.Synced}
|
|
|
|
|
|
|
|
logScope:
|
|
|
|
node = node
|
|
|
|
|
|
|
|
while true:
|
|
|
|
while node.status notin statuses:
|
2023-06-28 13:33:21 +00:00
|
|
|
await vc.waitNodes(nil, statuses, roles, true)
|
2023-06-08 08:44:32 +00:00
|
|
|
|
|
|
|
let response =
|
|
|
|
block:
|
|
|
|
var resp: HttpClientResponseRef
|
|
|
|
try:
|
|
|
|
resp = await node.client.subscribeEventStream({EventTopic.Block})
|
|
|
|
if resp.status == 200:
|
2023-06-28 13:33:21 +00:00
|
|
|
Opt.some(resp)
|
2023-06-08 08:44:32 +00:00
|
|
|
else:
|
|
|
|
let body = await resp.getBodyBytes()
|
|
|
|
await resp.closeWait()
|
|
|
|
let
|
|
|
|
plain = RestPlainResponse(status: resp.status,
|
|
|
|
contentType: resp.contentType, data: body)
|
|
|
|
reason = plain.getErrorMessage()
|
2024-04-19 11:31:15 +00:00
|
|
|
debug "Unable to obtain events stream", code = resp.status,
|
2023-06-08 08:44:32 +00:00
|
|
|
reason = reason
|
2023-06-28 13:33:21 +00:00
|
|
|
Opt.none(HttpClientResponseRef)
|
2024-10-16 17:20:39 +00:00
|
|
|
except HttpError as exc:
|
|
|
|
debug "Unable to obtain events stream", reason = $exc.msg
|
|
|
|
Opt.none(HttpClientResponseRef)
|
2023-06-08 08:44:32 +00:00
|
|
|
except RestError as exc:
|
|
|
|
if not(isNil(resp)): await resp.closeWait()
|
|
|
|
debug "Unable to obtain events stream", reason = $exc.msg
|
2023-06-28 13:33:21 +00:00
|
|
|
Opt.none(HttpClientResponseRef)
|
2023-06-08 08:44:32 +00:00
|
|
|
except CancelledError as exc:
|
|
|
|
if not(isNil(resp)): await resp.closeWait()
|
|
|
|
debug "Block monitoring loop has been interrupted"
|
|
|
|
raise exc
|
|
|
|
|
2023-06-28 13:33:21 +00:00
|
|
|
if response.isSome():
|
|
|
|
debug "Block monitoring connection has been established"
|
|
|
|
try:
|
|
|
|
await service.pollForEvents(node, response.get())
|
|
|
|
except CancelledError as exc:
|
|
|
|
raise exc
|
|
|
|
finally:
|
|
|
|
debug "Block monitoring connection has been lost"
|
|
|
|
await response.get().closeWait()
|
2023-06-08 08:44:32 +00:00
|
|
|
|
|
|
|
proc pollForBlockHeaders(service: BlockServiceRef, node: BeaconNodeServerRef,
|
|
|
|
slot: Slot, waitTime: Duration,
|
2024-10-16 17:20:39 +00:00
|
|
|
index: int): Future[bool] {.
|
|
|
|
async: (raises: [CancelledError]).} =
|
2023-06-08 08:44:32 +00:00
|
|
|
let vc = service.client
|
|
|
|
|
|
|
|
logScope:
|
|
|
|
node = node
|
|
|
|
slot = slot
|
|
|
|
wait_time = waitTime
|
|
|
|
schedule_index = index
|
|
|
|
|
|
|
|
trace "Polling for block header"
|
|
|
|
|
|
|
|
let bres =
|
|
|
|
try:
|
|
|
|
await sleepAsync(waitTime)
|
|
|
|
await node.client.getBlockHeader(BlockIdent.init(slot))
|
|
|
|
except RestError as exc:
|
|
|
|
debug "Unable to obtain block header",
|
|
|
|
reason = $exc.msg, error = $exc.name
|
|
|
|
return false
|
|
|
|
except RestResponseError as exc:
|
|
|
|
debug "Got an error while trying to obtain block header",
|
|
|
|
reason = exc.message, status = exc.status
|
|
|
|
return false
|
|
|
|
except CancelledError as exc:
|
|
|
|
raise exc
|
|
|
|
|
|
|
|
if bres.isNone():
|
|
|
|
trace "Beacon node does not yet have block"
|
|
|
|
return false
|
|
|
|
|
|
|
|
let blockHeader = bres.get()
|
|
|
|
|
|
|
|
let eventBlock = EventBeaconBlockObject(
|
|
|
|
slot: blockHeader.data.header.message.slot,
|
|
|
|
block_root: blockHeader.data.root,
|
|
|
|
optimistic: blockHeader.execution_optimistic
|
|
|
|
)
|
2023-06-28 13:33:21 +00:00
|
|
|
vc.registerBlock(eventBlock, node)
|
2024-10-16 17:20:39 +00:00
|
|
|
true
|
2023-06-08 08:44:32 +00:00
|
|
|
|
|
|
|
proc runBlockPollMonitor(service: BlockServiceRef,
|
2024-10-16 17:20:39 +00:00
|
|
|
node: BeaconNodeServerRef) {.
|
|
|
|
async: (raises: [CancelledError]).} =
|
2023-06-08 08:44:32 +00:00
|
|
|
let
|
|
|
|
vc = service.client
|
|
|
|
roles = {BeaconNodeRole.BlockProposalData}
|
|
|
|
statuses = {RestBeaconNodeStatus.Synced}
|
|
|
|
|
|
|
|
logScope:
|
|
|
|
node = node
|
|
|
|
|
|
|
|
while true:
|
2024-10-16 17:20:39 +00:00
|
|
|
let currentSlot {.used.} =
|
|
|
|
(await vc.checkedWaitForNextSlot(ZeroTimeDiff, false)).valueOr:
|
|
|
|
continue
|
2023-06-08 08:44:32 +00:00
|
|
|
|
|
|
|
while node.status notin statuses:
|
2023-06-28 13:33:21 +00:00
|
|
|
await vc.waitNodes(nil, statuses, roles, true)
|
2023-06-08 08:44:32 +00:00
|
|
|
|
|
|
|
let
|
|
|
|
currentTime = vc.beaconClock.now()
|
|
|
|
afterSlot = currentTime.slotOrZero()
|
|
|
|
|
|
|
|
if currentTime > afterSlot.attestation_deadline():
|
|
|
|
# Attestation time already, lets wait for next slot.
|
|
|
|
continue
|
|
|
|
|
|
|
|
let
|
|
|
|
pollTime1 = afterSlot.start_beacon_time() + BlockPollOffset1
|
|
|
|
pollTime2 = afterSlot.start_beacon_time() + BlockPollOffset2
|
|
|
|
pollTime3 = afterSlot.start_beacon_time() + BlockPollOffset3
|
|
|
|
|
|
|
|
var pendingTasks =
|
|
|
|
block:
|
|
|
|
var res: seq[FutureBase]
|
|
|
|
if currentTime <= pollTime1:
|
|
|
|
let stime = nanoseconds((pollTime1 - currentTime).nanoseconds)
|
|
|
|
res.add(FutureBase(
|
|
|
|
service.pollForBlockHeaders(node, afterSlot, stime, 0)))
|
|
|
|
if currentTime <= pollTime2:
|
|
|
|
let stime = nanoseconds((pollTime2 - currentTime).nanoseconds)
|
|
|
|
res.add(FutureBase(
|
|
|
|
service.pollForBlockHeaders(node, afterSlot, stime, 1)))
|
|
|
|
if currentTime <= pollTime3:
|
|
|
|
let stime = nanoseconds((pollTime3 - currentTime).nanoseconds)
|
|
|
|
res.add(FutureBase(
|
|
|
|
service.pollForBlockHeaders(node, afterSlot, stime, 2)))
|
|
|
|
res
|
|
|
|
try:
|
|
|
|
while true:
|
2024-10-16 17:20:39 +00:00
|
|
|
let completedFuture =
|
|
|
|
try:
|
|
|
|
await race(pendingTasks)
|
|
|
|
except ValueError:
|
|
|
|
raiseAssert "Number of pending tasks should not be zero"
|
2023-06-08 08:44:32 +00:00
|
|
|
let blockReceived =
|
|
|
|
block:
|
|
|
|
var res = false
|
|
|
|
for future in pendingTasks:
|
2023-06-09 23:43:28 +00:00
|
|
|
if not(future.completed()): continue
|
2024-10-16 17:20:39 +00:00
|
|
|
if not(cast[Future[bool]](future).value): continue
|
2023-06-08 08:44:32 +00:00
|
|
|
res = true
|
|
|
|
break
|
|
|
|
res
|
|
|
|
if blockReceived:
|
2023-09-24 07:28:09 +00:00
|
|
|
let pending =
|
|
|
|
pendingTasks.filterIt(not(it.finished())).mapIt(it.cancelAndWait())
|
|
|
|
# We use `noCancel` here because its cleanup and we have `break`
|
|
|
|
# after it.
|
|
|
|
await noCancel allFutures(pending)
|
2023-06-08 08:44:32 +00:00
|
|
|
break
|
|
|
|
pendingTasks.keepItIf(it != completedFuture)
|
|
|
|
if len(pendingTasks) == 0: break
|
|
|
|
except CancelledError as exc:
|
2023-09-24 07:28:09 +00:00
|
|
|
let pending =
|
|
|
|
pendingTasks.filterIt(not(it.finished())).mapIt(it.cancelAndWait())
|
|
|
|
await noCancel allFutures(pending)
|
2023-06-08 08:44:32 +00:00
|
|
|
raise exc
|
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc runBlockMonitor(service: BlockServiceRef) {.
|
|
|
|
async: (raises: [CancelledError]).} =
|
2023-06-08 08:44:32 +00:00
|
|
|
let
|
|
|
|
vc = service.client
|
2023-09-15 05:34:32 +00:00
|
|
|
blockNodes = vc.filterNodes(ResolvedBeaconNodeStatuses,
|
2023-06-08 08:44:32 +00:00
|
|
|
{BeaconNodeRole.BlockProposalData})
|
|
|
|
let pendingTasks =
|
|
|
|
case vc.config.monitoringType
|
|
|
|
of BlockMonitoringType.Disabled:
|
|
|
|
debug "Block monitoring disabled"
|
2024-10-16 17:20:39 +00:00
|
|
|
@[Future[void].Raising([CancelledError]).init("block.monitor.disabled")]
|
2023-06-08 08:44:32 +00:00
|
|
|
of BlockMonitoringType.Poll:
|
2024-10-16 17:20:39 +00:00
|
|
|
blockNodes.mapIt(service.runBlockPollMonitor(it))
|
2023-06-08 08:44:32 +00:00
|
|
|
of BlockMonitoringType.Event:
|
2024-10-16 17:20:39 +00:00
|
|
|
blockNodes.mapIt(service.runBlockEventMonitor(it))
|
2023-06-08 08:44:32 +00:00
|
|
|
|
|
|
|
try:
|
|
|
|
await allFutures(pendingTasks)
|
|
|
|
except CancelledError as exc:
|
2023-09-24 07:28:09 +00:00
|
|
|
let pending =
|
|
|
|
pendingTasks.filterIt(not(it.finished())).mapIt(it.cancelAndWait())
|
|
|
|
await noCancel allFutures(pending)
|
2023-06-08 08:44:32 +00:00
|
|
|
raise exc
|
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc mainLoop(service: BlockServiceRef) {.async: (raises: []).} =
|
2023-04-17 21:31:54 +00:00
|
|
|
let vc = service.client
|
|
|
|
service.state = ServiceState.Running
|
|
|
|
debug "Service started"
|
2023-06-08 08:44:32 +00:00
|
|
|
let future = service.runBlockMonitor()
|
2023-04-17 21:31:54 +00:00
|
|
|
try:
|
|
|
|
# Future is not going to be completed, so the only way to exit, is to
|
|
|
|
# cancel it.
|
|
|
|
await future
|
2023-12-05 11:45:47 +00:00
|
|
|
except CancelledError:
|
2023-04-17 21:31:54 +00:00
|
|
|
debug "Service interrupted"
|
|
|
|
|
|
|
|
# We going to cleanup all the pending proposer tasks.
|
2023-09-24 07:28:09 +00:00
|
|
|
var res: seq[FutureBase]
|
2023-04-17 21:31:54 +00:00
|
|
|
for epoch, data in vc.proposers.pairs():
|
|
|
|
for duty in data.duties.items():
|
2023-11-04 07:14:14 +00:00
|
|
|
if not(duty.proposeFut.finished()):
|
|
|
|
res.add(duty.proposeFut.cancelAndWait())
|
|
|
|
if not(duty.randaoFut.finished()):
|
|
|
|
res.add(duty.randaoFut.cancelAndWait())
|
2023-09-24 07:28:09 +00:00
|
|
|
await noCancel allFutures(res)
|
2023-04-17 21:31:54 +00:00
|
|
|
|
2024-10-16 17:20:39 +00:00
|
|
|
proc init*(
|
|
|
|
t: typedesc[BlockServiceRef],
|
|
|
|
vc: ValidatorClientRef
|
|
|
|
): Future[BlockServiceRef] {.async: (raises: []).} =
|
2023-04-17 21:31:54 +00:00
|
|
|
logScope: service = ServiceName
|
2024-10-16 17:20:39 +00:00
|
|
|
let res = BlockServiceRef(name: ServiceName, client: vc,
|
2023-04-17 21:31:54 +00:00
|
|
|
state: ServiceState.Initialized)
|
|
|
|
debug "Initializing service"
|
2024-10-16 17:20:39 +00:00
|
|
|
res
|
2023-04-17 21:31:54 +00:00
|
|
|
|
|
|
|
proc start*(service: BlockServiceRef) =
|
|
|
|
service.lifeFut = mainLoop(service)
|