2020-04-24 07:16:11 +00:00
|
|
|
# beacon_chain
|
2022-01-07 17:10:40 +00:00
|
|
|
# Copyright (c) 2018-2022 Status Research & Development GmbH
|
2020-04-24 07:16:11 +00:00
|
|
|
# Licensed and distributed under either of
|
|
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
|
|
|
{.push raises: [Defect].}
|
|
|
|
|
2019-02-19 23:35:02 +00:00
|
|
|
import
|
2020-07-09 09:29:32 +00:00
|
|
|
# Standard libraries
|
2021-03-06 07:32:55 +00:00
|
|
|
std/[options, tables, sequtils],
|
2020-07-09 09:29:32 +00:00
|
|
|
# Status libraries
|
2021-04-14 14:43:29 +00:00
|
|
|
metrics,
|
2021-10-19 14:09:26 +00:00
|
|
|
chronicles, stew/byteutils,
|
2020-07-09 09:29:32 +00:00
|
|
|
# Internal
|
2021-08-18 18:57:58 +00:00
|
|
|
../spec/[beaconstate, eth2_merkleization, forks, helpers, validator],
|
2022-01-07 17:10:40 +00:00
|
|
|
../spec/datatypes/[phase0, altair, bellatrix],
|
2021-06-11 17:51:46 +00:00
|
|
|
"."/[spec_cache, blockchain_dag, block_quarantine],
|
2021-10-19 14:09:26 +00:00
|
|
|
../fork_choice/fork_choice,
|
|
|
|
../beacon_clock
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2022-01-07 17:10:40 +00:00
|
|
|
export options, tables, phase0, altair, bellatrix, blockchain_dag, fork_choice
|
2021-10-19 14:09:26 +00:00
|
|
|
|
|
|
|
const
|
|
|
|
ATTESTATION_LOOKBACK* =
|
|
|
|
min(24'u64, SLOTS_PER_EPOCH) + MIN_ATTESTATION_INCLUSION_DELAY
|
|
|
|
## The number of slots we'll keep track of in terms of "free" attestations
|
|
|
|
## that potentially could be added to a newly created block
|
|
|
|
|
|
|
|
type
|
|
|
|
OnAttestationCallback* = proc(data: Attestation) {.gcsafe, raises: [Defect].}
|
|
|
|
|
|
|
|
Validation* = object
|
|
|
|
## Validations collect a set of signatures for a distict attestation - in
|
|
|
|
## eth2, a single bit is used to keep track of which signatures have been
|
|
|
|
## added to the aggregate meaning that only non-overlapping aggregates may
|
|
|
|
## be further combined.
|
|
|
|
aggregation_bits*: CommitteeValidatorsBits
|
|
|
|
aggregate_signature*: AggregateSignature
|
|
|
|
|
|
|
|
AttestationEntry* = object
|
|
|
|
## Each entry holds the known signatures for a particular, distinct vote
|
|
|
|
data*: AttestationData
|
|
|
|
committee_len*: int
|
|
|
|
singles*: Table[int, CookedSig] ## \
|
|
|
|
## On the attestation subnets, only attestations with a single vote are
|
|
|
|
## allowed - these can be collected separately to top up aggregates with -
|
|
|
|
## here we collect them by mapping index in committee to a vote
|
|
|
|
aggregates*: seq[Validation]
|
|
|
|
|
|
|
|
AttestationTable* = Table[Eth2Digest, AttestationEntry]
|
|
|
|
## Depending on the world view of the various validators, they may have
|
|
|
|
## voted on different states - this map keeps track of each vote keyed by
|
|
|
|
## hash_tree_root(AttestationData)
|
|
|
|
|
|
|
|
AttestationPool* = object
|
|
|
|
## The attestation pool keeps track of all attestations that potentially
|
|
|
|
## could be added to a block during block production.
|
|
|
|
## These attestations also contribute to the fork choice, which combines
|
|
|
|
## "free" attestations with those found in past blocks - these votes
|
|
|
|
## are tracked separately in the fork choice.
|
|
|
|
|
|
|
|
candidates*: array[ATTESTATION_LOOKBACK, AttestationTable] ## \
|
|
|
|
## We keep one item per slot such that indexing matches slot number
|
|
|
|
## together with startingSlot
|
|
|
|
|
|
|
|
startingSlot*: Slot ## \
|
|
|
|
## Generally, we keep attestations only until a slot has been finalized -
|
|
|
|
## after that, they may no longer affect fork choice.
|
|
|
|
|
|
|
|
dag*: ChainDAGRef
|
2021-12-06 09:49:01 +00:00
|
|
|
quarantine*: ref Quarantine
|
2021-10-19 14:09:26 +00:00
|
|
|
|
|
|
|
forkChoice*: ForkChoice
|
|
|
|
|
|
|
|
nextAttestationEpoch*: seq[tuple[subnet: Epoch, aggregate: Epoch]] ## \
|
|
|
|
## sequence based on validator indices
|
|
|
|
|
|
|
|
onAttestationAdded*: OnAttestationCallback
|
2020-07-27 16:04:44 +00:00
|
|
|
|
2019-09-12 01:45:04 +00:00
|
|
|
logScope: topics = "attpool"
|
|
|
|
|
2021-04-14 14:43:29 +00:00
|
|
|
declareGauge attestation_pool_block_attestation_packing_time,
|
|
|
|
"Time it took to create list of attestations for block"
|
|
|
|
|
2021-09-22 12:17:15 +00:00
|
|
|
proc init*(T: type AttestationPool, dag: ChainDAGRef,
|
2021-12-06 09:49:01 +00:00
|
|
|
quarantine: ref Quarantine,
|
2022-02-04 11:59:40 +00:00
|
|
|
onAttestation: OnAttestationCallback = nil,
|
|
|
|
proposerBoosting: bool = false): T =
|
2021-06-01 11:13:40 +00:00
|
|
|
## Initialize an AttestationPool from the dag `headState`
|
2020-06-10 06:58:12 +00:00
|
|
|
## The `finalized_root` works around the finalized_checkpoint of the genesis block
|
|
|
|
## holding a zero_root.
|
2021-06-01 11:13:40 +00:00
|
|
|
let finalizedEpochRef = dag.getFinalizedEpochRef()
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2020-08-18 14:56:32 +00:00
|
|
|
var forkChoice = ForkChoice.init(
|
2020-11-03 01:21:07 +00:00
|
|
|
finalizedEpochRef,
|
2022-02-04 11:59:40 +00:00
|
|
|
dag.finalizedHead.blck,
|
|
|
|
proposerBoosting)
|
2020-07-25 19:41:12 +00:00
|
|
|
|
2020-07-27 16:04:44 +00:00
|
|
|
# Feed fork choice with unfinalized history - during startup, block pool only
|
|
|
|
# keeps track of a single history so we just need to follow it
|
2021-06-01 11:13:40 +00:00
|
|
|
doAssert dag.heads.len == 1, "Init only supports a single history"
|
2020-07-27 16:04:44 +00:00
|
|
|
|
2020-07-25 19:41:12 +00:00
|
|
|
var blocks: seq[BlockRef]
|
2021-06-01 11:13:40 +00:00
|
|
|
var cur = dag.head
|
2020-10-29 11:09:03 +00:00
|
|
|
|
|
|
|
# When the chain is finalizing, the votes between the head block and the
|
|
|
|
# finalized checkpoint should be enough for a stable fork choice - when the
|
|
|
|
# chain is not finalizing, we want to seed it with as many votes as possible
|
|
|
|
# since the whole history of each branch might be significant. It is however
|
|
|
|
# a game of diminishing returns, and we have to weigh it against the time
|
|
|
|
# it takes to replay that many blocks during startup and thus miss _new_
|
|
|
|
# votes.
|
|
|
|
const ForkChoiceHorizon = 256
|
2021-06-01 11:13:40 +00:00
|
|
|
while cur != dag.finalizedHead.blck:
|
2020-07-25 19:41:12 +00:00
|
|
|
blocks.add cur
|
|
|
|
cur = cur.parent
|
|
|
|
|
2020-11-16 19:15:43 +00:00
|
|
|
info "Initializing fork choice", unfinalized_blocks = blocks.len
|
2020-08-03 18:39:43 +00:00
|
|
|
|
2020-10-29 11:09:03 +00:00
|
|
|
var epochRef = finalizedEpochRef
|
|
|
|
for i in 0..<blocks.len:
|
2020-08-03 18:39:43 +00:00
|
|
|
let
|
2021-07-14 12:18:52 +00:00
|
|
|
blckRef = blocks[blocks.len - i - 1]
|
2020-08-03 18:39:43 +00:00
|
|
|
status =
|
2020-11-02 17:51:08 +00:00
|
|
|
if i < (blocks.len - ForkChoiceHorizon) and (i mod 1024 != 0):
|
2020-10-29 11:09:03 +00:00
|
|
|
# Fork choice needs to know about the full block tree up to the
|
|
|
|
# finalization point, but doesn't really need to have overly accurate
|
|
|
|
# justification and finalization points until we get close to head -
|
|
|
|
# nonetheless, we'll make sure to pass a fresh finalization point now
|
|
|
|
# and then to make sure the fork choice data structure doesn't grow
|
|
|
|
# too big - getting an EpochRef can be expensive.
|
|
|
|
forkChoice.backend.process_block(
|
2021-07-14 12:18:52 +00:00
|
|
|
blckRef.root, blckRef.parent.root,
|
2021-11-25 18:41:39 +00:00
|
|
|
epochRef.current_justified_checkpoint,
|
|
|
|
epochRef.finalized_checkpoint)
|
2020-10-29 11:09:03 +00:00
|
|
|
else:
|
2022-01-05 18:38:04 +00:00
|
|
|
epochRef = dag.getEpochRef(blckRef, blckRef.slot.epoch, false).expect(
|
|
|
|
"Getting an EpochRef should always work for non-finalized blocks")
|
2022-03-11 12:08:17 +00:00
|
|
|
let blck = dag.getForkedBlock(blckRef.bid).expect(
|
|
|
|
"Should be able to load initial fork choice blocks")
|
|
|
|
withBlck(blck):
|
2021-07-14 12:18:52 +00:00
|
|
|
forkChoice.process_block(
|
limit by-root requests to non-finalized blocks (#3293)
* limit by-root requests to non-finalized blocks
Presently, we keep a mapping from block root to `BlockRef` in memory -
this has simplified reasoning about the dag, but is not sustainable with
the chain growing.
We can distinguish between two cases where by-root access is useful:
* unfinalized blocks - this is where the beacon chain is operating
generally, by validating incoming data as interesting for future fork
choice decisions - bounded by the length of the unfinalized period
* finalized blocks - historical access in the REST API etc - no bounds,
really
In this PR, we limit the by-root block index to the first use case:
finalized chain data can more efficiently be addressed by slot number.
Future work includes:
* limiting the `BlockRef` horizon in general - each instance is 40
bytes+overhead which adds up - this needs further refactoring to deal
with the tail vs state problem
* persisting the finalized slot-to-hash index - this one also keeps
growing unbounded (albeit slowly)
Anyway, this PR easily shaves ~128mb of memory usage at the time of
writing.
* No longer honor `BeaconBlocksByRoot` requests outside of the
non-finalized period - previously, Nimbus would generously return any
block through this libp2p request - per the spec, finalized blocks
should be fetched via `BeaconBlocksByRange` instead.
* return `Opt[BlockRef]` instead of `nil` when blocks can't be found -
this becomes a lot more common now and thus deserves more attention
* `dag.blocks` -> `dag.forkBlocks` - this index only carries unfinalized
blocks from now - `finalizedBlocks` covers the other `BlockRef`
instances
* in backfill, verify that the last backfilled block leads back to
genesis, or panic
* add backfill timings to log
* fix missing check that `BlockRef` block can be fetched with
`getForkedBlock` reliably
* shortcut doppelganger check when feature is not enabled
* in REST/JSON-RPC, fetch blocks without involving `BlockRef`
* fix dag.blocks ref
2022-01-21 11:33:16 +00:00
|
|
|
dag, epochRef, blckRef, blck.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-25 19:41:12 +00:00
|
|
|
|
2020-08-03 18:39:43 +00:00
|
|
|
doAssert status.isOk(), "Error in preloading the fork choice: " & $status.error
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2020-10-29 11:09:03 +00:00
|
|
|
info "Fork choice initialized",
|
2021-04-08 08:24:25 +00:00
|
|
|
justified_epoch = getStateField(
|
2022-03-16 07:20:40 +00:00
|
|
|
dag.headState, current_justified_checkpoint).epoch,
|
|
|
|
finalized_epoch = getStateField(dag.headState, finalized_checkpoint).epoch,
|
2022-02-16 22:24:44 +00:00
|
|
|
finalized_root = shortLog(dag.finalizedHead.blck.root)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2019-02-28 21:21:29 +00:00
|
|
|
T(
|
2021-06-01 11:13:40 +00:00
|
|
|
dag: dag,
|
2020-07-30 19:18:17 +00:00
|
|
|
quarantine: quarantine,
|
2021-09-22 12:17:15 +00:00
|
|
|
forkChoice: forkChoice,
|
|
|
|
onAttestationAdded: onAttestation
|
2019-02-28 21:21:29 +00:00
|
|
|
)
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2020-09-14 14:50:03 +00:00
|
|
|
proc addForkChoiceVotes(
|
2021-04-26 20:39:44 +00:00
|
|
|
pool: var AttestationPool, slot: Slot,
|
|
|
|
attesting_indices: openArray[ValidatorIndex], block_root: Eth2Digest,
|
2021-12-21 18:56:08 +00:00
|
|
|
wallTime: BeaconTime) =
|
2020-07-27 16:04:44 +00:00
|
|
|
# Add attestation votes to fork choice
|
2020-08-17 18:36:13 +00:00
|
|
|
if (let v = pool.forkChoice.on_attestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
pool.dag, slot, block_root, attesting_indices, wallTime);
|
2020-08-17 18:36:13 +00:00
|
|
|
v.isErr):
|
2020-09-14 14:50:03 +00:00
|
|
|
# This indicates that the fork choice and the chain dag are out of sync -
|
|
|
|
# this is most likely the result of a bug, but we'll try to keep going -
|
|
|
|
# hopefully the fork choice will heal itself over time.
|
|
|
|
error "Couldn't add attestation to fork choice, bug?", err = v.error()
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2021-04-14 14:43:29 +00:00
|
|
|
func candidateIdx(pool: AttestationPool, slot: Slot): Option[int] =
|
2020-07-28 13:54:32 +00:00
|
|
|
if slot >= pool.startingSlot and
|
|
|
|
slot < (pool.startingSlot + pool.candidates.lenu64):
|
2021-04-14 14:43:29 +00:00
|
|
|
some(int(slot mod pool.candidates.lenu64))
|
2020-07-28 13:54:32 +00:00
|
|
|
else:
|
2021-04-14 14:43:29 +00:00
|
|
|
none(int)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-07-28 13:54:32 +00:00
|
|
|
proc updateCurrent(pool: var AttestationPool, wallSlot: Slot) =
|
|
|
|
if wallSlot + 1 < pool.candidates.lenu64:
|
2021-04-12 20:25:09 +00:00
|
|
|
return # Genesis
|
|
|
|
|
|
|
|
let
|
|
|
|
newStartingSlot = wallSlot + 1 - pool.candidates.lenu64
|
2020-06-28 17:32:11 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
if newStartingSlot < pool.startingSlot:
|
2020-07-28 13:54:32 +00:00
|
|
|
error "Current slot older than attestation pool view, clock reset?",
|
2021-04-12 20:25:09 +00:00
|
|
|
startingSlot = pool.startingSlot, newStartingSlot, wallSlot
|
2020-06-28 17:32:11 +00:00
|
|
|
return
|
|
|
|
|
2020-07-28 13:54:32 +00:00
|
|
|
# As time passes we'll clear out any old attestations as they are no longer
|
|
|
|
# viable to be included in blocks
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
if newStartingSlot - pool.startingSlot >= pool.candidates.lenu64():
|
|
|
|
# In case many slots passed since the last update, avoid iterating over
|
|
|
|
# the same indices over and over
|
|
|
|
pool.candidates = default(type(pool.candidates))
|
|
|
|
else:
|
|
|
|
for i in pool.startingSlot..newStartingSlot:
|
|
|
|
pool.candidates[i.uint64 mod pool.candidates.lenu64] = AttestationTable()
|
|
|
|
|
|
|
|
pool.startingSlot = newStartingSlot
|
|
|
|
|
2022-02-17 11:53:55 +00:00
|
|
|
func oneIndex(bits: CommitteeValidatorsBits): Option[int] =
|
2021-04-12 20:25:09 +00:00
|
|
|
# Find the index of the set bit, iff one bit is set
|
|
|
|
var res = none(int)
|
|
|
|
for idx in 0..<bits.len():
|
|
|
|
if bits[idx]:
|
|
|
|
if res.isNone():
|
|
|
|
res = some(idx)
|
|
|
|
else: # More than one bit set!
|
|
|
|
return none(int)
|
|
|
|
res
|
|
|
|
|
|
|
|
func toAttestation(entry: AttestationEntry, validation: Validation): Attestation =
|
|
|
|
Attestation(
|
|
|
|
aggregation_bits: validation.aggregation_bits,
|
|
|
|
data: entry.data,
|
2021-04-26 20:39:44 +00:00
|
|
|
signature: validation.aggregate_signature.finish().toValidatorSig()
|
2021-04-12 20:25:09 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
func updateAggregates(entry: var AttestationEntry) =
|
|
|
|
# Upgrade the list of aggregates to ensure that there is at least one
|
|
|
|
# aggregate (assuming there are singles) and all aggregates have all
|
|
|
|
# singles incorporated
|
|
|
|
if entry.singles.len() == 0:
|
|
|
|
return
|
|
|
|
|
|
|
|
if entry.aggregates.len() == 0:
|
|
|
|
# If there are singles, we can create an aggregate from them that will
|
|
|
|
# represent our best knowledge about the current votes
|
|
|
|
for index_in_committee, signature in entry.singles:
|
|
|
|
if entry.aggregates.len() == 0:
|
|
|
|
# Create aggregate on first iteration..
|
|
|
|
entry.aggregates.add(
|
|
|
|
Validation(
|
|
|
|
aggregation_bits: CommitteeValidatorsBits.init(entry.committee_len),
|
|
|
|
aggregate_signature: AggregateSignature.init(signature)
|
|
|
|
))
|
|
|
|
else:
|
|
|
|
entry.aggregates[0].aggregate_signature.aggregate(signature)
|
|
|
|
|
|
|
|
entry.aggregates[0].aggregation_bits.setBit(index_in_committee)
|
|
|
|
else:
|
|
|
|
# There already exist aggregates - we'll try to top them up by adding
|
|
|
|
# singles to them - for example, it may happen that we're being asked to
|
|
|
|
# produce a block 4s after creating an aggregate and new information may
|
|
|
|
# have arrived by then.
|
|
|
|
# In theory, also aggregates could be combined but finding the best
|
|
|
|
# combination is hard, so we'll pragmatically use singles only here
|
|
|
|
var updated = false
|
|
|
|
for index_in_committee, signature in entry.singles:
|
|
|
|
for v in entry.aggregates.mitems():
|
|
|
|
if not v.aggregation_bits[index_in_committee]:
|
|
|
|
v.aggregation_bits.setBit(index_in_committee)
|
|
|
|
v.aggregate_signature.aggregate(signature)
|
|
|
|
updated = true
|
|
|
|
|
|
|
|
if updated:
|
|
|
|
# One or more aggregates were updated - time to remove the ones that are
|
|
|
|
# pure subsets of the others. This may lead to quadratic behaviour, but
|
|
|
|
# the number of aggregates for the entry is limited by the number of
|
|
|
|
# aggregators on the topic which is capped `is_aggregator` and
|
|
|
|
# TARGET_AGGREGATORS_PER_COMMITTEE
|
|
|
|
var i = 0
|
|
|
|
while i < entry.aggregates.len():
|
|
|
|
var j = 0
|
|
|
|
while j < entry.aggregates.len():
|
|
|
|
if i != j and entry.aggregates[i].aggregation_bits.isSubsetOf(
|
|
|
|
entry.aggregates[j].aggregation_bits):
|
|
|
|
entry.aggregates[i] = entry.aggregates[j]
|
|
|
|
entry.aggregates.del(j)
|
|
|
|
dec i # Rerun checks on the new `i` item
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
inc j
|
|
|
|
inc i
|
2020-09-14 11:13:30 +00:00
|
|
|
|
2022-02-25 16:15:39 +00:00
|
|
|
func covers(entry: AttestationEntry, bits: CommitteeValidatorsBits): bool =
|
|
|
|
for i in 0..<entry.aggregates.len():
|
|
|
|
if bits.isSubsetOf(entry.aggregates[i].aggregation_bits):
|
|
|
|
return true
|
|
|
|
false
|
|
|
|
|
2021-04-14 14:43:29 +00:00
|
|
|
proc addAttestation(entry: var AttestationEntry,
|
|
|
|
attestation: Attestation,
|
|
|
|
signature: CookedSig): bool =
|
|
|
|
logScope:
|
|
|
|
attestation = shortLog(attestation)
|
|
|
|
|
|
|
|
let
|
|
|
|
singleIndex = oneIndex(attestation.aggregation_bits)
|
|
|
|
|
|
|
|
if singleIndex.isSome():
|
|
|
|
if singleIndex.get() in entry.singles:
|
|
|
|
trace "Attestation already seen",
|
|
|
|
singles = entry.singles.len(),
|
|
|
|
aggregates = entry.aggregates.len()
|
|
|
|
|
|
|
|
return false
|
|
|
|
|
|
|
|
debug "Attestation resolved",
|
|
|
|
singles = entry.singles.len(),
|
|
|
|
aggregates = entry.aggregates.len()
|
|
|
|
|
|
|
|
entry.singles[singleIndex.get()] = signature
|
|
|
|
else:
|
|
|
|
# More than one vote in this attestation
|
2022-02-25 16:15:39 +00:00
|
|
|
if entry.covers(attestation.aggregation_bits):
|
|
|
|
return false
|
2021-04-14 14:43:29 +00:00
|
|
|
|
|
|
|
# Since we're adding a new aggregate, we can now remove existing
|
|
|
|
# aggregates that don't add any new votes
|
|
|
|
entry.aggregates.keepItIf(
|
|
|
|
not it.aggregation_bits.isSubsetOf(attestation.aggregation_bits))
|
|
|
|
|
|
|
|
entry.aggregates.add(Validation(
|
|
|
|
aggregation_bits: attestation.aggregation_bits,
|
|
|
|
aggregate_signature: AggregateSignature.init(signature)))
|
|
|
|
|
|
|
|
debug "Aggregate resolved",
|
|
|
|
singles = entry.singles.len(),
|
|
|
|
aggregates = entry.aggregates.len()
|
|
|
|
|
|
|
|
true
|
|
|
|
|
2020-08-27 07:34:12 +00:00
|
|
|
proc addAttestation*(pool: var AttestationPool,
|
|
|
|
attestation: Attestation,
|
2021-04-26 20:39:44 +00:00
|
|
|
attesting_indices: openArray[ValidatorIndex],
|
2021-04-09 12:59:24 +00:00
|
|
|
signature: CookedSig,
|
2021-12-21 18:56:08 +00:00
|
|
|
wallTime: BeaconTime) =
|
2020-09-14 14:50:03 +00:00
|
|
|
## Add an attestation to the pool, assuming it's been validated already.
|
|
|
|
##
|
2021-04-12 20:25:09 +00:00
|
|
|
## Assuming the votes in the attestation have not already been seen, the
|
|
|
|
## attestation will be added to the fork choice and lazily added to a list of
|
|
|
|
## attestations for future aggregation and block production.
|
2020-07-28 13:54:32 +00:00
|
|
|
logScope:
|
|
|
|
attestation = shortLog(attestation)
|
2020-07-22 07:51:45 +00:00
|
|
|
|
2021-04-26 20:39:44 +00:00
|
|
|
doAssert attestation.signature == signature.toValidatorSig(),
|
2021-04-12 20:25:09 +00:00
|
|
|
"Deserialized signature must match the one in the attestation"
|
|
|
|
|
2021-12-21 18:56:08 +00:00
|
|
|
updateCurrent(pool, wallTime.slotOrZero)
|
2020-07-28 13:54:32 +00:00
|
|
|
|
|
|
|
let candidateIdx = pool.candidateIdx(attestation.data.slot)
|
|
|
|
if candidateIdx.isNone:
|
2020-09-14 14:50:03 +00:00
|
|
|
debug "Skipping old attestation for block production",
|
2020-07-28 13:54:32 +00:00
|
|
|
startingSlot = pool.startingSlot
|
|
|
|
return
|
2019-02-28 21:21:29 +00:00
|
|
|
|
2021-04-14 14:43:29 +00:00
|
|
|
let attestation_data_root = hash_tree_root(attestation.data)
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2021-04-14 14:43:29 +00:00
|
|
|
# TODO withValue is an abomination but hard to use anything else too without
|
|
|
|
# creating an unnecessary AttestationEntry on the hot path and avoiding
|
|
|
|
# multiple lookups
|
|
|
|
pool.candidates[candidateIdx.get()].withValue(attestation_data_root, entry) do:
|
|
|
|
if not addAttestation(entry[], attestation, signature):
|
|
|
|
return
|
|
|
|
do:
|
|
|
|
if not addAttestation(
|
2021-08-10 20:46:35 +00:00
|
|
|
pool.candidates[candidateIdx.get()].mgetOrPut(
|
2021-04-14 14:43:29 +00:00
|
|
|
attestation_data_root,
|
|
|
|
AttestationEntry(
|
|
|
|
data: attestation.data,
|
|
|
|
committee_len: attestation.aggregation_bits.len())),
|
|
|
|
attestation, signature):
|
2021-04-12 20:25:09 +00:00
|
|
|
return
|
2019-08-19 16:41:13 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
pool.addForkChoiceVotes(
|
2021-04-26 20:39:44 +00:00
|
|
|
attestation.data.slot, attesting_indices,
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation.data.beacon_block_root, wallTime)
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2021-09-22 12:17:15 +00:00
|
|
|
# Send notification about new attestation via callback.
|
|
|
|
if not(isNil(pool.onAttestationAdded)):
|
|
|
|
pool.onAttestationAdded(attestation)
|
|
|
|
|
2022-02-25 16:15:39 +00:00
|
|
|
func covers*(
|
|
|
|
pool: var AttestationPool, data: Attestationdata,
|
|
|
|
bits: CommitteeValidatorsBits): bool =
|
|
|
|
## Return true iff the given attestation already is fully covered by one of
|
|
|
|
## the existing aggregates, making it redundant
|
|
|
|
## the `var` attestation pool is needed to use `withValue`, else Table becomes
|
|
|
|
## unusably inefficient
|
|
|
|
let candidateIdx = pool.candidateIdx(data.slot)
|
|
|
|
if candidateIdx.isNone:
|
|
|
|
return false
|
|
|
|
|
|
|
|
let attestation_data_root = hash_tree_root(data)
|
|
|
|
pool.candidates[candidateIdx.get()].withValue(attestation_data_root, entry):
|
|
|
|
if entry[].covers(bits):
|
|
|
|
return true
|
|
|
|
|
|
|
|
false
|
|
|
|
|
2020-07-25 19:41:12 +00:00
|
|
|
proc addForkChoice*(pool: var AttestationPool,
|
2020-08-03 18:39:43 +00:00
|
|
|
epochRef: EpochRef,
|
2020-07-25 19:41:12 +00:00
|
|
|
blckRef: BlockRef,
|
2021-11-05 07:34:34 +00:00
|
|
|
blck: ForkyTrustedBeaconBlock,
|
2021-12-21 18:56:08 +00:00
|
|
|
wallTime: BeaconTime) =
|
2020-07-09 09:29:32 +00:00
|
|
|
## Add a verified block to the fork choice context
|
2020-07-25 19:41:12 +00:00
|
|
|
let state = pool.forkChoice.process_block(
|
2021-12-21 18:56:08 +00:00
|
|
|
pool.dag, epochRef, blckRef, blck, wallTime)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2020-07-22 09:42:55 +00:00
|
|
|
if state.isErr:
|
2020-09-14 14:50:03 +00:00
|
|
|
# This indicates that the fork choice and the chain dag are out of sync -
|
|
|
|
# this is most likely the result of a bug, but we'll try to keep going -
|
|
|
|
# hopefully the fork choice will heal itself over time.
|
|
|
|
error "Couldn't add block to fork choice, bug?",
|
2020-07-25 19:41:12 +00:00
|
|
|
blck = shortLog(blck), err = state.error
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
iterator attestations*(pool: AttestationPool, slot: Option[Slot],
|
2022-01-08 23:28:49 +00:00
|
|
|
committee_index: Option[CommitteeIndex]): Attestation =
|
2021-04-14 14:43:29 +00:00
|
|
|
let candidateIndices =
|
|
|
|
if slot.isSome():
|
|
|
|
let candidateIdx = pool.candidateIdx(slot.get())
|
|
|
|
if candidateIdx.isSome():
|
|
|
|
candidateIdx.get() .. candidateIdx.get()
|
|
|
|
else:
|
|
|
|
1 .. 0
|
|
|
|
else:
|
|
|
|
0 ..< pool.candidates.len()
|
|
|
|
|
|
|
|
for candidateIndex in candidateIndices:
|
|
|
|
for _, entry in pool.candidates[candidateIndex]:
|
2022-01-08 23:28:49 +00:00
|
|
|
if committee_index.isNone() or entry.data.index == committee_index.get():
|
2021-04-12 20:25:09 +00:00
|
|
|
var singleAttestation = Attestation(
|
|
|
|
aggregation_bits: CommitteeValidatorsBits.init(entry.committee_len),
|
|
|
|
data: entry.data)
|
|
|
|
|
|
|
|
for index, signature in entry.singles:
|
|
|
|
singleAttestation.aggregation_bits.setBit(index)
|
2021-04-26 20:39:44 +00:00
|
|
|
singleAttestation.signature = signature.toValidatorSig()
|
2021-04-12 20:25:09 +00:00
|
|
|
yield singleAttestation
|
|
|
|
singleAttestation.aggregation_bits.clearBit(index)
|
|
|
|
|
|
|
|
for v in entry.aggregates:
|
|
|
|
yield entry.toAttestation(v)
|
|
|
|
|
|
|
|
type
|
|
|
|
AttestationCacheKey* = (Slot, uint64)
|
|
|
|
AttestationCache = Table[AttestationCacheKey, CommitteeValidatorsBits] ##\
|
|
|
|
## Cache for quick lookup during beacon block construction of attestations
|
|
|
|
## which have already been included, and therefore should be skipped.
|
|
|
|
|
|
|
|
func getAttestationCacheKey(ad: AttestationData): AttestationCacheKey =
|
|
|
|
# The committee is unique per slot and committee index which means we can use
|
|
|
|
# it as key for a participation cache - this is checked in `check_attestation`
|
|
|
|
(ad.slot, ad.index)
|
|
|
|
|
|
|
|
func add(
|
|
|
|
attCache: var AttestationCache, data: AttestationData,
|
|
|
|
aggregation_bits: CommitteeValidatorsBits) =
|
|
|
|
let key = data.getAttestationCacheKey()
|
|
|
|
attCache.withValue(key, v) do:
|
|
|
|
v[].incl(aggregation_bits)
|
|
|
|
do:
|
|
|
|
attCache[key] = aggregation_bits
|
|
|
|
|
2021-06-14 17:42:46 +00:00
|
|
|
func init(T: type AttestationCache, state: phase0.HashedBeaconState): T =
|
2021-04-12 20:25:09 +00:00
|
|
|
# Load attestations that are scheduled for being given rewards for
|
2021-06-11 17:51:46 +00:00
|
|
|
for i in 0..<state.data.previous_epoch_attestations.len():
|
2021-04-12 20:25:09 +00:00
|
|
|
result.add(
|
2021-06-11 17:51:46 +00:00
|
|
|
state.data.previous_epoch_attestations[i].data,
|
|
|
|
state.data.previous_epoch_attestations[i].aggregation_bits)
|
|
|
|
for i in 0..<state.data.current_epoch_attestations.len():
|
2021-04-12 20:25:09 +00:00
|
|
|
result.add(
|
2021-06-11 17:51:46 +00:00
|
|
|
state.data.current_epoch_attestations[i].data,
|
|
|
|
state.data.current_epoch_attestations[i].aggregation_bits)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
2021-06-17 17:13:14 +00:00
|
|
|
func init(
|
2021-09-27 14:22:58 +00:00
|
|
|
T: type AttestationCache,
|
2022-01-07 17:10:40 +00:00
|
|
|
state: altair.HashedBeaconState | bellatrix.HashedBeaconState,
|
2021-06-17 17:13:14 +00:00
|
|
|
cache: var StateCache): T =
|
|
|
|
# Load attestations that are scheduled for being given rewards for
|
|
|
|
let
|
|
|
|
prev_epoch = state.data.get_previous_epoch()
|
|
|
|
cur_epoch = state.data.get_current_epoch()
|
|
|
|
|
|
|
|
template update_attestation_pool_cache(
|
2022-01-08 23:28:49 +00:00
|
|
|
epoch: Epoch, participation_bitmap: untyped) =
|
2022-01-12 20:42:03 +00:00
|
|
|
let committees_per_slot = get_committee_count_per_slot(
|
|
|
|
state.data, epoch, cache)
|
|
|
|
for committee_index in get_committee_indices(committees_per_slot):
|
2022-01-11 10:01:54 +00:00
|
|
|
for slot in epoch.slots():
|
2022-01-08 23:28:49 +00:00
|
|
|
let committee = get_beacon_committee(
|
|
|
|
state.data, slot, committee_index, cache)
|
|
|
|
var
|
|
|
|
validator_bits = CommitteeValidatorsBits.init(committee.len)
|
|
|
|
for index_in_committee, validator_index in committee:
|
|
|
|
if participation_bitmap[validator_index] != 0:
|
|
|
|
# If any flag got set, there was an attestation from this validator.
|
|
|
|
validator_bits[index_in_committee] = true
|
|
|
|
result.add((slot, committee_index.uint64), validator_bits)
|
2021-06-17 17:13:14 +00:00
|
|
|
|
|
|
|
# This treats all types of rewards as equivalent, which isn't ideal
|
2022-01-08 23:28:49 +00:00
|
|
|
update_attestation_pool_cache(
|
|
|
|
prev_epoch, state.data.previous_epoch_participation)
|
|
|
|
update_attestation_pool_cache(
|
|
|
|
cur_epoch, state.data.current_epoch_participation)
|
2021-06-17 17:13:14 +00:00
|
|
|
|
2022-02-17 11:53:55 +00:00
|
|
|
func score(
|
2021-04-12 20:25:09 +00:00
|
|
|
attCache: var AttestationCache, data: AttestationData,
|
|
|
|
aggregation_bits: CommitteeValidatorsBits): int =
|
|
|
|
# The score of an attestation is loosely based on how many new votes it brings
|
|
|
|
# to the state - a more accurate score function would also look at inclusion
|
|
|
|
# distance and effective balance.
|
|
|
|
# TODO cache not var, but `withValue` requires it
|
2019-02-19 23:35:02 +00:00
|
|
|
let
|
2021-04-12 20:25:09 +00:00
|
|
|
key = data.getAttestationCacheKey()
|
|
|
|
bitsScore = aggregation_bits.countOnes()
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
attCache.withValue(key, value):
|
|
|
|
doAssert aggregation_bits.len() == value[].len(),
|
|
|
|
"check_attestation ensures committee length"
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
# How many votes were in the attestation minues the votes that are the same
|
|
|
|
return bitsScore - aggregation_bits.countOverlap(value[])
|
2020-03-31 18:39:02 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
# Not found in cache - fresh vote meaning all attestations count
|
|
|
|
bitsScore
|
2020-12-15 15:16:10 +00:00
|
|
|
|
|
|
|
proc getAttestationsForBlock*(pool: var AttestationPool,
|
2021-11-05 07:34:34 +00:00
|
|
|
state: ForkyHashedBeaconState,
|
2020-09-14 14:50:03 +00:00
|
|
|
cache: var StateCache): seq[Attestation] =
|
2020-03-31 18:39:02 +00:00
|
|
|
## Retrieve attestations that may be added to a new block at the slot of the
|
|
|
|
## given state
|
2022-03-02 10:00:21 +00:00
|
|
|
## https://github.com/ethereum/consensus-specs/blob/v1.1.10/specs/phase0/validator.md#attestations
|
2021-06-11 17:51:46 +00:00
|
|
|
let newBlockSlot = state.data.slot.uint64
|
2019-02-19 23:35:02 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
if newBlockSlot < MIN_ATTESTATION_INCLUSION_DELAY:
|
|
|
|
return # Too close to genesis
|
2020-07-27 16:04:44 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
let
|
|
|
|
# Attestations produced in a particular slot are added to the block
|
|
|
|
# at the slot where at least MIN_ATTESTATION_INCLUSION_DELAY have passed
|
|
|
|
maxAttestationSlot = newBlockSlot - MIN_ATTESTATION_INCLUSION_DELAY
|
2021-05-28 16:34:00 +00:00
|
|
|
startPackingTick = Moment.now()
|
2019-03-28 17:06:43 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
var
|
|
|
|
candidates: seq[tuple[
|
|
|
|
score: int, slot: Slot, entry: ptr AttestationEntry, validation: int]]
|
2021-06-21 08:35:24 +00:00
|
|
|
attCache =
|
|
|
|
when state is phase0.HashedBeaconState:
|
|
|
|
AttestationCache.init(state)
|
2022-01-07 17:10:40 +00:00
|
|
|
elif state is altair.HashedBeaconState or state is bellatrix.HashedBeaconState:
|
2021-06-21 08:35:24 +00:00
|
|
|
AttestationCache.init(state, cache)
|
|
|
|
else:
|
|
|
|
static: doAssert false
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
for i in 0..<ATTESTATION_LOOKBACK:
|
|
|
|
if i > maxAttestationSlot: # Around genesis..
|
|
|
|
break
|
2019-02-28 21:21:29 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
let
|
|
|
|
slot = Slot(maxAttestationSlot - i)
|
|
|
|
candidateIdx = pool.candidateIdx(slot)
|
|
|
|
|
|
|
|
if candidateIdx.isNone():
|
|
|
|
# Passed the collection horizon - shouldn't happen because it's based on
|
|
|
|
# ATTESTATION_LOOKBACK
|
|
|
|
break
|
|
|
|
|
|
|
|
for _, entry in pool.candidates[candidateIdx.get()].mpairs():
|
|
|
|
entry.updateAggregates()
|
|
|
|
|
|
|
|
for j in 0..<entry.aggregates.len():
|
|
|
|
let
|
|
|
|
attestation = entry.toAttestation(entry.aggregates[j])
|
|
|
|
|
|
|
|
# Attestations are checked based on the state that we're adding the
|
|
|
|
# attestation to - there might have been a fork between when we first
|
|
|
|
# saw the attestation and the time that we added it
|
2021-04-14 14:43:29 +00:00
|
|
|
if not check_attestation(
|
2021-06-11 17:51:46 +00:00
|
|
|
state.data, attestation, {skipBlsValidation}, cache).isOk():
|
2021-04-12 20:25:09 +00:00
|
|
|
continue
|
|
|
|
|
|
|
|
let score = attCache.score(
|
|
|
|
entry.data, entry.aggregates[j].aggregation_bits)
|
|
|
|
if score == 0:
|
|
|
|
# 0 score means the attestation would not bring any votes - discard
|
|
|
|
# it early
|
|
|
|
# Note; this must be done _after_ `check_attestation` as it relies on
|
|
|
|
# the committee to match the state that was used to build the cache
|
|
|
|
continue
|
|
|
|
|
|
|
|
# Careful, must not update the attestation table for the pointer to
|
|
|
|
# remain valid
|
|
|
|
candidates.add((score, slot, addr entry, j))
|
|
|
|
|
|
|
|
# Using a greedy algorithm, select as many attestations as possible that will
|
|
|
|
# fit in the block.
|
|
|
|
#
|
|
|
|
# For each round, we'll look for the best attestation and add it to the result
|
|
|
|
# then re-score the other candidates.
|
|
|
|
#
|
|
|
|
# A possible improvement here would be to use a maximum cover algorithm.
|
|
|
|
var
|
2021-06-11 17:51:46 +00:00
|
|
|
prevEpoch = state.data.get_previous_epoch()
|
2021-04-12 20:25:09 +00:00
|
|
|
prevEpochSpace =
|
2022-01-07 17:10:40 +00:00
|
|
|
when state is altair.HashedBeaconState or state is bellatrix.HashedBeaconState:
|
2021-06-21 08:35:24 +00:00
|
|
|
MAX_ATTESTATIONS
|
2021-09-27 14:22:58 +00:00
|
|
|
elif state is phase0.HashedBeaconState:
|
2021-06-21 08:35:24 +00:00
|
|
|
state.data.previous_epoch_attestations.maxLen -
|
|
|
|
state.data.previous_epoch_attestations.len()
|
2021-09-27 14:22:58 +00:00
|
|
|
else:
|
|
|
|
raiseAssert "invalid HashedBeaconState fork"
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
var res: seq[Attestation]
|
2021-04-14 14:43:29 +00:00
|
|
|
let totalCandidates = candidates.len()
|
2021-04-12 20:25:09 +00:00
|
|
|
while candidates.len > 0 and res.lenu64() < MAX_ATTESTATIONS:
|
|
|
|
block:
|
|
|
|
# Find the candidate with the highest score - slot is used as a
|
|
|
|
# tie-breaker so that more recent attestations are added first
|
|
|
|
let
|
|
|
|
candidate =
|
|
|
|
# Fast path for when all remaining candidates fit
|
|
|
|
if candidates.lenu64 < MAX_ATTESTATIONS: candidates.len - 1
|
|
|
|
else: maxIndex(candidates)
|
2021-05-28 12:51:15 +00:00
|
|
|
# TODO slot not used; replace with _ when
|
|
|
|
# https://github.com/nim-lang/Nim/issues/15972 and
|
|
|
|
# https://github.com/nim-lang/Nim/issues/16217 are
|
|
|
|
# fixed in Status's Nim.
|
2021-04-12 20:25:09 +00:00
|
|
|
(_, slot, entry, j) = candidates[candidate]
|
|
|
|
|
|
|
|
candidates.del(candidate) # careful, `del` reorders candidates
|
|
|
|
|
|
|
|
if entry[].data.target.epoch == prevEpoch:
|
|
|
|
if prevEpochSpace < 1:
|
|
|
|
continue # No need to rescore since we didn't add the attestation
|
|
|
|
|
|
|
|
prevEpochSpace -= 1
|
|
|
|
|
|
|
|
res.add(entry[].toAttestation(entry[].aggregates[j]))
|
|
|
|
|
|
|
|
# Update cache so that the new votes are taken into account when updating
|
|
|
|
# the score below
|
|
|
|
attCache.add(entry[].data, entry[].aggregates[j].aggregation_bits)
|
|
|
|
|
|
|
|
block:
|
|
|
|
# Because we added some votes, it's quite possible that some candidates
|
|
|
|
# are no longer interesting - update the scores of the existing candidates
|
|
|
|
for it in candidates.mitems():
|
|
|
|
it.score = attCache.score(
|
|
|
|
it.entry[].data,
|
|
|
|
it.entry[].aggregates[it.validation].aggregation_bits)
|
|
|
|
|
|
|
|
candidates.keepItIf:
|
|
|
|
# Only keep candidates that might add coverage
|
|
|
|
it.score > 0
|
|
|
|
|
2021-04-14 14:43:29 +00:00
|
|
|
let
|
2021-05-28 16:34:00 +00:00
|
|
|
packingDur = Moment.now() - startPackingTick
|
2021-04-14 14:43:29 +00:00
|
|
|
|
|
|
|
debug "Packed attestations for block",
|
2021-05-28 16:34:00 +00:00
|
|
|
newBlockSlot, packingDur, totalCandidates, attestations = res.len()
|
2021-04-14 14:43:29 +00:00
|
|
|
attestation_pool_block_attestation_packing_time.set(
|
2021-05-28 16:34:00 +00:00
|
|
|
packingDur.toFloatSeconds())
|
2021-04-14 14:43:29 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
res
|
|
|
|
|
2021-10-19 14:09:26 +00:00
|
|
|
proc getAttestationsForBlock*(pool: var AttestationPool,
|
|
|
|
state: ForkedHashedBeaconState,
|
2021-10-01 01:29:32 +00:00
|
|
|
cache: var StateCache): seq[Attestation] =
|
|
|
|
withState(state):
|
|
|
|
pool.getAttestationsForBlock(state, cache)
|
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
func bestValidation(aggregates: openArray[Validation]): (int, int) =
|
|
|
|
# Look for best validation based on number of votes in the aggregate
|
|
|
|
doAssert aggregates.len() > 0,
|
|
|
|
"updateAggregates should have created at least one aggregate"
|
|
|
|
var
|
|
|
|
bestIndex = 0
|
|
|
|
best = aggregates[bestIndex].aggregation_bits.countOnes()
|
|
|
|
|
|
|
|
for i in 1..<aggregates.len():
|
|
|
|
let count = aggregates[i].aggregation_bits.countOnes()
|
|
|
|
if count > best:
|
|
|
|
best = count
|
|
|
|
bestIndex = i
|
|
|
|
(bestIndex, best)
|
|
|
|
|
|
|
|
func getAggregatedAttestation*(pool: var AttestationPool,
|
2020-09-14 11:13:30 +00:00
|
|
|
slot: Slot,
|
2021-04-12 20:25:09 +00:00
|
|
|
attestation_data_root: Eth2Digest): Option[Attestation] =
|
|
|
|
let
|
|
|
|
candidateIdx = pool.candidateIdx(slot)
|
|
|
|
if candidateIdx.isNone:
|
|
|
|
return none(Attestation)
|
|
|
|
|
|
|
|
pool.candidates[candidateIdx.get].withValue(attestation_data_root, entry):
|
|
|
|
entry[].updateAggregates()
|
|
|
|
|
|
|
|
let (bestIndex, _) = bestValidation(entry[].aggregates)
|
|
|
|
|
|
|
|
# Found the right hash, no need to look further
|
|
|
|
return some(entry[].toAttestation(entry[].aggregates[bestIndex]))
|
|
|
|
|
2020-12-15 15:16:10 +00:00
|
|
|
none(Attestation)
|
2020-09-14 11:13:30 +00:00
|
|
|
|
2022-02-17 11:53:55 +00:00
|
|
|
func getAggregatedAttestation*(pool: var AttestationPool,
|
2020-08-21 01:22:26 +00:00
|
|
|
slot: Slot,
|
|
|
|
index: CommitteeIndex): Option[Attestation] =
|
2021-04-12 20:25:09 +00:00
|
|
|
## Select the attestation that has the most votes going for it in the given
|
|
|
|
## slot/index
|
2022-03-02 10:00:21 +00:00
|
|
|
## https://github.com/ethereum/consensus-specs/blob/v1.1.10/specs/phase0/validator.md#construct-aggregate
|
2021-04-12 20:25:09 +00:00
|
|
|
let
|
|
|
|
candidateIdx = pool.candidateIdx(slot)
|
|
|
|
if candidateIdx.isNone:
|
2020-08-21 01:22:26 +00:00
|
|
|
return none(Attestation)
|
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
var res: Option[Attestation]
|
|
|
|
for _, entry in pool.candidates[candidateIdx.get].mpairs():
|
|
|
|
doAssert entry.data.slot == slot
|
2022-01-08 23:28:49 +00:00
|
|
|
if index != entry.data.index:
|
2020-08-21 01:22:26 +00:00
|
|
|
continue
|
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
entry.updateAggregates()
|
2020-08-21 01:22:26 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
let (bestIndex, best) = bestValidation(entry.aggregates)
|
2020-08-21 01:22:26 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
if res.isNone() or best > res.get().aggregation_bits.countOnes():
|
|
|
|
res = some(entry.toAttestation(entry.aggregates[bestIndex]))
|
2020-08-21 01:22:26 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
res
|
2020-08-21 01:22:26 +00:00
|
|
|
|
limit by-root requests to non-finalized blocks (#3293)
* limit by-root requests to non-finalized blocks
Presently, we keep a mapping from block root to `BlockRef` in memory -
this has simplified reasoning about the dag, but is not sustainable with
the chain growing.
We can distinguish between two cases where by-root access is useful:
* unfinalized blocks - this is where the beacon chain is operating
generally, by validating incoming data as interesting for future fork
choice decisions - bounded by the length of the unfinalized period
* finalized blocks - historical access in the REST API etc - no bounds,
really
In this PR, we limit the by-root block index to the first use case:
finalized chain data can more efficiently be addressed by slot number.
Future work includes:
* limiting the `BlockRef` horizon in general - each instance is 40
bytes+overhead which adds up - this needs further refactoring to deal
with the tail vs state problem
* persisting the finalized slot-to-hash index - this one also keeps
growing unbounded (albeit slowly)
Anyway, this PR easily shaves ~128mb of memory usage at the time of
writing.
* No longer honor `BeaconBlocksByRoot` requests outside of the
non-finalized period - previously, Nimbus would generously return any
block through this libp2p request - per the spec, finalized blocks
should be fetched via `BeaconBlocksByRange` instead.
* return `Opt[BlockRef]` instead of `nil` when blocks can't be found -
this becomes a lot more common now and thus deserves more attention
* `dag.blocks` -> `dag.forkBlocks` - this index only carries unfinalized
blocks from now - `finalizedBlocks` covers the other `BlockRef`
instances
* in backfill, verify that the last backfilled block leads back to
genesis, or panic
* add backfill timings to log
* fix missing check that `BlockRef` block can be fetched with
`getForkedBlock` reliably
* shortcut doppelganger check when feature is not enabled
* in REST/JSON-RPC, fetch blocks without involving `BlockRef`
* fix dag.blocks ref
2022-01-21 11:33:16 +00:00
|
|
|
proc selectHead*(pool: var AttestationPool, wallTime: BeaconTime): Opt[BlockRef] =
|
2020-08-26 15:23:34 +00:00
|
|
|
## Trigger fork choice and returns the new head block.
|
|
|
|
## Can return `nil`
|
2021-12-21 18:56:08 +00:00
|
|
|
let newHead = pool.forkChoice.get_head(pool.dag, wallTime)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2020-07-22 09:42:55 +00:00
|
|
|
if newHead.isErr:
|
|
|
|
error "Couldn't select head", err = newHead.error
|
limit by-root requests to non-finalized blocks (#3293)
* limit by-root requests to non-finalized blocks
Presently, we keep a mapping from block root to `BlockRef` in memory -
this has simplified reasoning about the dag, but is not sustainable with
the chain growing.
We can distinguish between two cases where by-root access is useful:
* unfinalized blocks - this is where the beacon chain is operating
generally, by validating incoming data as interesting for future fork
choice decisions - bounded by the length of the unfinalized period
* finalized blocks - historical access in the REST API etc - no bounds,
really
In this PR, we limit the by-root block index to the first use case:
finalized chain data can more efficiently be addressed by slot number.
Future work includes:
* limiting the `BlockRef` horizon in general - each instance is 40
bytes+overhead which adds up - this needs further refactoring to deal
with the tail vs state problem
* persisting the finalized slot-to-hash index - this one also keeps
growing unbounded (albeit slowly)
Anyway, this PR easily shaves ~128mb of memory usage at the time of
writing.
* No longer honor `BeaconBlocksByRoot` requests outside of the
non-finalized period - previously, Nimbus would generously return any
block through this libp2p request - per the spec, finalized blocks
should be fetched via `BeaconBlocksByRange` instead.
* return `Opt[BlockRef]` instead of `nil` when blocks can't be found -
this becomes a lot more common now and thus deserves more attention
* `dag.blocks` -> `dag.forkBlocks` - this index only carries unfinalized
blocks from now - `finalizedBlocks` covers the other `BlockRef`
instances
* in backfill, verify that the last backfilled block leads back to
genesis, or panic
* add backfill timings to log
* fix missing check that `BlockRef` block can be fetched with
`getForkedBlock` reliably
* shortcut doppelganger check when feature is not enabled
* in REST/JSON-RPC, fetch blocks without involving `BlockRef`
* fix dag.blocks ref
2022-01-21 11:33:16 +00:00
|
|
|
err()
|
2020-07-22 09:42:55 +00:00
|
|
|
else:
|
limit by-root requests to non-finalized blocks (#3293)
* limit by-root requests to non-finalized blocks
Presently, we keep a mapping from block root to `BlockRef` in memory -
this has simplified reasoning about the dag, but is not sustainable with
the chain growing.
We can distinguish between two cases where by-root access is useful:
* unfinalized blocks - this is where the beacon chain is operating
generally, by validating incoming data as interesting for future fork
choice decisions - bounded by the length of the unfinalized period
* finalized blocks - historical access in the REST API etc - no bounds,
really
In this PR, we limit the by-root block index to the first use case:
finalized chain data can more efficiently be addressed by slot number.
Future work includes:
* limiting the `BlockRef` horizon in general - each instance is 40
bytes+overhead which adds up - this needs further refactoring to deal
with the tail vs state problem
* persisting the finalized slot-to-hash index - this one also keeps
growing unbounded (albeit slowly)
Anyway, this PR easily shaves ~128mb of memory usage at the time of
writing.
* No longer honor `BeaconBlocksByRoot` requests outside of the
non-finalized period - previously, Nimbus would generously return any
block through this libp2p request - per the spec, finalized blocks
should be fetched via `BeaconBlocksByRange` instead.
* return `Opt[BlockRef]` instead of `nil` when blocks can't be found -
this becomes a lot more common now and thus deserves more attention
* `dag.blocks` -> `dag.forkBlocks` - this index only carries unfinalized
blocks from now - `finalizedBlocks` covers the other `BlockRef`
instances
* in backfill, verify that the last backfilled block leads back to
genesis, or panic
* add backfill timings to log
* fix missing check that `BlockRef` block can be fetched with
`getForkedBlock` reliably
* shortcut doppelganger check when feature is not enabled
* in REST/JSON-RPC, fetch blocks without involving `BlockRef`
* fix dag.blocks ref
2022-01-21 11:33:16 +00:00
|
|
|
let ret = pool.dag.getBlockRef(newHead.get())
|
|
|
|
if ret.isErr():
|
2020-10-28 07:55:36 +00:00
|
|
|
# This should normally not happen, but if the chain dag and fork choice
|
|
|
|
# get out of sync, we'll need to try to download the selected head - in
|
|
|
|
# the meantime, return nil to indicate that no new head was chosen
|
|
|
|
warn "Fork choice selected unknown head, trying to sync", root = newHead.get()
|
2021-12-06 09:49:01 +00:00
|
|
|
pool.quarantine[].addMissing(newHead.get())
|
2020-10-28 07:55:36 +00:00
|
|
|
|
|
|
|
ret
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2020-07-25 19:41:12 +00:00
|
|
|
proc prune*(pool: var AttestationPool) =
|
|
|
|
if (let v = pool.forkChoice.prune(); v.isErr):
|
2020-09-14 14:50:03 +00:00
|
|
|
# If pruning fails, it's likely the result of a bug - this shouldn't happen
|
|
|
|
# but we'll keep running hoping that the fork chocie will recover eventually
|
|
|
|
error "Couldn't prune fork choice, bug?", err = v.error()
|