2019-02-19 23:35:02 +00:00
|
|
|
# beacon_chain
|
2024-01-06 14:26:56 +00:00
|
|
|
# Copyright (c) 2018-2024 Status Research & Development GmbH
|
2019-02-19 23:35:02 +00:00
|
|
|
# Licensed and distributed under either of
|
2019-11-25 15:30:02 +00:00
|
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
2019-02-19 23:35:02 +00:00
|
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
2024-02-18 01:16:49 +00:00
|
|
|
{.push raises: [].}
|
2019-11-14 10:47:55 +00:00
|
|
|
{.used.}
|
|
|
|
|
2019-02-19 23:35:02 +00:00
|
|
|
import
|
2021-01-25 18:45:48 +00:00
|
|
|
# Status lib
|
2021-04-12 20:25:09 +00:00
|
|
|
unittest2,
|
2020-08-27 07:34:12 +00:00
|
|
|
chronicles, chronos,
|
2023-09-12 18:16:04 +00:00
|
|
|
stew/[byteutils, endians2],
|
2023-06-19 22:43:50 +00:00
|
|
|
taskpools,
|
2021-01-25 18:45:48 +00:00
|
|
|
# Internal
|
2021-04-26 20:39:44 +00:00
|
|
|
../beacon_chain/gossip_processing/[gossip_validation],
|
2020-07-30 19:18:17 +00:00
|
|
|
../beacon_chain/fork_choice/[fork_choice_types, fork_choice],
|
2021-04-12 20:25:09 +00:00
|
|
|
../beacon_chain/consensus_object_pools/[
|
2021-06-11 17:51:46 +00:00
|
|
|
block_quarantine, blockchain_dag, block_clearance, attestation_pool],
|
2021-11-18 12:02:43 +00:00
|
|
|
../beacon_chain/spec/[beaconstate, helpers, state_transition, validator],
|
2021-12-21 18:56:08 +00:00
|
|
|
../beacon_chain/beacon_clock,
|
2021-01-25 18:45:48 +00:00
|
|
|
# Test utilities
|
2024-05-17 12:37:41 +00:00
|
|
|
./testutil, ./testdbutil, ./testblockutil, ./consensus_spec/fixtures_utils
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2024-10-14 14:20:26 +00:00
|
|
|
from std/sequtils import mapIt, toSeq
|
2024-02-09 23:46:51 +00:00
|
|
|
from ./testbcutil import addHeadBlock
|
|
|
|
|
2024-09-06 12:15:30 +00:00
|
|
|
func combine(tgt: var (phase0.Attestation | electra.Attestation),
|
|
|
|
src: phase0.Attestation | electra.Attestation) =
|
2020-08-15 17:33:58 +00:00
|
|
|
## Combine the signature and participation bitfield, with the assumption that
|
|
|
|
## the same data is being signed - if the signatures overlap, they are not
|
|
|
|
## combined.
|
|
|
|
|
|
|
|
doAssert tgt.data == src.data
|
|
|
|
|
|
|
|
# In a BLS aggregate signature, one needs to count how many times a
|
|
|
|
# particular public key has been added - since we use a single bit per key, we
|
|
|
|
# can only it once, thus we can never combine signatures that overlap already!
|
2021-04-12 20:25:09 +00:00
|
|
|
doAssert not tgt.aggregation_bits.overlaps(src.aggregation_bits)
|
2020-08-15 17:33:58 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
tgt.aggregation_bits.incl(src.aggregation_bits)
|
|
|
|
|
2022-02-16 22:24:44 +00:00
|
|
|
var agg {.noinit.}: AggregateSignature
|
2021-04-26 20:39:44 +00:00
|
|
|
agg.init(tgt.signature.load().get())
|
|
|
|
agg.aggregate(src.signature.load.get())
|
|
|
|
tgt.signature = agg.finish().toValidatorSig()
|
2020-08-15 17:33:58 +00:00
|
|
|
|
2024-04-17 20:44:29 +00:00
|
|
|
func loadSig(a: phase0.Attestation): CookedSig =
|
2021-04-26 20:39:44 +00:00
|
|
|
a.signature.load.get()
|
2024-05-17 12:37:41 +00:00
|
|
|
func loadSig(a: electra.Attestation): CookedSig =
|
|
|
|
a.signature.load.get()
|
2021-04-09 12:59:24 +00:00
|
|
|
|
2021-03-09 14:36:17 +00:00
|
|
|
proc pruneAtFinalization(dag: ChainDAGRef, attPool: AttestationPool) =
|
|
|
|
if dag.needStateCachesAndForkChoicePruning():
|
|
|
|
dag.pruneStateCachesDAG()
|
|
|
|
# pool[].prune() # We test logic without attestation pool / fork choice pruning
|
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
suite "Attestation pool processing" & preset():
|
2020-04-29 11:44:07 +00:00
|
|
|
## For now just test that we can compile and execute block processing with
|
|
|
|
## mock data.
|
2019-12-19 14:13:35 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
setup:
|
2021-04-12 20:25:09 +00:00
|
|
|
# Genesis state that results in 6 members per committee
|
2023-06-19 22:43:50 +00:00
|
|
|
let rng = HmacDrbgContext.new()
|
2020-04-29 11:44:07 +00:00
|
|
|
var
|
2021-12-20 19:20:31 +00:00
|
|
|
validatorMonitor = newClone(ValidatorMonitor.init())
|
|
|
|
dag = init(
|
|
|
|
ChainDAGRef, defaultRuntimeConfig, makeTestDB(SLOTS_PER_EPOCH * 6),
|
|
|
|
validatorMonitor, {})
|
2021-09-17 00:13:52 +00:00
|
|
|
taskpool = Taskpool.new()
|
2023-08-03 08:36:45 +00:00
|
|
|
verifier = BatchVerifier.init(rng, taskpool)
|
2021-12-06 09:49:01 +00:00
|
|
|
quarantine = newClone(Quarantine.init())
|
2021-06-01 11:13:40 +00:00
|
|
|
pool = newClone(AttestationPool.init(dag, quarantine))
|
|
|
|
state = newClone(dag.headState)
|
2020-09-07 15:04:33 +00:00
|
|
|
cache = StateCache()
|
2021-10-13 14:24:36 +00:00
|
|
|
info = ForkedEpochInfo()
|
2020-04-29 11:44:07 +00:00
|
|
|
# Slot 0 is a finalized slot - won't be making attestations for it..
|
2020-05-19 15:46:29 +00:00
|
|
|
check:
|
2021-06-11 17:51:46 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
dag.cfg, state[], getStateField(state[], slot) + 1, cache, info,
|
2022-01-17 11:19:58 +00:00
|
|
|
{}).isOk()
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2022-11-01 12:23:40 +00:00
|
|
|
test "Attestation from different branch" & preset():
|
|
|
|
# Create two alternate histories with different shufflings
|
|
|
|
check process_slots(
|
|
|
|
dag.cfg, state[], (SLOTS_PER_EPOCH - 2).Slot, cache, info, {}).isOk
|
|
|
|
var state2 = newClone(state[])
|
|
|
|
|
|
|
|
const epoch = 3.Epoch
|
|
|
|
template fillToEpoch(
|
|
|
|
state: ref ForkedHashedBeaconState, cache: var StateCache) =
|
|
|
|
while getStateField(state[], slot).epoch <= epoch:
|
|
|
|
check process_slots(
|
|
|
|
dag.cfg, state[], getStateField(state[], slot) + 1, cache, info,
|
|
|
|
{}).isOk
|
|
|
|
let
|
|
|
|
parent_root = withState(state[]): forkyState.latest_block_root
|
|
|
|
attestations = makeFullAttestations(
|
|
|
|
state[], parent_root, getStateField(state[], slot), cache)
|
|
|
|
blck = addTestBlock(
|
|
|
|
state[], cache, attestations = attestations, cfg = dag.cfg)
|
|
|
|
check dag.addHeadBlock(
|
|
|
|
verifier, blck.phase0Data, OnPhase0BlockAdded(nil)).isOk
|
|
|
|
|
|
|
|
# History 1 contains all odd blocks
|
|
|
|
state.fillToEpoch(cache)
|
|
|
|
|
|
|
|
# History 2 contains all even blocks
|
|
|
|
var cache2 = StateCache()
|
|
|
|
check process_slots(
|
|
|
|
dag.cfg, state2[], getStateField(state2[], slot) + 1, cache2, info,
|
|
|
|
{}).isOk
|
|
|
|
state2.fillToEpoch(cache2)
|
|
|
|
|
|
|
|
# The shuffling for epoch 3 among both chains should now be different
|
|
|
|
let
|
|
|
|
dependent_root1 = withState(state[]): forkyState.attester_dependent_root
|
|
|
|
dependent_root2 = withState(state2[]): forkyState.attester_dependent_root
|
|
|
|
check dependent_root1 != dependent_root2
|
|
|
|
|
|
|
|
# Fill pool with attestations from both chains
|
|
|
|
let
|
|
|
|
cIndex = 0.CommitteeIndex
|
|
|
|
att1 = block:
|
|
|
|
let
|
|
|
|
slot = getStateField(state[], slot)
|
|
|
|
parent_root = withState(state[]): forkyState.latest_block_root
|
|
|
|
committee = get_beacon_committee(state[], slot, cIndex, cache)
|
|
|
|
makeAttestation(state[], parent_root, committee[0], cache)
|
|
|
|
att2 = block:
|
|
|
|
let
|
|
|
|
slot = getStateField(state2[], slot)
|
|
|
|
parent_root = withState(state2[]): forkyState.latest_block_root
|
|
|
|
committee = get_beacon_committee(state2[], slot, cIndex, cache2)
|
|
|
|
makeAttestation(state2[], parent_root, committee[0], cache2)
|
|
|
|
maxSlot = max(att1.data.slot, att2.data.slot)
|
|
|
|
|
|
|
|
# Advance time so attestations become valid
|
|
|
|
check:
|
|
|
|
process_slots(
|
|
|
|
dag.cfg, state[], maxSlot + MIN_ATTESTATION_INCLUSION_DELAY,
|
|
|
|
cache, info, {}).isOk
|
|
|
|
process_slots(
|
|
|
|
dag.cfg, state2[], maxSlot + MIN_ATTESTATION_INCLUSION_DELAY,
|
|
|
|
cache2, info, {}).isOk
|
|
|
|
|
|
|
|
# They should remain valid only within a compatible state
|
|
|
|
withState(state[]):
|
|
|
|
check:
|
|
|
|
check_attestation(forkyState.data, att1, {}, cache).isOk
|
|
|
|
check_attestation(forkyState.data, att2, {}, cache).isErr
|
|
|
|
withState(state2[]):
|
|
|
|
check:
|
|
|
|
check_attestation(forkyState.data, att1, {}, cache2).isErr
|
|
|
|
check_attestation(forkyState.data, att2, {}, cache2).isOk
|
|
|
|
|
|
|
|
# If signature checks are skipped, state incompatibility is not detected
|
|
|
|
let flags = {skipBlsValidation}
|
|
|
|
withState(state[]):
|
|
|
|
check:
|
|
|
|
check_attestation(forkyState.data, att1, flags, cache).isOk
|
|
|
|
check_attestation(forkyState.data, att2, flags, cache).isOk
|
|
|
|
withState(state2[]):
|
|
|
|
check:
|
|
|
|
check_attestation(forkyState.data, att1, flags, cache2).isOk
|
|
|
|
check_attestation(forkyState.data, att2, flags, cache2).isOk
|
|
|
|
|
|
|
|
# An additional compatibility check catches that (used in block production)
|
|
|
|
withState(state[]):
|
|
|
|
check:
|
2023-06-04 10:39:12 +00:00
|
|
|
check_attestation_compatible(dag, forkyState, att1).isOk
|
|
|
|
check_attestation_compatible(dag, forkyState, att2).isErr
|
2022-11-01 12:23:40 +00:00
|
|
|
withState(state2[]):
|
|
|
|
check:
|
2023-06-04 10:39:12 +00:00
|
|
|
check_attestation_compatible(dag, forkyState, att1).isErr
|
|
|
|
check_attestation_compatible(dag, forkyState, att2).isOk
|
2022-11-01 12:23:40 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Can add and retrieve simple attestations" & preset():
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
|
|
|
# Create an attestation for slot 1!
|
2021-04-12 20:25:09 +00:00
|
|
|
bc0 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
attestation = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2020-08-27 07:34:12 +00:00
|
|
|
pool[].addAttestation(
|
2021-04-12 20:25:09 +00:00
|
|
|
attestation, @[bc0[0]], attestation.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation.data.slot.start_beacon_time)
|
2019-12-19 14:13:35 +00:00
|
|
|
|
2020-05-19 15:46:29 +00:00
|
|
|
check:
|
2021-04-12 20:25:09 +00:00
|
|
|
# Added attestation, should get it back
|
2022-07-06 16:11:44 +00:00
|
|
|
toSeq(pool[].attestations(Opt.none(Slot), Opt.none(CommitteeIndex))) ==
|
2021-04-12 20:25:09 +00:00
|
|
|
@[attestation]
|
|
|
|
toSeq(pool[].attestations(
|
2022-07-06 16:11:44 +00:00
|
|
|
Opt.some(attestation.data.slot), Opt.none(CommitteeIndex))) ==
|
|
|
|
@[attestation]
|
2021-04-12 20:25:09 +00:00
|
|
|
toSeq(pool[].attestations(
|
2022-07-06 16:11:44 +00:00
|
|
|
Opt.some(attestation.data.slot), Opt.some(attestation.data.index.CommitteeIndex))) ==
|
2021-04-12 20:25:09 +00:00
|
|
|
@[attestation]
|
2022-07-06 16:11:44 +00:00
|
|
|
toSeq(pool[].attestations(Opt.none(Slot), Opt.some(attestation.data.index.CommitteeIndex))) ==
|
2021-04-12 20:25:09 +00:00
|
|
|
@[attestation]
|
2022-07-06 16:11:44 +00:00
|
|
|
toSeq(pool[].attestations(Opt.some(
|
|
|
|
attestation.data.slot + 1), Opt.none(CommitteeIndex))) == []
|
2021-04-12 20:25:09 +00:00
|
|
|
toSeq(pool[].attestations(
|
2022-07-06 16:11:44 +00:00
|
|
|
Opt.none(Slot), Opt.some(CommitteeIndex(attestation.data.index + 1)))) == []
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
2022-01-17 11:19:58 +00:00
|
|
|
info, {}).isOk()
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2019-12-19 14:13:35 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
|
|
|
attestations.len == 1
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(1.Slot, 0.CommitteeIndex).isSome()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
let
|
|
|
|
root1 = addTestBlock(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], cache, attestations = attestations,
|
2021-11-18 12:02:43 +00:00
|
|
|
nextSlot = false).phase0Data.root
|
2021-04-12 20:25:09 +00:00
|
|
|
bc1 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
att1 = makeAttestation(state[], root1, bc1[0], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
check:
|
2022-09-10 06:12:07 +00:00
|
|
|
withState(state[]): forkyState.latest_block_root == root1
|
2021-11-18 12:02:43 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
2022-01-17 11:19:58 +00:00
|
|
|
info, {}).isOk()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
2022-09-10 06:12:07 +00:00
|
|
|
withState(state[]): forkyState.latest_block_root == root1
|
2021-11-18 12:02:43 +00:00
|
|
|
|
2021-04-12 20:25:09 +00:00
|
|
|
check:
|
|
|
|
# shouldn't include already-included attestations
|
2022-03-16 07:20:40 +00:00
|
|
|
pool[].getAttestationsForBlock(state[], cache) == []
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att1, @[bc1[0]], att1.loadSig, att1.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
# but new ones should go in
|
2022-03-16 07:20:40 +00:00
|
|
|
pool[].getAttestationsForBlock(state[], cache).len() == 1
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
att2 = makeAttestation(state[], root1, bc1[1], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att2, @[bc1[1]], att2.loadSig, att2.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
combined = pool[].getAttestationsForBlock(state[], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
# New attestations should be combined with old attestations
|
|
|
|
combined.len() == 1
|
|
|
|
combined[0].aggregation_bits.countOnes() == 2
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
combined[0], @[bc1[1], bc1[0]], combined[0].loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
combined[0].data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
# readding the combined attestation shouldn't have an effect
|
2022-03-16 07:20:40 +00:00
|
|
|
pool[].getAttestationsForBlock(state[], cache).len() == 1
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
let
|
|
|
|
# Someone votes for a different root
|
2022-06-18 04:57:37 +00:00
|
|
|
att3 = makeAttestation(state[], ZERO_HASH, bc1[2], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att3, @[bc1[2]], att3.loadSig, att3.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
# We should now get both attestations for the block, but the aggregate
|
|
|
|
# should be the one with the most votes
|
2022-03-16 07:20:40 +00:00
|
|
|
pool[].getAttestationsForBlock(state[], cache).len() == 2
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(2.Slot, 0.CommitteeIndex).
|
2021-04-12 20:25:09 +00:00
|
|
|
get().aggregation_bits.countOnes() == 2
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(2.Slot, hash_tree_root(att2.data)).
|
2021-04-12 20:25:09 +00:00
|
|
|
get().aggregation_bits.countOnes() == 2
|
|
|
|
|
|
|
|
let
|
|
|
|
# Someone votes for a different root
|
2022-06-18 04:57:37 +00:00
|
|
|
att4 = makeAttestation(state[], ZERO_HASH, bc1[2], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att4, @[bc1[2]], att3.loadSig, att3.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Working with aggregates" & preset():
|
2021-04-12 20:25:09 +00:00
|
|
|
let
|
|
|
|
# Create an attestation for slot 1!
|
|
|
|
bc0 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
var
|
2022-03-16 07:20:40 +00:00
|
|
|
att0 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
att0x = att0
|
2022-03-16 07:20:40 +00:00
|
|
|
att1 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[1], cache)
|
|
|
|
att2 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[2], cache)
|
|
|
|
att3 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[3], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
# Both attestations include member 2 but neither is a subset of the other
|
|
|
|
att0.combine(att2)
|
|
|
|
att1.combine(att2)
|
|
|
|
|
2022-02-25 16:15:39 +00:00
|
|
|
check:
|
|
|
|
not pool[].covers(att0.data, att0.aggregation_bits)
|
|
|
|
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att0, @[bc0[0], bc0[2]], att0.loadSig, att0.data.slot.start_beacon_time)
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att1, @[bc0[1], bc0[2]], att1.loadSig, att1.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
2022-01-17 11:19:58 +00:00
|
|
|
info, {}).isOk()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
check:
|
2022-02-25 16:15:39 +00:00
|
|
|
pool[].covers(att0.data, att0.aggregation_bits)
|
2022-03-16 07:20:40 +00:00
|
|
|
pool[].getAttestationsForBlock(state[], cache).len() == 2
|
2021-04-12 20:25:09 +00:00
|
|
|
# Can get either aggregate here, random!
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(1.Slot, 0.CommitteeIndex).isSome()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
# Add in attestation 3 - both aggregates should now have it added
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att3, @[bc0[3]], att3.loadSig, att3.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
block:
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
check:
|
|
|
|
attestations.len() == 2
|
|
|
|
attestations[0].aggregation_bits.countOnes() == 3
|
|
|
|
# Can get either aggregate here, random!
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(1.Slot, 0.CommitteeIndex).isSome()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
# Add in attestation 0 as single - attestation 1 is now a superset of the
|
|
|
|
# aggregates in the pool, so everything else should be removed
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att0x, @[bc0[0]], att0x.loadSig, att0x.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
block:
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
check:
|
|
|
|
attestations.len() == 1
|
|
|
|
attestations[0].aggregation_bits.countOnes() == 4
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(1.Slot, 0.CommitteeIndex).isSome()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Everyone voting for something different" & preset():
|
2021-04-12 20:25:09 +00:00
|
|
|
var attestations: int
|
|
|
|
for i in 0..<SLOTS_PER_EPOCH:
|
|
|
|
var root: Eth2Digest
|
|
|
|
root.data[0..<8] = toBytesBE(i.uint64)
|
|
|
|
let
|
|
|
|
bc0 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
for j in 0..<bc0.len():
|
|
|
|
root.data[8..<16] = toBytesBE(j.uint64)
|
2022-03-16 07:20:40 +00:00
|
|
|
let att = makeAttestation(state[], root, bc0[j], cache)
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addAttestation(
|
2022-01-11 10:01:54 +00:00
|
|
|
att, @[bc0[j]], att.loadSig, att.data.slot.start_beacon_time)
|
2021-04-12 20:25:09 +00:00
|
|
|
inc attestations
|
|
|
|
|
|
|
|
check:
|
Implement split preset/config support (#2710)
* Implement split preset/config support
This is the initial bulk refactor to introduce runtime config values in
a number of places, somewhat replacing the existing mechanism of loading
network metadata.
It still needs more work, this is the initial refactor that introduces
runtime configuration in some of the places that need it.
The PR changes the way presets and constants work, to match the spec. In
particular, a "preset" now refers to the compile-time configuration
while a "cfg" or "RuntimeConfig" is the dynamic part.
A single binary can support either mainnet or minimal, but not both.
Support for other presets has been removed completely (can be readded,
in case there's need).
There's a number of outstanding tasks:
* `SECONDS_PER_SLOT` still needs fixing
* loading custom runtime configs needs redoing
* checking constants against YAML file
* yeerongpilly support
`build/nimbus_beacon_node --network=yeerongpilly --discv5:no --log-level=DEBUG`
* load fork epoch from config
* fix fork digest sent in status
* nicer error string for request failures
* fix tools
* one more
* fixup
* fixup
* fixup
* use "standard" network definition folder in local testnet
Files are loaded from their standard locations, including genesis etc,
to conform to the format used in the `eth2-networks` repo.
* fix launch scripts, allow unknown config values
* fix base config of rest test
* cleanups
* bundle mainnet config using common loader
* fix spec links and names
* only include supported preset in binary
* drop yeerongpilly, add altair-devnet-0, support boot_enr.yaml
2021-07-12 13:01:38 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + 1, cache, info, {}).isOk()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
|
|
|
doAssert attestations.uint64 > MAX_ATTESTATIONS,
|
|
|
|
"6*SLOTS_PER_EPOCH validators > 128 mainnet MAX_ATTESTATIONS"
|
|
|
|
check:
|
|
|
|
# Fill block with attestations
|
2022-03-16 07:20:40 +00:00
|
|
|
pool[].getAttestationsForBlock(state[], cache).lenu64() ==
|
2021-04-12 20:25:09 +00:00
|
|
|
MAX_ATTESTATIONS
|
2024-10-07 07:53:38 +00:00
|
|
|
pool[].getPhase0AggregatedAttestation(
|
2022-03-16 07:20:40 +00:00
|
|
|
getStateField(state[], slot) - 1, 0.CommitteeIndex).isSome()
|
2021-04-12 20:25:09 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Attestations may arrive in any order" & preset():
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
|
|
|
# Create an attestation for slot 1!
|
|
|
|
bc0 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
attestation0 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2020-05-19 15:46:29 +00:00
|
|
|
check:
|
2021-06-11 17:51:46 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[], getStateField(state[], slot) + 1,
|
2022-01-17 11:19:58 +00:00
|
|
|
cache, info, {}).isOk()
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
bc1 = get_beacon_committee(state[],
|
|
|
|
getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
attestation1 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc1[0], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
# test reverse order
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation1, @[bc1[0]], attestation1.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation1.data.slot.start_beacon_time)
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation0, @[bc0[0]], attestation0.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation0.data.slot.start_beacon_time)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
attestations.len == 1
|
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Attestations should be combined" & preset():
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
|
|
|
# Create an attestation for slot 1!
|
|
|
|
bc0 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
2021-06-11 17:51:46 +00:00
|
|
|
attestation0 =
|
2022-03-16 07:20:40 +00:00
|
|
|
makeAttestation(state[], state[].latest_block_root, bc0[0], cache)
|
2021-06-11 17:51:46 +00:00
|
|
|
attestation1 =
|
2022-03-16 07:20:40 +00:00
|
|
|
makeAttestation(state[], state[].latest_block_root, bc0[1], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation0, @[bc0[0]], attestation0.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation0.data.slot.start_beacon_time)
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation1, @[bc0[1]], attestation1.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation1.data.slot.start_beacon_time)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2020-05-19 15:46:29 +00:00
|
|
|
check:
|
2021-05-07 11:36:21 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
2022-01-17 11:19:58 +00:00
|
|
|
MIN_ATTESTATION_INCLUSION_DELAY.Slot + 1, cache, info, {}).isOk()
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
attestations.len == 1
|
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Attestations may overlap, bigger first" & preset():
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
var
|
|
|
|
# Create an attestation for slot 1!
|
|
|
|
bc0 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
attestation0 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
|
|
|
attestation1 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[1], cache)
|
2019-12-19 14:13:35 +00:00
|
|
|
|
2021-01-22 13:29:04 +00:00
|
|
|
attestation0.combine(attestation1)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation0, @[bc0[0]], attestation0.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation0.data.slot.start_beacon_time)
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation1, @[bc0[1]], attestation1.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation1.data.slot.start_beacon_time)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-05-19 15:46:29 +00:00
|
|
|
check:
|
2021-05-07 11:36:21 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
2022-01-17 11:19:58 +00:00
|
|
|
MIN_ATTESTATION_INCLUSION_DELAY.Slot + 1, cache, info, {}).isOk()
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
|
|
|
attestations.len == 1
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Attestations may overlap, smaller first" & preset():
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-04-29 11:44:07 +00:00
|
|
|
var
|
|
|
|
# Create an attestation for slot 1!
|
2022-03-16 07:20:40 +00:00
|
|
|
bc0 = get_beacon_committee(state[],
|
|
|
|
getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
attestation0 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
|
|
|
attestation1 = makeAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[1], cache)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2021-01-22 13:29:04 +00:00
|
|
|
attestation0.combine(attestation1)
|
2019-12-19 14:13:35 +00:00
|
|
|
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation1, @[bc0[1]], attestation1.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation1.data.slot.start_beacon_time)
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation0, @[bc0[0]], attestation0.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation0.data.slot.start_beacon_time)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-05-19 15:46:29 +00:00
|
|
|
check:
|
2021-05-07 11:36:21 +00:00
|
|
|
process_slots(
|
2022-03-16 07:20:40 +00:00
|
|
|
defaultRuntimeConfig, state[],
|
2022-01-17 11:19:58 +00:00
|
|
|
MIN_ATTESTATION_INCLUSION_DELAY.Slot + 1, cache, info, {}).isOk()
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2022-03-16 07:20:40 +00:00
|
|
|
let attestations = pool[].getAttestationsForBlock(state[], cache)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
|
|
|
attestations.len == 1
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Fork choice returns latest block with no attestations":
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
b1 = addTestBlock(state[], cache).phase0Data
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
b1Add = dag.addHeadBlock(verifier, b1) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-22 09:42:55 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head =
|
|
|
|
pool[].selectOptimisticHead(b1Add[].slot.start_beacon_time).get().blck
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
2020-07-09 09:29:32 +00:00
|
|
|
head == b1Add[]
|
2019-12-19 14:13:35 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
b2 = addTestBlock(state[], cache).phase0Data
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
b2Add = dag.addHeadBlock(verifier, b2) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-06-10 06:58:12 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head2 =
|
|
|
|
pool[].selectOptimisticHead(b2Add[].slot.start_beacon_time).get().blck
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
2020-07-09 09:29:32 +00:00
|
|
|
head2 == b2Add[]
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Fork choice returns block with attestation":
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
b10 = makeTestBlock(state[], cache).phase0Data
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
b10Add = dag.addHeadBlock(verifier, b10) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-06-10 06:58:12 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head =
|
|
|
|
pool[].selectOptimisticHead(b10Add[].slot.start_beacon_time).get().blck
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
2020-07-09 09:29:32 +00:00
|
|
|
head == b10Add[]
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2022-04-12 10:06:30 +00:00
|
|
|
# Add a block too late to be timely enough to be proposer-boosted, which
|
|
|
|
# would otherwise cause it to be selected as head
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
b11 = makeTestBlock(state[], cache,
|
2020-06-29 17:30:19 +00:00
|
|
|
graffiti = GraffitiBytes [1'u8, 0, 0, 0 ,0 ,0 ,0 ,0 ,0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
|
2021-10-18 16:37:27 +00:00
|
|
|
).phase0Data
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
b11Add = dag.addHeadBlock(verifier, b11) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
2022-04-12 10:06:30 +00:00
|
|
|
blckRef.slot.start_beacon_time + SECONDS_PER_SLOT.int64.seconds)
|
2019-12-19 13:02:28 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
bc1 = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot) - 1, 1.CommitteeIndex,
|
2021-06-11 17:51:46 +00:00
|
|
|
cache)
|
2022-03-16 07:20:40 +00:00
|
|
|
attestation0 = makeAttestation(state[], b10.root, bc1[0], cache)
|
2019-06-03 08:26:38 +00:00
|
|
|
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation0, @[bc1[0]], attestation0.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation0.data.slot.start_beacon_time)
|
2019-06-03 08:26:38 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head2 =
|
|
|
|
pool[].selectOptimisticHead(b10Add[].slot.start_beacon_time).get().blck
|
2019-06-03 08:26:38 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
check:
|
|
|
|
# Single vote for b10 and no votes for b11
|
2020-07-09 09:29:32 +00:00
|
|
|
head2 == b10Add[]
|
2019-06-03 08:26:38 +00:00
|
|
|
|
2020-04-29 11:44:07 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
attestation1 = makeAttestation(state[], b11.root, bc1[1], cache)
|
|
|
|
attestation2 = makeAttestation(state[], b11.root, bc1[2], cache)
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation1, @[bc1[1]], attestation1.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation1.data.slot.start_beacon_time)
|
2019-06-03 08:26:38 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head3 =
|
|
|
|
pool[].selectOptimisticHead(b10Add[].slot.start_beacon_time).get().blck
|
2020-07-25 19:41:12 +00:00
|
|
|
let bigger = if b11.root.data < b10.root.data: b10Add else: b11Add
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
check:
|
2020-06-10 06:58:12 +00:00
|
|
|
# Ties broken lexicographically in spec -> ?
|
2020-07-25 19:41:12 +00:00
|
|
|
head3 == bigger[]
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2021-04-09 12:59:24 +00:00
|
|
|
pool[].addAttestation(
|
2021-12-21 18:56:08 +00:00
|
|
|
attestation2, @[bc1[2]], attestation2.loadSig,
|
2022-01-11 10:01:54 +00:00
|
|
|
attestation2.data.slot.start_beacon_time)
|
2020-04-29 11:44:07 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head4 =
|
|
|
|
pool[].selectOptimisticHead(b11Add[].slot.start_beacon_time).get().blck
|
2020-04-29 11:44:07 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
# Two votes for b11
|
2020-07-09 09:29:32 +00:00
|
|
|
head4 == b11Add[]
|
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Trying to add a block twice tags the second as an error":
|
2020-07-15 10:44:18 +00:00
|
|
|
var cache = StateCache()
|
2020-07-09 09:29:32 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
b10 = makeTestBlock(state[], cache).phase0Data
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
b10Add = dag.addHeadBlock(verifier, b10) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2022-07-06 10:33:02 +00:00
|
|
|
pool[].addForkChoice(
|
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head =
|
|
|
|
pool[].selectOptimisticHead(b10Add[].slot.start_beacon_time).get().blck
|
2020-07-09 09:29:32 +00:00
|
|
|
|
|
|
|
check:
|
|
|
|
head == b10Add[]
|
|
|
|
|
|
|
|
# -------------------------------------------------------------
|
|
|
|
# Add back the old block to ensure we have a duplicate error
|
|
|
|
let b10_clone = b10 # Assumes deep copy
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
let b10Add_clone = dag.addHeadBlock(verifier, b10_clone) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-22 09:42:55 +00:00
|
|
|
|
2022-11-10 17:40:27 +00:00
|
|
|
doAssert: b10Add_clone.error == VerifierError.Duplicate
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2021-04-28 16:41:02 +00:00
|
|
|
test "Trying to add a duplicate block from an old pruned epoch is tagged as an error":
|
2020-08-26 15:23:34 +00:00
|
|
|
# Note: very sensitive to stack usage
|
|
|
|
|
2022-04-14 15:39:37 +00:00
|
|
|
dag.updateFlags.incl {skipBlsValidation}
|
2020-07-23 17:01:07 +00:00
|
|
|
var cache = StateCache()
|
2020-07-09 09:29:32 +00:00
|
|
|
let
|
2022-03-16 07:20:40 +00:00
|
|
|
b10 = addTestBlock(state[], cache).phase0Data
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
b10Add = dag.addHeadBlock(verifier, b10) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head =
|
|
|
|
pool[].selectOptimisticHead(b10Add[].slot.start_beacon_time).get().blck
|
2020-07-09 09:29:32 +00:00
|
|
|
|
|
|
|
doAssert: head == b10Add[]
|
|
|
|
|
|
|
|
# -------------------------------------------------------------
|
|
|
|
let b10_clone = b10 # Assumes deep copy
|
|
|
|
|
|
|
|
# -------------------------------------------------------------
|
|
|
|
# Pass an epoch
|
2024-04-17 20:44:29 +00:00
|
|
|
var attestations: seq[phase0.Attestation]
|
2020-07-09 09:29:32 +00:00
|
|
|
|
|
|
|
for epoch in 0 ..< 5:
|
2022-01-11 10:01:54 +00:00
|
|
|
let start_slot = start_slot(Epoch epoch)
|
2020-07-25 19:41:12 +00:00
|
|
|
let committees_per_slot =
|
2022-03-16 07:20:40 +00:00
|
|
|
get_committee_count_per_slot(state[], Epoch epoch, cache)
|
2020-07-09 09:29:32 +00:00
|
|
|
for slot in start_slot ..< start_slot + SLOTS_PER_EPOCH:
|
2020-08-27 07:34:12 +00:00
|
|
|
let new_block = addTestBlock(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], cache, attestations = attestations).phase0Data
|
2020-07-09 09:29:32 +00:00
|
|
|
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
let blockRef = dag.addHeadBlock(verifier, new_block) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2022-08-25 23:34:02 +00:00
|
|
|
let head =
|
|
|
|
pool[].selectOptimisticHead(
|
|
|
|
blockRef[].slot.start_beacon_time).get().blck
|
2021-04-12 20:25:09 +00:00
|
|
|
doAssert: head == blockRef[]
|
2023-03-02 16:13:35 +00:00
|
|
|
dag.updateHead(head, quarantine[], [])
|
2021-06-01 11:13:40 +00:00
|
|
|
pruneAtFinalization(dag, pool[])
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2022-04-14 10:47:14 +00:00
|
|
|
attestations.setLen(0)
|
2022-01-08 23:28:49 +00:00
|
|
|
for committee_index in get_committee_indices(committees_per_slot):
|
2020-07-09 09:29:32 +00:00
|
|
|
let committee = get_beacon_committee(
|
2022-03-16 07:20:40 +00:00
|
|
|
state[], getStateField(state[], slot), committee_index,
|
2021-04-16 08:49:37 +00:00
|
|
|
cache)
|
2020-07-09 09:29:32 +00:00
|
|
|
|
|
|
|
# Create a bitfield filled with the given count per attestation,
|
|
|
|
# exactly on the right-most part of the committee field.
|
|
|
|
var aggregation_bits = init(CommitteeValidatorsBits, committee.len)
|
|
|
|
for v in 0 ..< committee.len * 2 div 3 + 1:
|
|
|
|
aggregation_bits[v] = true
|
|
|
|
|
2024-04-17 20:44:29 +00:00
|
|
|
attestations.add phase0.Attestation(
|
2020-07-09 09:29:32 +00:00
|
|
|
aggregation_bits: aggregation_bits,
|
2022-03-16 07:20:40 +00:00
|
|
|
data: makeAttestationData(state[], getStateField(state[], slot),
|
2022-01-08 23:28:49 +00:00
|
|
|
committee_index, blockRef.get().root)
|
2020-07-09 09:29:32 +00:00
|
|
|
# signature: ValidatorSig()
|
|
|
|
)
|
|
|
|
|
2020-07-15 10:44:18 +00:00
|
|
|
cache = StateCache()
|
2020-07-09 09:29:32 +00:00
|
|
|
|
|
|
|
# -------------------------------------------------------------
|
|
|
|
# Prune
|
|
|
|
|
2021-06-01 11:13:40 +00:00
|
|
|
doAssert: dag.finalizedHead.slot != 0
|
2020-07-09 09:29:32 +00:00
|
|
|
|
2020-07-25 19:41:12 +00:00
|
|
|
pool[].prune()
|
2020-08-27 07:34:12 +00:00
|
|
|
doAssert: b10.root notin pool.forkChoice.backend
|
2020-07-09 09:29:32 +00:00
|
|
|
|
|
|
|
# Add back the old block to ensure we have a duplicate error
|
Backfill support for ChainDAG (#3171)
In the ChainDAG, 3 block pointers are kept: genesis, tail and head. This
PR adds one more block pointer: the backfill block which represents the
block that has been backfilled so far.
When doing a checkpoint sync, a random block is given as starting point
- this is the tail block, and we require that the tail block has a
corresponding state.
When backfilling, we end up with blocks without corresponding states,
hence we cannot use `tail` as a backfill pointer - there is no state.
Nonetheless, we need to keep track of where we are in the backfill
process between restarts, such that we can answer GetBeaconBlocksByRange
requests.
This PR adds the basic support for backfill handling - it needs to be
integrated with backfill sync, and the REST API needs to be adjusted to
take advantage of the new backfilled blocks when responding to certain
requests.
Future work will also enable moving the tail in either direction:
* pruning means moving the tail forward in time and removing states
* backwards means recreating past states from genesis, such that
intermediate states are recreated step by step all the way to the tail -
at that point, tail, genesis and backfill will match up.
* backfilling is done when backfill != genesis - later, this will be the
WSS checkpoint instead
2021-12-13 13:36:06 +00:00
|
|
|
let b10Add_clone = dag.addHeadBlock(verifier, b10_clone) do (
|
2021-08-12 13:08:20 +00:00
|
|
|
blckRef: BlockRef, signedBlock: phase0.TrustedSignedBeaconBlock,
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef: EpochRef, unrealized: FinalityCheckpoints):
|
2020-07-22 09:42:55 +00:00
|
|
|
# Callback add to fork choice if valid
|
2021-12-21 18:56:08 +00:00
|
|
|
pool[].addForkChoice(
|
2022-07-06 10:33:02 +00:00
|
|
|
epochRef, blckRef, unrealized, signedBlock.message,
|
|
|
|
blckRef.slot.start_beacon_time)
|
2020-07-22 09:42:55 +00:00
|
|
|
|
2024-05-17 12:37:41 +00:00
|
|
|
doAssert: b10Add_clone.error == VerifierError.Duplicate
|
|
|
|
|
|
|
|
suite "Attestation pool electra processing" & preset():
|
|
|
|
## For now just test that we can compile and execute block processing with
|
|
|
|
## mock data.
|
|
|
|
|
|
|
|
setup:
|
2024-08-23 11:26:35 +00:00
|
|
|
# Genesis state that results in 6 members per committee (2 committees total)
|
|
|
|
const TOTAL_COMMITTEES = 2
|
2024-05-17 12:37:41 +00:00
|
|
|
let rng = HmacDrbgContext.new()
|
|
|
|
var
|
|
|
|
validatorMonitor = newClone(ValidatorMonitor.init())
|
|
|
|
cfg = genesisTestRuntimeConfig(ConsensusFork.Electra)
|
|
|
|
dag = init(
|
|
|
|
ChainDAGRef, cfg,
|
2024-09-19 05:46:38 +00:00
|
|
|
makeTestDB(
|
|
|
|
TOTAL_COMMITTEES * TARGET_COMMITTEE_SIZE * SLOTS_PER_EPOCH, cfg = cfg),
|
2024-05-17 12:37:41 +00:00
|
|
|
validatorMonitor, {})
|
|
|
|
taskpool = Taskpool.new()
|
|
|
|
verifier = BatchVerifier.init(rng, taskpool)
|
|
|
|
quarantine = newClone(Quarantine.init())
|
|
|
|
pool = newClone(AttestationPool.init(dag, quarantine))
|
|
|
|
state = newClone(dag.headState)
|
|
|
|
cache = StateCache()
|
|
|
|
info = ForkedEpochInfo()
|
|
|
|
# Slot 0 is a finalized slot - won't be making attestations for it..
|
|
|
|
check:
|
|
|
|
process_slots(
|
2024-08-23 11:26:35 +00:00
|
|
|
dag.cfg,
|
|
|
|
state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY,
|
|
|
|
cache,
|
|
|
|
info,
|
2024-05-17 12:37:41 +00:00
|
|
|
{}).isOk()
|
|
|
|
|
|
|
|
test "Can add and retrieve simple electra attestations" & preset():
|
|
|
|
let
|
|
|
|
# Create an attestation for slot 1!
|
|
|
|
bc0 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
attestation = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
attestation, @[bc0[0]], attestation.loadSig,
|
|
|
|
attestation.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
check:
|
|
|
|
process_slots(
|
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
|
|
|
info, {}).isOk()
|
|
|
|
|
|
|
|
let attestations = pool[].getElectraAttestationsForBlock(state[], cache)
|
|
|
|
|
|
|
|
check:
|
|
|
|
attestations.len == 1
|
|
|
|
|
|
|
|
let
|
|
|
|
root1 = addTestBlock(
|
|
|
|
state[], cache, electraAttestations = attestations,
|
|
|
|
nextSlot = false).electraData.root
|
|
|
|
bc1 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
att1 = makeElectraAttestation(state[], root1, bc1[0], cache)
|
|
|
|
|
|
|
|
check:
|
|
|
|
withState(state[]): forkyState.latest_block_root == root1
|
|
|
|
|
|
|
|
process_slots(
|
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
|
|
|
info, {}).isOk()
|
|
|
|
|
|
|
|
withState(state[]): forkyState.latest_block_root == root1
|
|
|
|
|
|
|
|
check:
|
|
|
|
# shouldn't include already-included attestations
|
|
|
|
pool[].getElectraAttestationsForBlock(state[], cache) == []
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
att1, @[bc1[0]], att1.loadSig, att1.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
check:
|
|
|
|
# but new ones should go in
|
|
|
|
pool[].getElectraAttestationsForBlock(state[], cache).len() == 1
|
|
|
|
|
|
|
|
let
|
|
|
|
att2 = makeElectraAttestation(state[], root1, bc1[1], cache)
|
|
|
|
pool[].addAttestation(
|
|
|
|
att2, @[bc1[1]], att2.loadSig, att2.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
let
|
|
|
|
combined = pool[].getElectraAttestationsForBlock(state[], cache)
|
|
|
|
|
|
|
|
check:
|
|
|
|
# New attestations should be combined with old attestations
|
|
|
|
combined.len() == 1
|
|
|
|
combined[0].aggregation_bits.countOnes() == 2
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
combined[0], @[bc1[1], bc1[0]], combined[0].loadSig,
|
|
|
|
combined[0].data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
check:
|
|
|
|
# readding the combined attestation shouldn't have an effect
|
|
|
|
pool[].getElectraAttestationsForBlock(state[], cache).len() == 1
|
|
|
|
|
|
|
|
let
|
|
|
|
# Someone votes for a different root
|
|
|
|
att3 = makeElectraAttestation(state[], ZERO_HASH, bc1[2], cache)
|
|
|
|
pool[].addAttestation(
|
|
|
|
att3, @[bc1[2]], att3.loadSig, att3.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
check:
|
|
|
|
# We should now get both attestations for the block, but the aggregate
|
|
|
|
# should be the one with the most votes
|
|
|
|
pool[].getElectraAttestationsForBlock(state[], cache).len() == 2
|
2024-10-04 05:15:48 +00:00
|
|
|
pool[].getElectraAggregatedAttestation(2.Slot, hash_tree_root(combined[0].data),
|
2024-09-06 12:15:30 +00:00
|
|
|
0.CommitteeIndex).get().aggregation_bits.countOnes() == 2
|
|
|
|
pool[].getElectraAggregatedAttestation(2.Slot, hash_tree_root(att2.data), 0.CommitteeIndex).
|
|
|
|
get().aggregation_bits.countOnes() == 2
|
|
|
|
# requests to get and aggregate from different committees should be empty
|
|
|
|
pool[].getElectraAggregatedAttestation(
|
|
|
|
2.Slot, combined[0].data.beacon_block_root, 1.CommitteeIndex).isNone()
|
2024-05-17 12:37:41 +00:00
|
|
|
|
2024-08-23 11:26:35 +00:00
|
|
|
|
|
|
|
test "Attestations with disjoint comittee bits and equal data into single on-chain aggregate" & preset():
|
|
|
|
let
|
|
|
|
bc0 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
|
|
|
|
bc1 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 1.CommitteeIndex, cache)
|
|
|
|
|
|
|
|
# atestation from committee 1
|
|
|
|
attestation_1 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
|
|
|
|
|
|
|
# atestation from different committee with same data as
|
|
|
|
# attestaton 1
|
|
|
|
attestation_2 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc1[1], cache)
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
attestation_1, @[bc0[0]], attestation_1.loadSig,
|
|
|
|
attestation_1.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
attestation_2, @[bc0[1]], attestation_2.loadSig,
|
|
|
|
attestation_2.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
check:
|
|
|
|
process_slots(
|
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
|
|
|
info, {}).isOk()
|
|
|
|
|
|
|
|
let attestations = pool[].getElectraAttestationsForBlock(state[], cache)
|
|
|
|
|
|
|
|
check:
|
|
|
|
# A single inal chain aggregated attestation should be created
|
|
|
|
# with same data and joint committee,aggregation bits
|
|
|
|
attestations.len == 1
|
|
|
|
attestations[0].aggregation_bits.countOnes() == 2
|
|
|
|
attestations[0].committee_bits.countOnes() == 2
|
|
|
|
|
|
|
|
test "Aggregated attestations with disjoint comittee bits into a single on-chain aggregate" & preset():
|
2024-10-14 14:20:26 +00:00
|
|
|
proc verifyAttestationSignature(attestation: electra.Attestation): bool =
|
|
|
|
withState(state[]):
|
|
|
|
when consensusFork == ConsensusFork.Electra:
|
|
|
|
let
|
|
|
|
fork = pool.dag.cfg.forkAtEpoch(forkyState.data.slot.epoch)
|
|
|
|
attesting_indices = get_attesting_indices(
|
|
|
|
forkyState.data, attestation.data, attestation.aggregation_bits,
|
|
|
|
attestation.committee_bits, cache)
|
|
|
|
verify_attestation_signature(
|
|
|
|
fork, pool.dag.genesis_validators_root, attestation.data,
|
|
|
|
attesting_indices.mapIt(forkyState.data.validators.item(it).pubkey),
|
|
|
|
attestation.signature)
|
|
|
|
else:
|
|
|
|
raiseAssert "must be electra"
|
2024-08-23 11:26:35 +00:00
|
|
|
|
|
|
|
let
|
|
|
|
bc0 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
|
|
|
|
bc1 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 1.CommitteeIndex, cache)
|
|
|
|
|
2024-10-04 05:15:48 +00:00
|
|
|
# attestation from first committee
|
2024-08-23 11:26:35 +00:00
|
|
|
attestation_1 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
|
|
|
|
|
|
|
# another attestation from first committee with same data
|
|
|
|
attestation_2 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[1], cache)
|
|
|
|
|
2024-10-04 05:15:48 +00:00
|
|
|
# attestation from different committee with same data as
|
|
|
|
# attestation 1
|
2024-08-23 11:26:35 +00:00
|
|
|
attestation_3 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc1[1], cache)
|
|
|
|
|
2024-10-14 14:20:26 +00:00
|
|
|
check:
|
|
|
|
verifyAttestationSignature(attestation_1)
|
|
|
|
verifyAttestationSignature(attestation_2)
|
|
|
|
verifyAttestationSignature(attestation_3)
|
|
|
|
|
2024-08-23 11:26:35 +00:00
|
|
|
pool[].addAttestation(
|
|
|
|
attestation_1, @[bc0[0]], attestation_1.loadSig,
|
|
|
|
attestation_1.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
attestation_2, @[bc0[1]], attestation_2.loadSig,
|
|
|
|
attestation_2.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
attestation_3, @[bc1[1]], attestation_3.loadSig,
|
|
|
|
attestation_3.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
check:
|
|
|
|
process_slots(
|
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
|
|
|
info, {}).isOk()
|
|
|
|
|
|
|
|
let attestations = pool[].getElectraAttestationsForBlock(state[], cache)
|
|
|
|
|
|
|
|
check:
|
2024-10-14 14:20:26 +00:00
|
|
|
verifyAttestationSignature(attestations[0])
|
2024-10-25 14:04:35 +00:00
|
|
|
check_attestation(
|
|
|
|
state[].electraData.data, attestations[0], {}, cache, true).isOk
|
2024-10-14 14:20:26 +00:00
|
|
|
|
2024-08-23 11:26:35 +00:00
|
|
|
# A single final chain aggregated attestation should be created
|
|
|
|
# with same data, 2 committee bits and 3 aggregation bits
|
|
|
|
attestations.len == 1
|
|
|
|
attestations[0].aggregation_bits.countOnes() == 3
|
2024-09-06 12:15:30 +00:00
|
|
|
attestations[0].committee_bits.countOnes() == 2
|
|
|
|
|
|
|
|
test "Working with electra aggregates" & preset():
|
|
|
|
let
|
|
|
|
# Create an attestation for slot 1!
|
|
|
|
bc0 = get_beacon_committee(
|
|
|
|
state[], getStateField(state[], slot), 0.CommitteeIndex, cache)
|
|
|
|
|
|
|
|
var
|
|
|
|
att0 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[0], cache)
|
|
|
|
att0x = att0
|
|
|
|
att1 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[1], cache)
|
|
|
|
att2 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[2], cache)
|
|
|
|
att3 = makeElectraAttestation(
|
|
|
|
state[], state[].latest_block_root, bc0[3], cache)
|
|
|
|
|
2024-10-14 14:20:26 +00:00
|
|
|
proc verifyAttestationSignature(attestation: electra.Attestation): bool =
|
|
|
|
withState(state[]):
|
|
|
|
when consensusFork == ConsensusFork.Electra:
|
|
|
|
let
|
|
|
|
fork = pool.dag.cfg.forkAtEpoch(forkyState.data.slot.epoch)
|
|
|
|
attesting_indices = get_attesting_indices(
|
|
|
|
forkyState.data, attestation.data, attestation.aggregation_bits,
|
|
|
|
attestation.committee_bits, cache)
|
|
|
|
verify_attestation_signature(
|
|
|
|
fork, pool.dag.genesis_validators_root, attestation.data,
|
|
|
|
attesting_indices.mapIt(forkyState.data.validators.item(it).pubkey),
|
|
|
|
attestation.signature)
|
|
|
|
else:
|
|
|
|
raiseAssert "must be electra"
|
|
|
|
|
|
|
|
check:
|
|
|
|
verifyAttestationSignature(att0)
|
|
|
|
verifyAttestationSignature(att0x)
|
|
|
|
verifyAttestationSignature(att1)
|
|
|
|
verifyAttestationSignature(att2)
|
|
|
|
verifyAttestationSignature(att3)
|
|
|
|
|
2024-09-06 12:15:30 +00:00
|
|
|
# Both attestations include member 2 but neither is a subset of the other
|
|
|
|
att0.combine(att2)
|
|
|
|
att1.combine(att2)
|
|
|
|
|
|
|
|
check:
|
2024-10-14 14:20:26 +00:00
|
|
|
verifyAttestationSignature(att0)
|
|
|
|
verifyAttestationSignature(att1)
|
2024-09-06 12:15:30 +00:00
|
|
|
not pool[].covers(att0.data, att0.aggregation_bits)
|
|
|
|
|
|
|
|
pool[].addAttestation(
|
|
|
|
att0, @[bc0[0], bc0[2]], att0.loadSig, att0.data.slot.start_beacon_time)
|
|
|
|
pool[].addAttestation(
|
|
|
|
att1, @[bc0[1], bc0[2]], att1.loadSig, att1.data.slot.start_beacon_time)
|
|
|
|
|
2024-10-14 14:20:26 +00:00
|
|
|
for att in pool[].electraAttestations(Opt.none Slot, Opt.none CommitteeIndex):
|
|
|
|
check: verifyAttestationSignature(att)
|
|
|
|
|
2024-09-06 12:15:30 +00:00
|
|
|
check:
|
|
|
|
process_slots(
|
|
|
|
defaultRuntimeConfig, state[],
|
|
|
|
getStateField(state[], slot) + MIN_ATTESTATION_INCLUSION_DELAY, cache,
|
|
|
|
info, {}).isOk()
|
|
|
|
|
2024-10-14 14:20:26 +00:00
|
|
|
for att in pool[].electraAttestations(Opt.none Slot, Opt.none CommitteeIndex):
|
|
|
|
check: verifyAttestationSignature(att)
|
|
|
|
|
2024-09-06 12:15:30 +00:00
|
|
|
check:
|
|
|
|
pool[].getElectraAttestationsForBlock(state[], cache).len() == 1
|
|
|
|
# Can get either aggregate here, random!
|
2024-10-14 14:20:26 +00:00
|
|
|
verifyAttestationSignature(pool[].getElectraAggregatedAttestation(
|
|
|
|
1.Slot, hash_tree_root(att0.data), 0.CommitteeIndex).get)
|
2024-09-06 12:15:30 +00:00
|
|
|
|
|
|
|
# Add in attestation 3 - both aggregates should now have it added
|
|
|
|
pool[].addAttestation(
|
|
|
|
att3, @[bc0[3]], att3.loadSig, att3.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
block:
|
|
|
|
let attestations = pool[].getElectraAttestationsForBlock(state[], cache)
|
|
|
|
check:
|
|
|
|
attestations.len() == 1
|
2024-10-15 04:02:36 +00:00
|
|
|
attestations[0].aggregation_bits.countOnes() == 3
|
2024-10-25 14:04:35 +00:00
|
|
|
check_attestation(
|
|
|
|
state[].electraData.data, attestations[0], {}, cache, true).isOk
|
2024-10-15 04:02:36 +00:00
|
|
|
verifyAttestationSignature(attestations[0])
|
2024-09-06 12:15:30 +00:00
|
|
|
# Can get either aggregate here, random!
|
2024-10-15 04:02:36 +00:00
|
|
|
verifyAttestationSignature(pool[].getElectraAggregatedAttestation(
|
|
|
|
1.Slot, hash_tree_root(attestations[0].data), 0.CommitteeIndex).get)
|
2024-09-06 12:15:30 +00:00
|
|
|
|
|
|
|
# Add in attestation 0 as single - attestation 1 is now a superset of the
|
|
|
|
# aggregates in the pool, so everything else should be removed
|
|
|
|
pool[].addAttestation(
|
|
|
|
att0x, @[bc0[0]], att0x.loadSig, att0x.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
block:
|
|
|
|
let attestations = pool[].getElectraAttestationsForBlock(state[], cache)
|
|
|
|
check:
|
|
|
|
attestations.len() == 1
|
|
|
|
attestations[0].aggregation_bits.countOnes() == 4
|
2024-10-25 14:04:35 +00:00
|
|
|
check_attestation(
|
|
|
|
state[].electraData.data, attestations[0], {}, cache, true).isOk
|
2024-10-14 14:20:26 +00:00
|
|
|
verifyAttestationSignature(attestations[0])
|
2024-10-15 04:02:36 +00:00
|
|
|
verifyAttestationSignature(pool[].getElectraAggregatedAttestation(
|
|
|
|
1.Slot, hash_tree_root(attestations[0].data), 0.CommitteeIndex).get)
|
2024-10-04 05:15:48 +00:00
|
|
|
|
|
|
|
# Someone votes for a different root
|
2024-10-15 04:02:36 +00:00
|
|
|
let att4 = makeElectraAttestation(state[], ZERO_HASH, bc0[4], cache)
|
|
|
|
check: verifyAttestationSignature(att4)
|
2024-10-04 05:15:48 +00:00
|
|
|
pool[].addAttestation(
|
|
|
|
att4, @[bc0[4]], att4.loadSig, att4.data.slot.start_beacon_time)
|
|
|
|
|
|
|
|
# Total aggregations size should be one for that root
|
|
|
|
check:
|
|
|
|
pool[].getElectraAggregatedAttestation(1.Slot, hash_tree_root(att4.data),
|
|
|
|
0.CommitteeIndex).get().aggregation_bits.countOnes() == 1
|