Implement beacon sync stub
- Prepare a test env for beacon sync in engine api simulator. - Wiring beacon sync to the rest of subsystems.
This commit is contained in:
parent
6139152143
commit
5fb0fc65ba
|
@ -51,7 +51,7 @@ template testLatestHeader(client: untyped, expectedHash: Web3Hash) =
|
||||||
expect = expectedHash,
|
expect = expectedHash,
|
||||||
get = lastHash
|
get = lastHash
|
||||||
|
|
||||||
#proc sendTx(t: TestEnv, recipient: EthAddress, val: UInt256, data: openArray[byte] = []): bool =
|
#proc sendTx(env: TestEnv, recipient: EthAddress, val: UInt256, data: openArray[byte] = []): bool =
|
||||||
# t.tx = t.makeTx(recipient, val, data)
|
# t.tx = t.makeTx(recipient, val, data)
|
||||||
# let rr = env.client.sendTransaction(t.tx)
|
# let rr = env.client.sendTransaction(t.tx)
|
||||||
# if rr.isErr:
|
# if rr.isErr:
|
||||||
|
@ -59,7 +59,7 @@ template testLatestHeader(client: untyped, expectedHash: Web3Hash) =
|
||||||
# return false
|
# return false
|
||||||
# return true
|
# return true
|
||||||
#
|
#
|
||||||
#proc sendTx(t: TestEnv, val: UInt256): bool =
|
#proc sendTx(env: TestEnv, val: UInt256): bool =
|
||||||
# t.sendTx(prevRandaoContractAddr, val)
|
# t.sendTx(prevRandaoContractAddr, val)
|
||||||
|
|
||||||
# Invalid Terminal Block in ForkchoiceUpdated:
|
# Invalid Terminal Block in ForkchoiceUpdated:
|
||||||
|
@ -88,7 +88,7 @@ proc invalidTerminalBlockForkchoiceUpdated*(env: TestEnv): bool =
|
||||||
|
|
||||||
#[
|
#[
|
||||||
# Invalid GetPayload Under PoW: Client must reject GetPayload directives under PoW.
|
# Invalid GetPayload Under PoW: Client must reject GetPayload directives under PoW.
|
||||||
proc invalidGetPayloadUnderPoW(t: TestEnv): TestStatus =
|
proc invalidGetPayloadUnderPoW(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# We start in PoW and try to get an invalid Payload, which should produce an error but nothing should be disrupted.
|
# We start in PoW and try to get an invalid Payload, which should produce an error but nothing should be disrupted.
|
||||||
|
@ -101,7 +101,7 @@ proc invalidGetPayloadUnderPoW(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
# Invalid Terminal Block in NewPayload:
|
# Invalid Terminal Block in NewPayload:
|
||||||
# Client must reject NewPayload directives if the referenced ParentHash does not meet the TTD requirement.
|
# Client must reject NewPayload directives if the referenced ParentHash does not meet the TTD requirement.
|
||||||
proc invalidTerminalBlockNewPayload(t: TestEnv): TestStatus =
|
proc invalidTerminalBlockNewPayload(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let gBlock = t.gHeader
|
let gBlock = t.gHeader
|
||||||
|
@ -126,16 +126,16 @@ proc invalidTerminalBlockNewPayload(t: TestEnv): TestStatus =
|
||||||
# Check that PoW chain progresses
|
# Check that PoW chain progresses
|
||||||
testCond t.verifyPoWProgress(t.gHeader.blockHash)
|
testCond t.verifyPoWProgress(t.gHeader.blockHash)
|
||||||
|
|
||||||
proc unknownHeadBlockHash(t: TestEnv): TestStatus =
|
proc unknownHeadBlockHash(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
var randomHash: common.Hash256
|
var randomHash: common.Hash256
|
||||||
testCond randomBytes(randomHash.data) == 32
|
testCond randomBytes(randomHash.data) == 32
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let forkchoiceStateUnknownHeadHash = ForkchoiceStateV1(
|
let forkchoiceStateUnknownHeadHash = ForkchoiceStateV1(
|
||||||
headBlockHash: BlockHash randomHash.data,
|
headBlockHash: BlockHash randomHash.data,
|
||||||
safeBlockHash: clMock.latestForkchoice.finalizedBlockHash,
|
safeBlockHash: clMock.latestForkchoice.finalizedBlockHash,
|
||||||
|
@ -163,17 +163,17 @@ proc unknownHeadBlockHash(t: TestEnv): TestStatus =
|
||||||
testCond s.payloadStatus.status == PayloadExecutionStatus.syncing
|
testCond s.payloadStatus.status == PayloadExecutionStatus.syncing
|
||||||
testCond s.payloadId.isNone
|
testCond s.payloadId.isNone
|
||||||
|
|
||||||
proc unknownSafeBlockHash(t: TestEnv): TestStatus =
|
proc unknownSafeBlockHash(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
let produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
let produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
||||||
# Run test after a new payload has been broadcast
|
# Run test after a new payload has been broadcast
|
||||||
|
@ -196,17 +196,17 @@ proc unknownSafeBlockHash(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
testCond produceSingleBlockRes
|
testCond produceSingleBlockRes
|
||||||
|
|
||||||
proc unknownFinalizedBlockHash(t: TestEnv): TestStatus =
|
proc unknownFinalizedBlockHash(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
let produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
let produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
||||||
# Run test after a new payload has been broadcast
|
# Run test after a new payload has been broadcast
|
||||||
|
@ -250,15 +250,15 @@ type
|
||||||
alternativePayloads: seq[ExecutableData]
|
alternativePayloads: seq[ExecutableData]
|
||||||
|
|
||||||
template inconsistentForkchoiceStateGen(procname: untyped, inconsistency: Inconsistency) =
|
template inconsistentForkchoiceStateGen(procname: untyped, inconsistency: Inconsistency) =
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
var pList = PayloadList()
|
var pList = PayloadList()
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
|
@ -320,14 +320,14 @@ inconsistentForkchoiceStateGen(inconsistentForkchoiceState3, Inconsistency.Final
|
||||||
|
|
||||||
# Verify behavior on a forkchoiceUpdated with invalid payload attributes
|
# Verify behavior on a forkchoiceUpdated with invalid payload attributes
|
||||||
template invalidPayloadAttributesGen(procname: untyped, syncingCond: bool) =
|
template invalidPayloadAttributesGen(procname: untyped, syncingCond: bool) =
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
|
@ -378,14 +378,14 @@ template invalidPayloadAttributesGen(procname: untyped, syncingCond: bool) =
|
||||||
invalidPayloadAttributesGen(invalidPayloadAttributes1, false)
|
invalidPayloadAttributesGen(invalidPayloadAttributes1, false)
|
||||||
invalidPayloadAttributesGen(invalidPayloadAttributes2, true)
|
invalidPayloadAttributesGen(invalidPayloadAttributes2, true)
|
||||||
|
|
||||||
proc preTTDFinalizedBlockHash(t: TestEnv): TestStatus =
|
proc preTTDFinalizedBlockHash(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let
|
let
|
||||||
|
@ -396,7 +396,7 @@ proc preTTDFinalizedBlockHash(t: TestEnv): TestStatus =
|
||||||
finalizedBlockHash: gHash,
|
finalizedBlockHash: gHash,
|
||||||
)
|
)
|
||||||
client = env.client
|
client = env.client
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
|
|
||||||
var res = client.forkchoiceUpdatedV1(forkchoiceState)
|
var res = client.forkchoiceUpdatedV1(forkchoiceState)
|
||||||
testFCU(res, invalid, some(common.Hash256()))
|
testFCU(res, invalid, some(common.Hash256()))
|
||||||
|
@ -435,17 +435,17 @@ type
|
||||||
hash: common.Hash256
|
hash: common.Hash256
|
||||||
|
|
||||||
template badHashOnNewPayloadGen(procname: untyped, syncingCond: bool, sideChain: bool) =
|
template badHashOnNewPayloadGen(procname: untyped, syncingCond: bool, sideChain: bool) =
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
let shadow = Shadow()
|
let shadow = Shadow()
|
||||||
|
|
||||||
|
@ -530,18 +530,18 @@ badHashOnNewPayloadGen(badHashOnNewPayload2, true, false)
|
||||||
badHashOnNewPayloadGen(badHashOnNewPayload3, false, true)
|
badHashOnNewPayloadGen(badHashOnNewPayload3, false, true)
|
||||||
badHashOnNewPayloadGen(badHashOnNewPayload4, true, true)
|
badHashOnNewPayloadGen(badHashOnNewPayload4, true, true)
|
||||||
|
|
||||||
proc parentHashOnExecPayload(t: TestEnv): TestStatus =
|
proc parentHashOnExecPayload(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
var produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
var produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
||||||
# Run test after the new payload has been obtained
|
# Run test after the new payload has been obtained
|
||||||
|
@ -560,14 +560,14 @@ proc parentHashOnExecPayload(t: TestEnv): TestStatus =
|
||||||
testCond produceSingleBlockRes
|
testCond produceSingleBlockRes
|
||||||
|
|
||||||
# Attempt to re-org to a chain containing an invalid transition payload
|
# Attempt to re-org to a chain containing an invalid transition payload
|
||||||
proc invalidTransitionPayload(t: TestEnv): TestStatus =
|
proc invalidTransitionPayload(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by main client
|
# Wait until TTD is reached by main client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
# Produce two blocks before trying to re-org
|
# Produce two blocks before trying to re-org
|
||||||
|
@ -603,14 +603,14 @@ proc invalidTransitionPayload(t: TestEnv): TestStatus =
|
||||||
testCond pbRes
|
testCond pbRes
|
||||||
|
|
||||||
template invalidPayloadTestCaseGen(procName: untyped, payloadField: InvalidPayloadField, emptyTxs: bool = false) =
|
template invalidPayloadTestCaseGen(procName: untyped, payloadField: InvalidPayloadField, emptyTxs: bool = false) =
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
template txProc(): bool =
|
template txProc(): bool =
|
||||||
|
@ -763,19 +763,19 @@ invalidPayloadTestCaseGen(invalidPayload15, InvalidTransactionValue)
|
||||||
|
|
||||||
# Test to verify Block information available at the Eth RPC after NewPayload
|
# Test to verify Block information available at the Eth RPC after NewPayload
|
||||||
template blockStatusExecPayloadGen(procname: untyped, transitionBlock: bool) =
|
template blockStatusExecPayloadGen(procname: untyped, transitionBlock: bool) =
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test, only if we are not testing the transition block
|
# Produce blocks before starting the test, only if we are not testing the transition block
|
||||||
when not transitionBlock:
|
when not transitionBlock:
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
let shadow = Shadow()
|
let shadow = Shadow()
|
||||||
|
|
||||||
|
@ -828,14 +828,14 @@ type
|
||||||
template invalidMissingAncestorReOrgGen(procName: untyped,
|
template invalidMissingAncestorReOrgGen(procName: untyped,
|
||||||
invalid_index: int, payloadField: InvalidPayloadField, p2psync: bool, emptyTxs: bool) =
|
invalid_index: int, payloadField: InvalidPayloadField, p2psync: bool, emptyTxs: bool) =
|
||||||
|
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
|
@ -942,19 +942,19 @@ invalidMissingAncestorReOrgGen(invalidMissingAncestor2, 9, InvalidStateRoot, fal
|
||||||
invalidMissingAncestorReOrgGen(invalidMissingAncestor3, 10, InvalidStateRoot, false, true)
|
invalidMissingAncestorReOrgGen(invalidMissingAncestor3, 10, InvalidStateRoot, false, true)
|
||||||
|
|
||||||
template blockStatusHeadBlockGen(procname: untyped, transitionBlock: bool) =
|
template blockStatusHeadBlockGen(procname: untyped, transitionBlock: bool) =
|
||||||
proc procName(t: TestEnv): TestStatus =
|
proc procName(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test, only if we are not testing the transition block
|
# Produce blocks before starting the test, only if we are not testing the transition block
|
||||||
when not transitionBlock:
|
when not transitionBlock:
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
let shadow = Shadow()
|
let shadow = Shadow()
|
||||||
|
|
||||||
|
@ -981,10 +981,10 @@ template blockStatusHeadBlockGen(procname: untyped, transitionBlock: bool) =
|
||||||
blockStatusHeadBlockGen(blockStatusHeadBlock1, false)
|
blockStatusHeadBlockGen(blockStatusHeadBlock1, false)
|
||||||
blockStatusHeadBlockGen(blockStatusHeadBlock2, true)
|
blockStatusHeadBlockGen(blockStatusHeadBlock2, true)
|
||||||
|
|
||||||
proc blockStatusSafeBlock(t: TestEnv): TestStatus =
|
proc blockStatusSafeBlock(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
# On PoW mode, `safe` tag shall return error.
|
# On PoW mode, `safe` tag shall return error.
|
||||||
|
@ -993,7 +993,7 @@ proc blockStatusSafeBlock(t: TestEnv): TestStatus =
|
||||||
testCond rr.isErr
|
testCond rr.isErr
|
||||||
|
|
||||||
# Wait until this client catches up with latest PoS Block
|
# Wait until this client catches up with latest PoS Block
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# First ForkchoiceUpdated sent was equal to 0x00..00, `safe` should return error now
|
# First ForkchoiceUpdated sent was equal to 0x00..00, `safe` should return error now
|
||||||
|
@ -1012,10 +1012,10 @@ proc blockStatusSafeBlock(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
testCond pbres
|
testCond pbres
|
||||||
|
|
||||||
proc blockStatusFinalizedBlock(t: TestEnv): TestStatus =
|
proc blockStatusFinalizedBlock(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
# On PoW mode, `finalized` tag shall return error.
|
# On PoW mode, `finalized` tag shall return error.
|
||||||
|
@ -1024,7 +1024,7 @@ proc blockStatusFinalizedBlock(t: TestEnv): TestStatus =
|
||||||
testCond rr.isErr
|
testCond rr.isErr
|
||||||
|
|
||||||
# Wait until this client catches up with latest PoS Block
|
# Wait until this client catches up with latest PoS Block
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# First ForkchoiceUpdated sent was equal to 0x00..00, `finalized` should return error now
|
# First ForkchoiceUpdated sent was equal to 0x00..00, `finalized` should return error now
|
||||||
|
@ -1043,18 +1043,18 @@ proc blockStatusFinalizedBlock(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
testCond pbres
|
testCond pbres
|
||||||
|
|
||||||
proc blockStatusReorg(t: TestEnv): TestStatus =
|
proc blockStatusReorg(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
var produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
var produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
||||||
# Run test after a forkchoice with new HeadBlockHash has been broadcasted
|
# Run test after a forkchoice with new HeadBlockHash has been broadcasted
|
||||||
|
@ -1130,18 +1130,18 @@ proc blockStatusReorg(t: TestEnv): TestStatus =
|
||||||
))
|
))
|
||||||
testCond produceSingleBlockRes
|
testCond produceSingleBlockRes
|
||||||
|
|
||||||
proc reExecPayloads(t: TestEnv): TestStatus =
|
proc reExecPayloads(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until this client catches up with latest PoS
|
# Wait until this client catches up with latest PoS
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# How many Payloads we are going to re-execute
|
# How many Payloads we are going to re-execute
|
||||||
var payloadReExecCount = 10
|
var payloadReExecCount = 10
|
||||||
|
|
||||||
# Create those blocks
|
# Create those blocks
|
||||||
let produceBlockRes = t.clMock.produceBlocks(payloadReExecCount, BlockProcessCallbacks())
|
let produceBlockRes = env.clMock.produceBlocks(payloadReExecCount, BlockProcessCallbacks())
|
||||||
testCond produceBlockRes
|
testCond produceBlockRes
|
||||||
|
|
||||||
# Re-execute the payloads
|
# Re-execute the payloads
|
||||||
|
@ -1154,7 +1154,7 @@ proc reExecPayloads(t: TestEnv): TestStatus =
|
||||||
info "Started re-executing payloads at block", number=lastBlock
|
info "Started re-executing payloads at block", number=lastBlock
|
||||||
|
|
||||||
let
|
let
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
start = lastBlock - payloadReExecCount + 1
|
start = lastBlock - payloadReExecCount + 1
|
||||||
|
|
||||||
for i in start..lastBlock:
|
for i in start..lastBlock:
|
||||||
|
@ -1171,18 +1171,18 @@ proc reExecPayloads(t: TestEnv): TestStatus =
|
||||||
testCond true:
|
testCond true:
|
||||||
error "(test issue) Payload does not exist", index=i
|
error "(test issue) Payload does not exist", index=i
|
||||||
|
|
||||||
proc multipleNewCanonicalPayloads(t: TestEnv): TestStatus =
|
proc multipleNewCanonicalPayloads(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
let produce5BlockRes = t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
let produce5BlockRes = env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
testCond produce5BlockRes
|
testCond produce5BlockRes
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
var produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
var produceSingleBlockRes = clMock.produceSingleBlock(BlockProcessCallbacks(
|
||||||
# Run test after a new payload has been obtained
|
# Run test after a new payload has been obtained
|
||||||
|
@ -1213,11 +1213,11 @@ proc multipleNewCanonicalPayloads(t: TestEnv): TestStatus =
|
||||||
# At the end the clMocker continues to try to execute fcU with the original payload, which should not fail
|
# At the end the clMocker continues to try to execute fcU with the original payload, which should not fail
|
||||||
testCond produceSingleBlockRes
|
testCond produceSingleBlockRes
|
||||||
|
|
||||||
proc outOfOrderPayloads(t: TestEnv): TestStatus =
|
proc outOfOrderPayloads(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# First prepare payloads on a first client, which will also contain multiple transactions
|
# First prepare payloads on a first client, which will also contain multiple transactions
|
||||||
|
@ -1232,7 +1232,7 @@ proc outOfOrderPayloads(t: TestEnv): TestStatus =
|
||||||
var recipient: EthAddress
|
var recipient: EthAddress
|
||||||
doAssert randomBytes(recipient) == 20
|
doAssert randomBytes(recipient) == 20
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
var produceBlockRes = clMock.produceBlocks(payloadCount, BlockProcessCallbacks(
|
var produceBlockRes = clMock.produceBlocks(payloadCount, BlockProcessCallbacks(
|
||||||
# We send the transactions after we got the Payload ID, before the clMocker gets the prepared Payload
|
# We send the transactions after we got the Payload ID, before the clMocker gets the prepared Payload
|
||||||
|
@ -1257,14 +1257,14 @@ proc outOfOrderPayloads(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
# Test that performing a re-org back into a previous block of the canonical chain does not produce errors and the chain
|
# Test that performing a re-org back into a previous block of the canonical chain does not produce errors and the chain
|
||||||
# is still capable of progressing.
|
# is still capable of progressing.
|
||||||
proc reorgBack(t: TestEnv): TestStatus =
|
proc reorgBack(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
testCond clMock.produceSingleBlock(BlockProcessCallbacks())
|
testCond clMock.produceSingleBlock(BlockProcessCallbacks())
|
||||||
|
@ -1298,16 +1298,16 @@ type
|
||||||
SideChainList = ref object
|
SideChainList = ref object
|
||||||
sidechainPayloads: seq[ExecutionPayloadV1]
|
sidechainPayloads: seq[ExecutionPayloadV1]
|
||||||
|
|
||||||
proc reorgBackFromSyncing(t: TestEnv): TestStatus =
|
proc reorgBackFromSyncing(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce an alternative chain
|
# Produce an alternative chain
|
||||||
let pList = SideChainList()
|
let pList = SideChainList()
|
||||||
let clMock = t.clMock
|
let clMock = env.clMock
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
let r1 = clMock.produceBlocks(10, BlockProcessCallbacks(
|
let r1 = clMock.produceBlocks(10, BlockProcessCallbacks(
|
||||||
|
@ -1371,15 +1371,15 @@ type
|
||||||
noTxnPayload: ExecutionPayloadV1
|
noTxnPayload: ExecutionPayloadV1
|
||||||
txHash: common.Hash256
|
txHash: common.Hash256
|
||||||
|
|
||||||
proc transactionReorg(t: TestEnv): TestStatus =
|
proc transactionReorg(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
testCond t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
testCond env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
|
|
||||||
# Create transactions that modify the state in order to testCond after the reorg.
|
# Create transactions that modify the state in order to testCond after the reorg.
|
||||||
const
|
const
|
||||||
|
@ -1388,7 +1388,7 @@ proc transactionReorg(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
let
|
let
|
||||||
client = env.client
|
client = env.client
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
shadow = TxReorgShadow()
|
shadow = TxReorgShadow()
|
||||||
|
|
||||||
for i in 0..<txCount:
|
for i in 0..<txCount:
|
||||||
|
@ -1468,7 +1468,7 @@ proc transactionReorg(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
testCond pbres
|
testCond pbres
|
||||||
|
|
||||||
proc testCondPrevRandaoValue(t: TestEnv, expectedPrevRandao: common.Hash256, blockNumber: uint64): bool =
|
proc testCondPrevRandaoValue(env: TestEnv, expectedPrevRandao: common.Hash256, blockNumber: uint64): bool =
|
||||||
let storageKey = blockNumber.u256
|
let storageKey = blockNumber.u256
|
||||||
let client = env.client
|
let client = env.client
|
||||||
|
|
||||||
|
@ -1485,19 +1485,19 @@ proc testCondPrevRandaoValue(t: TestEnv, expectedPrevRandao: common.Hash256, blo
|
||||||
return false
|
return false
|
||||||
true
|
true
|
||||||
|
|
||||||
proc sidechainReorg(t: TestEnv): TestStatus =
|
proc sidechainReorg(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Produce blocks before starting the test
|
# Produce blocks before starting the test
|
||||||
testCond t.clMock.produceBlocks(5, BlockProcessCallbacks())
|
testCond env.clMock.produceBlocks(5, BlockProcessCallbacks())
|
||||||
|
|
||||||
let
|
let
|
||||||
client = env.client
|
client = env.client
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
|
|
||||||
# Produce two payloads, send fcU with first payload, testCond transaction outcome, then reorg, testCond transaction outcome again
|
# Produce two payloads, send fcU with first payload, testCond transaction outcome, then reorg, testCond transaction outcome again
|
||||||
|
|
||||||
|
@ -1575,11 +1575,11 @@ proc sidechainReorg(t: TestEnv): TestStatus =
|
||||||
let latestBlockNum = clMock.latestHeadNumber.uint64
|
let latestBlockNum = clMock.latestHeadNumber.uint64
|
||||||
testCond testCondPrevRandaoValue(t, clMock.prevRandaoHistory[latestBlockNum], latestBlockNum)
|
testCond testCondPrevRandaoValue(t, clMock.prevRandaoHistory[latestBlockNum], latestBlockNum)
|
||||||
|
|
||||||
proc suggestedFeeRecipient(t: TestEnv): TestStatus =
|
proc suggestedFeeRecipient(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
let ok = waitFor t.clMock.waitForTTD()
|
let ok = waitFor env.clMock.waitForTTD()
|
||||||
testCond ok
|
testCond ok
|
||||||
|
|
||||||
# Amount of transactions to send
|
# Amount of transactions to send
|
||||||
|
@ -1592,7 +1592,7 @@ proc suggestedFeeRecipient(t: TestEnv): TestStatus =
|
||||||
|
|
||||||
let
|
let
|
||||||
client = env.client
|
client = env.client
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
|
|
||||||
# Send multiple transactions
|
# Send multiple transactions
|
||||||
for i in 0..<txCount:
|
for i in 0..<txCount:
|
||||||
|
@ -1649,21 +1649,21 @@ proc suggestedFeeRecipient(t: TestEnv): TestStatus =
|
||||||
error "balance does not match fees",
|
error "balance does not match fees",
|
||||||
feeRecipientBalance, feeRecipientFees
|
feeRecipientBalance, feeRecipientFees
|
||||||
|
|
||||||
proc sendTxAsync(t: TestEnv): Future[void] {.async.} =
|
proc sendTxAsync(env: TestEnv): Future[void] {.async.} =
|
||||||
let
|
let
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
period = chronos.milliseconds(500)
|
period = chronos.milliseconds(500)
|
||||||
|
|
||||||
while not clMock.ttdReached:
|
while not clMock.ttdReached:
|
||||||
await sleepAsync(period)
|
await sleepAsync(period)
|
||||||
discard t.sendTx(0.u256)
|
discard t.sendTx(0.u256)
|
||||||
|
|
||||||
proc prevRandaoOpcodeTx(t: TestEnv): TestStatus =
|
proc prevRandaoOpcodeTx(env: TestEnv): TestStatus =
|
||||||
result = TestStatus.OK
|
result = TestStatus.OK
|
||||||
|
|
||||||
let
|
let
|
||||||
client = env.client
|
client = env.client
|
||||||
clMock = t.clMock
|
clMock = env.clMock
|
||||||
sendTxFuture = sendTxAsync(t)
|
sendTxFuture = sendTxAsync(t)
|
||||||
|
|
||||||
# Wait until TTD is reached by this client
|
# Wait until TTD is reached by this client
|
||||||
|
|
|
@ -2,6 +2,7 @@ import
|
||||||
std/[os, math],
|
std/[os, math],
|
||||||
eth/keys,
|
eth/keys,
|
||||||
eth/p2p as eth_p2p,
|
eth/p2p as eth_p2p,
|
||||||
|
chronos,
|
||||||
json_rpc/[rpcserver, rpcclient],
|
json_rpc/[rpcserver, rpcclient],
|
||||||
stew/[results, byteutils],
|
stew/[results, byteutils],
|
||||||
../../../nimbus/[
|
../../../nimbus/[
|
||||||
|
@ -14,6 +15,8 @@ import
|
||||||
core/block_import,
|
core/block_import,
|
||||||
rpc,
|
rpc,
|
||||||
sync/protocol,
|
sync/protocol,
|
||||||
|
sync/beacon,
|
||||||
|
sync/handlers,
|
||||||
beacon/beacon_engine,
|
beacon/beacon_engine,
|
||||||
common
|
common
|
||||||
],
|
],
|
||||||
|
@ -39,12 +42,14 @@ type
|
||||||
EngineEnv* = ref object
|
EngineEnv* = ref object
|
||||||
conf : NimbusConf
|
conf : NimbusConf
|
||||||
com : CommonRef
|
com : CommonRef
|
||||||
|
node : EthereumNode
|
||||||
server : RpcHttpServer
|
server : RpcHttpServer
|
||||||
sealer : SealingEngineRef
|
sealer : SealingEngineRef
|
||||||
ttd : DifficultyInt
|
ttd : DifficultyInt
|
||||||
tx : Transaction
|
tx : Transaction
|
||||||
nonce : uint64
|
nonce : uint64
|
||||||
client : RpcHttpClient
|
client : RpcHttpClient
|
||||||
|
sync : BeaconSyncRef
|
||||||
|
|
||||||
const
|
const
|
||||||
baseFolder = "hive_integration/nodocker/engine"
|
baseFolder = "hive_integration/nodocker/engine"
|
||||||
|
@ -66,7 +71,11 @@ proc makeCom*(conf: NimbusConf): CommonRef =
|
||||||
)
|
)
|
||||||
|
|
||||||
proc envConfig*(): NimbusConf =
|
proc envConfig*(): NimbusConf =
|
||||||
makeConfig(@["--engine-signer:658bdf435d810c91414ec09147daa6db62406379", "--custom-network:" & genesisFile])
|
makeConfig(@[
|
||||||
|
"--engine-signer:658bdf435d810c91414ec09147daa6db62406379",
|
||||||
|
"--custom-network:" & genesisFile,
|
||||||
|
"--listen-address: 127.0.0.1",
|
||||||
|
])
|
||||||
|
|
||||||
proc envConfig*(conf: ChainConfig): NimbusConf =
|
proc envConfig*(conf: ChainConfig): NimbusConf =
|
||||||
result = envConfig()
|
result = envConfig()
|
||||||
|
@ -82,14 +91,19 @@ proc newEngineEnv*(conf: var NimbusConf, chainFile: string, enableAuth: bool): E
|
||||||
echo error
|
echo error
|
||||||
quit(QuitFailure)
|
quit(QuitFailure)
|
||||||
|
|
||||||
let
|
var
|
||||||
node = setupEthNode(conf, ctx, eth)
|
node = setupEthNode(conf, ctx)
|
||||||
com = makeCom(conf)
|
com = makeCom(conf)
|
||||||
chain = newChain(com)
|
chain = newChain(com)
|
||||||
|
|
||||||
com.initializeEmptyDb()
|
com.initializeEmptyDb()
|
||||||
let txPool = TxPoolRef.new(com, conf.engineSigner)
|
let txPool = TxPoolRef.new(com, conf.engineSigner)
|
||||||
|
|
||||||
|
node.addEthHandlerCapability(
|
||||||
|
node.peerPool,
|
||||||
|
chain,
|
||||||
|
txPool)
|
||||||
|
|
||||||
# txPool must be informed of active head
|
# txPool must be informed of active head
|
||||||
# so it can know the latest account state
|
# so it can know the latest account state
|
||||||
let head = com.db.getCanonicalHead()
|
let head = com.db.getCanonicalHead()
|
||||||
|
@ -101,15 +115,15 @@ proc newEngineEnv*(conf: var NimbusConf, chainFile: string, enableAuth: bool): E
|
||||||
quit(QuitFailure)
|
quit(QuitFailure)
|
||||||
|
|
||||||
let
|
let
|
||||||
hooks = if enableAuth: @[httpJwtAuth(key)]
|
hooks = if enableAuth: @[httpJwtAuth(key)]
|
||||||
else: @[]
|
else: @[]
|
||||||
server = newRpcHttpServer(["127.0.0.1:" & $conf.rpcPort], hooks)
|
server = newRpcHttpServer(["127.0.0.1:" & $conf.rpcPort], hooks)
|
||||||
sealer = SealingEngineRef.new(
|
sealer = SealingEngineRef.new(
|
||||||
chain, ctx, conf.engineSigner,
|
chain, ctx, conf.engineSigner,
|
||||||
txPool, EngineStopped)
|
txPool, EngineStopped)
|
||||||
|
sync = BeaconSyncRef.init(node, chain, ctx.rng, conf.maxPeers, id=conf.tcpPort.int)
|
||||||
beaconEngine = BeaconEngineRef.new(txPool, chain)
|
beaconEngine = BeaconEngineRef.new(txPool, chain)
|
||||||
|
|
||||||
|
|
||||||
setupEthRpc(node, ctx, com, txPool, server)
|
setupEthRpc(node, ctx, com, txPool, server)
|
||||||
setupEngineAPI(beaconEngine, server)
|
setupEngineAPI(beaconEngine, server)
|
||||||
setupDebugRpc(com, server)
|
setupDebugRpc(com, server)
|
||||||
|
@ -126,15 +140,22 @@ proc newEngineEnv*(conf: var NimbusConf, chainFile: string, enableAuth: bool): E
|
||||||
let client = newRpcHttpClient()
|
let client = newRpcHttpClient()
|
||||||
waitFor client.connect("127.0.0.1", conf.rpcPort, false)
|
waitFor client.connect("127.0.0.1", conf.rpcPort, false)
|
||||||
|
|
||||||
|
sync.start()
|
||||||
|
node.startListening()
|
||||||
|
|
||||||
EngineEnv(
|
EngineEnv(
|
||||||
conf : conf,
|
conf : conf,
|
||||||
com : com,
|
com : com,
|
||||||
|
node : node,
|
||||||
server : server,
|
server : server,
|
||||||
sealer : sealer,
|
sealer : sealer,
|
||||||
client : client
|
client : client,
|
||||||
|
sync : sync
|
||||||
)
|
)
|
||||||
|
|
||||||
proc close*(env: EngineEnv) =
|
proc close*(env: EngineEnv) =
|
||||||
|
env.node.stopListening()
|
||||||
|
env.sync.stop()
|
||||||
waitFor env.client.close()
|
waitFor env.client.close()
|
||||||
waitFor env.sealer.stop()
|
waitFor env.sealer.stop()
|
||||||
waitFor env.server.closeWait()
|
waitFor env.server.closeWait()
|
||||||
|
@ -157,6 +178,12 @@ func ttd*(env: EngineEnv): UInt256 =
|
||||||
func com*(env: EngineEnv): CommonRef =
|
func com*(env: EngineEnv): CommonRef =
|
||||||
env.com
|
env.com
|
||||||
|
|
||||||
|
func node*(env: EngineEnv): ENode =
|
||||||
|
env.node.listeningAddress
|
||||||
|
|
||||||
|
proc connect*(env: EngineEnv, node: ENode) =
|
||||||
|
waitFor env.node.connectToNode(node)
|
||||||
|
|
||||||
func gwei(n: int64): GasInt {.compileTime.} =
|
func gwei(n: int64): GasInt {.compileTime.} =
|
||||||
GasInt(n * (10 ^ 9))
|
GasInt(n * (10 ^ 9))
|
||||||
|
|
||||||
|
|
|
@ -78,10 +78,6 @@ proc close*(env: TestEnv) =
|
||||||
for eng in env.clients:
|
for eng in env.clients:
|
||||||
eng.close()
|
eng.close()
|
||||||
|
|
||||||
proc addEngine*(env: TestEnv): EngineEnv =
|
|
||||||
var conf = envConfig(env.conf.networkParams.config)
|
|
||||||
env.addEngine(conf)
|
|
||||||
|
|
||||||
func client*(env: TestEnv): RpcHttpClient =
|
func client*(env: TestEnv): RpcHttpClient =
|
||||||
env.clients.first.client
|
env.clients.first.client
|
||||||
|
|
||||||
|
@ -91,6 +87,12 @@ func engine*(env: TestEnv): EngineEnv =
|
||||||
proc setupCLMock*(env: TestEnv) =
|
proc setupCLMock*(env: TestEnv) =
|
||||||
env.clmock = newCLMocker(env.clients, env.engine.com)
|
env.clmock = newCLMocker(env.clients, env.engine.com)
|
||||||
|
|
||||||
|
proc addEngine*(env: TestEnv): EngineEnv =
|
||||||
|
var conf = env.conf # clone the conf
|
||||||
|
let eng = env.addEngine(conf)
|
||||||
|
eng.connect(env.engine.node)
|
||||||
|
eng
|
||||||
|
|
||||||
proc makeTx*(env: TestEnv, eng: EngineEnv, tc: BaseTx, nonce: AccountNonce): Transaction =
|
proc makeTx*(env: TestEnv, eng: EngineEnv, tc: BaseTx, nonce: AccountNonce): Transaction =
|
||||||
eng.makeTx(env.vaultKey, tc, nonce)
|
eng.makeTx(env.vaultKey, tc, nonce)
|
||||||
|
|
||||||
|
|
|
@ -247,7 +247,7 @@ let wdTestList* = [
|
||||||
wdAbleAccountCount: MAINNET_MAX_WITHDRAWAL_COUNT_PER_BLOCK,
|
wdAbleAccountCount: MAINNET_MAX_WITHDRAWAL_COUNT_PER_BLOCK,
|
||||||
syncSteps: 1,
|
syncSteps: 1,
|
||||||
)),
|
)),
|
||||||
#[TestDesc(
|
TestDesc(
|
||||||
name: "Sync after 128 blocks - Withdrawals on Block 2 - Multiple Withdrawal Accounts",
|
name: "Sync after 128 blocks - Withdrawals on Block 2 - Multiple Withdrawal Accounts",
|
||||||
about: "- Spawn a first client\n" &
|
about: "- Spawn a first client\n" &
|
||||||
"- Go through withdrawals fork on Block 2\n" &
|
"- Go through withdrawals fork on Block 2\n" &
|
||||||
|
@ -262,7 +262,7 @@ let wdTestList* = [
|
||||||
wdPerBlock: MAINNET_MAX_WITHDRAWAL_COUNT_PER_BLOCK,
|
wdPerBlock: MAINNET_MAX_WITHDRAWAL_COUNT_PER_BLOCK,
|
||||||
wdAbleAccountCount: 1024,
|
wdAbleAccountCount: 1024,
|
||||||
syncSteps: 1,
|
syncSteps: 1,
|
||||||
)),]#
|
)),
|
||||||
|
|
||||||
# EVM Tests (EIP-3651, EIP-3855, EIP-3860)
|
# EVM Tests (EIP-3651, EIP-3855, EIP-3860)
|
||||||
TestDesc(
|
TestDesc(
|
||||||
|
|
|
@ -27,7 +27,7 @@ import
|
||||||
./core/[chain, sealer, clique/clique_desc,
|
./core/[chain, sealer, clique/clique_desc,
|
||||||
clique/clique_sealer, tx_pool, block_import],
|
clique/clique_sealer, tx_pool, block_import],
|
||||||
./beacon/beacon_engine,
|
./beacon/beacon_engine,
|
||||||
./sync/[legacy, full, protocol, snap, stateless,
|
./sync/[beacon, legacy, full, protocol, snap, stateless,
|
||||||
protocol/les_protocol, handlers, peers],
|
protocol/les_protocol, handlers, peers],
|
||||||
./evm/async/data_sources/json_rpc_data_source
|
./evm/async/data_sources/json_rpc_data_source
|
||||||
|
|
||||||
|
@ -60,6 +60,7 @@ type
|
||||||
legaSyncRef: LegacySyncRef
|
legaSyncRef: LegacySyncRef
|
||||||
snapSyncRef: SnapSyncRef
|
snapSyncRef: SnapSyncRef
|
||||||
fullSyncRef: FullSyncRef
|
fullSyncRef: FullSyncRef
|
||||||
|
beaconSyncRef: BeaconSyncRef
|
||||||
statelessSyncRef: StatelessSyncRef
|
statelessSyncRef: StatelessSyncRef
|
||||||
beaconEngine: BeaconEngineRef
|
beaconEngine: BeaconEngineRef
|
||||||
|
|
||||||
|
@ -105,7 +106,7 @@ proc manageAccounts(nimbus: NimbusNode, conf: NimbusConf) =
|
||||||
quit(QuitFailure)
|
quit(QuitFailure)
|
||||||
|
|
||||||
proc setupP2P(nimbus: NimbusNode, conf: NimbusConf,
|
proc setupP2P(nimbus: NimbusNode, conf: NimbusConf,
|
||||||
protocols: set[ProtocolFlag]) =
|
com: CommonRef, protocols: set[ProtocolFlag]) =
|
||||||
## Creating P2P Server
|
## Creating P2P Server
|
||||||
let kpres = nimbus.ctx.getNetKeys(conf.netKey, conf.dataDir.string)
|
let kpres = nimbus.ctx.getNetKeys(conf.netKey, conf.dataDir.string)
|
||||||
if kpres.isErr:
|
if kpres.isErr:
|
||||||
|
@ -190,8 +191,13 @@ proc setupP2P(nimbus: NimbusNode, conf: NimbusConf,
|
||||||
# FIXME-Adam: what needs to go here?
|
# FIXME-Adam: what needs to go here?
|
||||||
nimbus.statelessSyncRef = StatelessSyncRef.init()
|
nimbus.statelessSyncRef = StatelessSyncRef.init()
|
||||||
of SyncMode.Default:
|
of SyncMode.Default:
|
||||||
nimbus.legaSyncRef = LegacySyncRef.new(
|
if com.forkGTE(MergeFork):
|
||||||
nimbus.ethNode, nimbus.chainRef)
|
nimbus.beaconSyncRef = BeaconSyncRef.init(
|
||||||
|
nimbus.ethNode, nimbus.chainRef, nimbus.ctx.rng, conf.maxPeers,
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
nimbus.legaSyncRef = LegacySyncRef.new(
|
||||||
|
nimbus.ethNode, nimbus.chainRef)
|
||||||
|
|
||||||
# Connect directly to the static nodes
|
# Connect directly to the static nodes
|
||||||
let staticPeers = conf.getStaticPeers()
|
let staticPeers = conf.getStaticPeers()
|
||||||
|
@ -426,17 +432,20 @@ proc start(nimbus: NimbusNode, conf: NimbusConf) =
|
||||||
else:
|
else:
|
||||||
basicServices(nimbus, conf, com)
|
basicServices(nimbus, conf, com)
|
||||||
manageAccounts(nimbus, conf)
|
manageAccounts(nimbus, conf)
|
||||||
setupP2P(nimbus, conf, protocols)
|
setupP2P(nimbus, conf, com, protocols)
|
||||||
localServices(nimbus, conf, com, protocols)
|
localServices(nimbus, conf, com, protocols)
|
||||||
|
|
||||||
if conf.maxPeers > 0:
|
if conf.maxPeers > 0:
|
||||||
case conf.syncMode:
|
case conf.syncMode:
|
||||||
of SyncMode.Default:
|
of SyncMode.Default:
|
||||||
nimbus.legaSyncRef.start
|
if com.forkGTE(MergeFork):
|
||||||
nimbus.ethNode.setEthHandlerNewBlocksAndHashes(
|
nimbus.beaconSyncRef.start
|
||||||
legacy.newBlockHandler,
|
else:
|
||||||
legacy.newBlockHashesHandler,
|
nimbus.legaSyncRef.start
|
||||||
cast[pointer](nimbus.legaSyncRef))
|
nimbus.ethNode.setEthHandlerNewBlocksAndHashes(
|
||||||
|
legacy.newBlockHandler,
|
||||||
|
legacy.newBlockHashesHandler,
|
||||||
|
cast[pointer](nimbus.legaSyncRef))
|
||||||
of SyncMode.Full:
|
of SyncMode.Full:
|
||||||
nimbus.fullSyncRef.start
|
nimbus.fullSyncRef.start
|
||||||
of SyncMode.Stateless:
|
of SyncMode.Stateless:
|
||||||
|
@ -474,6 +483,8 @@ proc stop*(nimbus: NimbusNode, conf: NimbusConf) {.async, gcsafe.} =
|
||||||
nimbus.snapSyncRef.stop()
|
nimbus.snapSyncRef.stop()
|
||||||
if nimbus.fullSyncRef.isNil.not:
|
if nimbus.fullSyncRef.isNil.not:
|
||||||
nimbus.fullSyncRef.stop()
|
nimbus.fullSyncRef.stop()
|
||||||
|
if nimbus.beaconSyncRef.isNil.not:
|
||||||
|
nimbus.beaconSyncRef.stop()
|
||||||
|
|
||||||
proc process*(nimbus: NimbusNode, conf: NimbusConf) =
|
proc process*(nimbus: NimbusNode, conf: NimbusConf) =
|
||||||
# Main event loop
|
# Main event loop
|
||||||
|
|
|
@ -0,0 +1,150 @@
|
||||||
|
# Nimbus
|
||||||
|
# Copyright (c) 2023 Status Research & Development GmbH
|
||||||
|
# Licensed under either of
|
||||||
|
# * Apache License, version 2.0, ([LICENSE-APACHE](LICENSE-APACHE) or
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0)
|
||||||
|
# * MIT license ([LICENSE-MIT](LICENSE-MIT) or
|
||||||
|
# http://opensource.org/licenses/MIT)
|
||||||
|
# at your option. This file may not be copied, modified, or distributed
|
||||||
|
# except according to those terms.
|
||||||
|
|
||||||
|
{.push raises: [].}
|
||||||
|
|
||||||
|
import
|
||||||
|
eth/p2p,
|
||||||
|
chronicles,
|
||||||
|
chronos,
|
||||||
|
stew/[interval_set, sorted_set],
|
||||||
|
./beacon/[worker, worker_desc],
|
||||||
|
"."/[sync_desc, sync_sched, protocol]
|
||||||
|
|
||||||
|
logScope:
|
||||||
|
topics = "beacon-sync"
|
||||||
|
|
||||||
|
type
|
||||||
|
BeaconSyncRef* = RunnerSyncRef[BeaconCtxData,BeaconBuddyData]
|
||||||
|
|
||||||
|
const
|
||||||
|
extraTraceMessages = false # or true
|
||||||
|
## Enable additional logging noise
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Private logging helpers
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
template traceMsg(f, info: static[string]; args: varargs[untyped]) =
|
||||||
|
trace "Snap scheduler " & f & "() " & info, args
|
||||||
|
|
||||||
|
template traceMsgCtx(f, info: static[string]; c: BeaconCtxRef) =
|
||||||
|
when extraTraceMessages:
|
||||||
|
block:
|
||||||
|
let
|
||||||
|
poolMode {.inject.} = c.poolMode
|
||||||
|
daemon {.inject.} = c.daemon
|
||||||
|
f.traceMsg info, poolMode, daemon
|
||||||
|
|
||||||
|
template traceMsgBuddy(f, info: static[string]; b: BeaconBuddyRef) =
|
||||||
|
when extraTraceMessages:
|
||||||
|
block:
|
||||||
|
let
|
||||||
|
peer {.inject.} = b.peer
|
||||||
|
runState {.inject.} = b.ctrl.state
|
||||||
|
multiOk {.inject.} = b.ctrl.multiOk
|
||||||
|
poolMode {.inject.} = b.ctx.poolMode
|
||||||
|
daemon {.inject.} = b.ctx.daemon
|
||||||
|
f.traceMsg info, peer, runState, multiOk, poolMode, daemon
|
||||||
|
|
||||||
|
|
||||||
|
template tracerFrameCtx(f: static[string]; c: BeaconCtxRef; code: untyped) =
|
||||||
|
f.traceMsgCtx "begin", c
|
||||||
|
code
|
||||||
|
f.traceMsgCtx "end", c
|
||||||
|
|
||||||
|
template tracerFrameBuddy(f: static[string]; b: BeaconBuddyRef; code: untyped) =
|
||||||
|
f.traceMsgBuddy "begin", b
|
||||||
|
code
|
||||||
|
f.traceMsgBuddy "end", b
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Virtual methods/interface, `mixin` functions
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc runSetup(ctx: BeaconCtxRef): bool =
|
||||||
|
tracerFrameCtx("runSetup", ctx):
|
||||||
|
result = worker.setup(ctx)
|
||||||
|
|
||||||
|
proc runRelease(ctx: BeaconCtxRef) =
|
||||||
|
tracerFrameCtx("runRelease", ctx):
|
||||||
|
worker.release(ctx)
|
||||||
|
|
||||||
|
proc runDaemon(ctx: BeaconCtxRef) {.async.} =
|
||||||
|
tracerFrameCtx("runDaemon", ctx):
|
||||||
|
await worker.runDaemon(ctx)
|
||||||
|
|
||||||
|
proc runStart(buddy: BeaconBuddyRef): bool =
|
||||||
|
tracerFrameBuddy("runStart", buddy):
|
||||||
|
result = worker.start(buddy)
|
||||||
|
|
||||||
|
proc runStop(buddy: BeaconBuddyRef) =
|
||||||
|
tracerFrameBuddy("runStop", buddy):
|
||||||
|
worker.stop(buddy)
|
||||||
|
|
||||||
|
proc runPool(buddy: BeaconBuddyRef; last: bool; laps: int): bool =
|
||||||
|
tracerFrameBuddy("runPool", buddy):
|
||||||
|
result = worker.runPool(buddy, last, laps)
|
||||||
|
|
||||||
|
proc runSingle(buddy: BeaconBuddyRef) {.async.} =
|
||||||
|
tracerFrameBuddy("runSingle", buddy):
|
||||||
|
await worker.runSingle(buddy)
|
||||||
|
|
||||||
|
proc runMulti(buddy: BeaconBuddyRef) {.async.} =
|
||||||
|
tracerFrameBuddy("runMulti", buddy):
|
||||||
|
await worker.runMulti(buddy)
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Private helpers
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc updateBeaconHeaderCB(ctx: BeaconSyncRef): SyncReqNewHeadCB =
|
||||||
|
## Update beacon header. This function is intended as a call back function
|
||||||
|
## for the RPC module.
|
||||||
|
result = proc(h: BlockHeader) {.gcsafe, raises: [].} =
|
||||||
|
try:
|
||||||
|
debugEcho "REQUEST SYNC TO: ", h.blockNumber
|
||||||
|
debugEcho "REQUEST SYNC TO: ", h.blockHash
|
||||||
|
except CatchableError as ex:
|
||||||
|
debugEcho ex.msg
|
||||||
|
|
||||||
|
proc enableRpcMagic(ctx: BeaconSyncRef) =
|
||||||
|
## Helper for `setup()`: Enable external pivot update via RPC
|
||||||
|
let com = ctx.ctx.chain.com
|
||||||
|
com.syncReqNewHead = ctx.updateBeaconHeaderCB
|
||||||
|
com.syncReqRelaxV2 = true
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Public functions
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc init*(
|
||||||
|
T: type BeaconSyncRef;
|
||||||
|
ethNode: EthereumNode;
|
||||||
|
chain: ChainRef;
|
||||||
|
rng: ref HmacDrbgContext;
|
||||||
|
maxPeers: int;
|
||||||
|
id: int = 0): T =
|
||||||
|
new result
|
||||||
|
result.initSync(ethNode, chain, maxPeers, none(string))
|
||||||
|
result.ctx.pool.rng = rng
|
||||||
|
result.ctx.pool.id = id
|
||||||
|
|
||||||
|
proc start*(ctx: BeaconSyncRef) =
|
||||||
|
## Beacon Sync always begin with stop mode
|
||||||
|
doAssert ctx.startSync() # Initialize subsystems
|
||||||
|
ctx.enableRpcMagic() # Allow external pivot update via RPC
|
||||||
|
|
||||||
|
proc stop*(ctx: BeaconSyncRef) =
|
||||||
|
ctx.stopSync()
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# End
|
||||||
|
# ------------------------------------------------------------------------------
|
|
@ -0,0 +1,171 @@
|
||||||
|
# Nimbus
|
||||||
|
# Copyright (c) 2023 Status Research & Development GmbH
|
||||||
|
# Licensed and distributed under either of
|
||||||
|
# * MIT license (license terms in the root directory or at
|
||||||
|
# https://opensource.org/licenses/MIT).
|
||||||
|
# * Apache v2 license (license terms in the root directory or at
|
||||||
|
# https://www.apache.org/licenses/LICENSE-2.0).
|
||||||
|
# at your option. This file may not be copied, modified, or distributed
|
||||||
|
# except according to those terms.
|
||||||
|
|
||||||
|
{.push raises:[].}
|
||||||
|
|
||||||
|
import
|
||||||
|
chronicles,
|
||||||
|
chronos,
|
||||||
|
eth/p2p,
|
||||||
|
".."/[protocol, sync_desc],
|
||||||
|
./worker_desc
|
||||||
|
|
||||||
|
logScope:
|
||||||
|
topics = "beacon-buddy"
|
||||||
|
|
||||||
|
const
|
||||||
|
extraTraceMessages = false # or true
|
||||||
|
## Enabled additional logging noise
|
||||||
|
|
||||||
|
FirstPivotSeenTimeout = 3.minutes
|
||||||
|
## Turn on relaxed pivot negotiation after some waiting time when there
|
||||||
|
## was a `peer` seen but was rejected. This covers a rare event. Typically
|
||||||
|
## useless peers do not appear ready for negotiation.
|
||||||
|
|
||||||
|
FirstPivotAcceptedTimeout = 50.seconds
|
||||||
|
## Turn on relaxed pivot negotiation after some waiting time when there
|
||||||
|
## was a `peer` accepted but no second one yet.
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Private helpers
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc pp(n: BlockNumber): string =
|
||||||
|
## Dedicated pretty printer (`$` is defined elsewhere using `UInt256`)
|
||||||
|
if n == high(BlockNumber): "high" else:"#" & $n
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Private functions
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Public start/stop and admin functions
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc setup*(ctx: BeaconCtxRef): bool =
|
||||||
|
## Global set up
|
||||||
|
#ctx.pool.pivot = BestPivotCtxRef.init(ctx.pool.rng)
|
||||||
|
|
||||||
|
true
|
||||||
|
|
||||||
|
proc release*(ctx: BeaconCtxRef) =
|
||||||
|
## Global clean up
|
||||||
|
#ctx.pool.pivot = nil
|
||||||
|
|
||||||
|
proc start*(buddy: BeaconBuddyRef): bool =
|
||||||
|
## Initialise worker peer
|
||||||
|
let
|
||||||
|
ctx = buddy.ctx
|
||||||
|
peer = buddy.peer
|
||||||
|
if peer.supports(protocol.eth) and
|
||||||
|
peer.state(protocol.eth).initialized:
|
||||||
|
|
||||||
|
ctx.daemon = true
|
||||||
|
return true
|
||||||
|
|
||||||
|
proc stop*(buddy: BeaconBuddyRef) =
|
||||||
|
## Clean up this peer
|
||||||
|
buddy.ctrl.stopped = true
|
||||||
|
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# Public functions
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc runDaemon*(ctx: BeaconCtxRef) {.async.} =
|
||||||
|
## Global background job that will be re-started as long as the variable
|
||||||
|
## `ctx.daemon` is set `true`. If that job was stopped due to re-setting
|
||||||
|
## `ctx.daemon` to `false`, it will be restarted next after it was reset
|
||||||
|
## as `true` not before there is some activity on the `runPool()`,
|
||||||
|
## `runSingle()`, or `runMulti()` functions.
|
||||||
|
##
|
||||||
|
|
||||||
|
debugEcho "RUNDAEMON: ", ctx.pool.id
|
||||||
|
ctx.daemon = false
|
||||||
|
|
||||||
|
# Without waiting, this function repeats every 50ms (as set with the constant
|
||||||
|
# `sync_sched.execLoopTimeElapsedMin`.) Larger waiting time cleans up logging.
|
||||||
|
await sleepAsync 300.milliseconds
|
||||||
|
|
||||||
|
|
||||||
|
proc runSingle*(buddy: BeaconBuddyRef) {.async.} =
|
||||||
|
## This peer worker is invoked if the peer-local flag `buddy.ctrl.multiOk`
|
||||||
|
## is set `false` which is the default mode. This flag is updated by the
|
||||||
|
## worker when deemed appropriate.
|
||||||
|
## * For all workers, there can be only one `runSingle()` function active
|
||||||
|
## simultaneously for all worker peers.
|
||||||
|
## * There will be no `runMulti()` function active for the same worker peer
|
||||||
|
## simultaneously
|
||||||
|
## * There will be no `runPool()` iterator active simultaneously.
|
||||||
|
##
|
||||||
|
## Note that this function runs in `async` mode.
|
||||||
|
##
|
||||||
|
let
|
||||||
|
ctx = buddy.ctx
|
||||||
|
peer {.used.} = buddy.peer
|
||||||
|
|
||||||
|
debugEcho "RUNSINGLE: ", ctx.pool.id
|
||||||
|
|
||||||
|
if buddy.ctrl.stopped:
|
||||||
|
when extraTraceMessages:
|
||||||
|
trace "Single mode stopped", peer, pivotState=ctx.pool.pivotState
|
||||||
|
return # done with this buddy
|
||||||
|
|
||||||
|
var napping = 2.seconds
|
||||||
|
when extraTraceMessages:
|
||||||
|
trace "Single mode end", peer, napping
|
||||||
|
|
||||||
|
# Without waiting, this function repeats every 50ms (as set with the constant
|
||||||
|
# `sync_sched.execLoopTimeElapsedMin`.)
|
||||||
|
await sleepAsync napping
|
||||||
|
|
||||||
|
|
||||||
|
proc runPool*(buddy: BeaconBuddyRef; last: bool; laps: int): bool =
|
||||||
|
## Once started, the function `runPool()` is called for all worker peers in
|
||||||
|
## sequence as the body of an iteration as long as the function returns
|
||||||
|
## `false`. There will be no other worker peer functions activated
|
||||||
|
## simultaneously.
|
||||||
|
##
|
||||||
|
## This procedure is started if the global flag `buddy.ctx.poolMode` is set
|
||||||
|
## `true` (default is `false`.) It will be automatically reset before the
|
||||||
|
## the loop starts. Re-setting it again results in repeating the loop. The
|
||||||
|
## argument `lap` (starting with `0`) indicated the currend lap of the
|
||||||
|
## repeated loops.
|
||||||
|
##
|
||||||
|
## The argument `last` is set `true` if the last entry is reached.
|
||||||
|
##
|
||||||
|
## Note that this function does not run in `async` mode.
|
||||||
|
##
|
||||||
|
let
|
||||||
|
ctx = buddy.ctx
|
||||||
|
|
||||||
|
debugEcho "RUNPOOL: ", ctx.pool.id
|
||||||
|
|
||||||
|
true # Stop after running once regardless of peer
|
||||||
|
|
||||||
|
proc runMulti*(buddy: BeaconBuddyRef) {.async.} =
|
||||||
|
## This peer worker is invoked if the `buddy.ctrl.multiOk` flag is set
|
||||||
|
## `true` which is typically done after finishing `runSingle()`. This
|
||||||
|
## instance can be simultaneously active for all peer workers.
|
||||||
|
##
|
||||||
|
let
|
||||||
|
ctx = buddy.ctx
|
||||||
|
|
||||||
|
debugEcho "RUNMULTI: ", ctx.pool.id
|
||||||
|
|
||||||
|
# Update persistent database
|
||||||
|
#while not buddy.ctrl.stopped:
|
||||||
|
# Allow thread switch as `persistBlocks()` might be slow
|
||||||
|
await sleepAsync(10.milliseconds)
|
||||||
|
|
||||||
|
# ------------------------------------------------------------------------------
|
||||||
|
# End
|
||||||
|
# ------------------------------------------------------------------------------
|
|
@ -0,0 +1,33 @@
|
||||||
|
# Nimbus
|
||||||
|
# Copyright (c) 2023 Status Research & Development GmbH
|
||||||
|
# Licensed and distributed under either of
|
||||||
|
# * MIT license (license terms in the root directory or at
|
||||||
|
# https://opensource.org/licenses/MIT).
|
||||||
|
# * Apache v2 license (license terms in the root directory or at
|
||||||
|
# https://www.apache.org/licenses/LICENSE-2.0).
|
||||||
|
# at your option. This file may not be copied, modified, or distributed
|
||||||
|
# except according to those terms.
|
||||||
|
|
||||||
|
{.push raises:[].}
|
||||||
|
|
||||||
|
import
|
||||||
|
eth/p2p,
|
||||||
|
chronos,
|
||||||
|
../sync_desc
|
||||||
|
|
||||||
|
type
|
||||||
|
BeaconBuddyData* = object
|
||||||
|
## Local descriptor data extension
|
||||||
|
|
||||||
|
BeaconCtxData* = object
|
||||||
|
## Globally shared data extension
|
||||||
|
rng*: ref HmacDrbgContext ## Random generator, pre-initialised
|
||||||
|
id*: int
|
||||||
|
|
||||||
|
BeaconBuddyRef* = BuddyRef[BeaconCtxData,BeaconBuddyData]
|
||||||
|
## Extended worker peer descriptor
|
||||||
|
|
||||||
|
BeaconCtxRef* = CtxRef[BeaconCtxData]
|
||||||
|
## Extended global descriptor
|
||||||
|
|
||||||
|
# End
|
Loading…
Reference in New Issue