426 lines
16 KiB
Nim
426 lines
16 KiB
Nim
# beacon_chain
|
|
# Copyright (c) 2018-2021 Status Research & Development GmbH
|
|
# Licensed and distributed under either of
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
import
|
|
stew/io2, presto, metrics, metrics/chronos_httpserver,
|
|
libp2p/crypto/crypto,
|
|
./rpc/rest_key_management_api,
|
|
./validator_client/[
|
|
common, fallback_service, duties_service, fork_service,
|
|
doppelganger_service, attestation_service, sync_committee_service]
|
|
|
|
proc initGenesis(vc: ValidatorClientRef): Future[RestGenesis] {.async.} =
|
|
info "Initializing genesis", nodes_count = len(vc.beaconNodes)
|
|
var nodes = vc.beaconNodes
|
|
while true:
|
|
var pendingRequests: seq[Future[RestResponse[GetGenesisResponse]]]
|
|
for node in nodes:
|
|
debug "Requesting genesis information", endpoint = node
|
|
pendingRequests.add(node.client.getGenesis())
|
|
|
|
try:
|
|
await allFutures(pendingRequests)
|
|
except CancelledError as exc:
|
|
var pending: seq[Future[void]]
|
|
debug "Genesis information request was interrupted"
|
|
for future in pendingRequests:
|
|
if not(future.finished()):
|
|
pending.add(future.cancelAndWait())
|
|
await allFutures(pending)
|
|
raise exc
|
|
|
|
let (errorNodes, genesisList) =
|
|
block:
|
|
var gres: seq[RestGenesis]
|
|
var bres: seq[BeaconNodeServerRef]
|
|
for i in 0 ..< len(pendingRequests):
|
|
let fut = pendingRequests[i]
|
|
if fut.done():
|
|
let resp = fut.read()
|
|
if resp.status == 200:
|
|
debug "Received genesis information", endpoint = nodes[i],
|
|
genesis_time = resp.data.data.genesis_time,
|
|
genesis_fork_version = resp.data.data.genesis_fork_version,
|
|
genesis_root = resp.data.data.genesis_validators_root
|
|
gres.add(resp.data.data)
|
|
else:
|
|
debug "Received unsuccessful response code", endpoint = nodes[i],
|
|
response_code = resp.status
|
|
bres.add(nodes[i])
|
|
elif fut.failed():
|
|
let error = fut.readError()
|
|
debug "Could not obtain genesis information from beacon node",
|
|
endpoint = nodes[i], error_name = error.name,
|
|
error_msg = error.msg
|
|
bres.add(nodes[i])
|
|
else:
|
|
debug "Interrupted while requesting information from beacon node",
|
|
endpoint = nodes[i]
|
|
bres.add(nodes[i])
|
|
(bres, gres)
|
|
|
|
if len(genesisList) == 0:
|
|
let sleepDuration = 2.seconds
|
|
info "Could not obtain network genesis information from nodes, repeating",
|
|
sleep_time = sleepDuration
|
|
await sleepAsync(sleepDuration)
|
|
nodes = errorNodes
|
|
else:
|
|
# Boyer-Moore majority vote algorithm
|
|
var melem: RestGenesis
|
|
var counter = 0
|
|
for item in genesisList:
|
|
if counter == 0:
|
|
melem = item
|
|
inc(counter)
|
|
else:
|
|
if melem == item:
|
|
inc(counter)
|
|
else:
|
|
dec(counter)
|
|
return melem
|
|
|
|
proc initValidators(vc: ValidatorClientRef): Future[bool] {.async.} =
|
|
info "Loading validators", validatorsDir = vc.config.validatorsDir()
|
|
var duplicates: seq[ValidatorPubKey]
|
|
for keystore in listLoadableKeystores(vc.config, vc.keystoreCache):
|
|
vc.addValidator(keystore)
|
|
return true
|
|
|
|
proc initClock(vc: ValidatorClientRef): Future[BeaconClock] {.async.} =
|
|
# This procedure performs initialization of BeaconClock using current genesis
|
|
# information. It also performs waiting for genesis.
|
|
let res = BeaconClock.init(vc.beaconGenesis.genesis_time)
|
|
let currentSlot = res.now().slotOrZero()
|
|
let currentEpoch = currentSlot.epoch()
|
|
info "Initializing beacon clock",
|
|
genesis_time = vc.beaconGenesis.genesis_time,
|
|
current_slot = currentSlot, current_epoch = currentEpoch
|
|
let genesisTime = res.fromNow(start_beacon_time(Slot(0)))
|
|
if genesisTime.inFuture:
|
|
notice "Waiting for genesis", genesisIn = genesisTime.offset
|
|
await sleepAsync(genesisTime.offset)
|
|
return res
|
|
|
|
proc initMetrics(vc: ValidatorClientRef): Future[bool] {.async.} =
|
|
if vc.config.metricsEnabled:
|
|
let
|
|
metricsAddress = vc.config.metricsAddress
|
|
metricsPort = vc.config.metricsPort
|
|
url = "http://" & $metricsAddress & ":" & $metricsPort & "/metrics"
|
|
info "Starting metrics HTTP server", url = url
|
|
let server =
|
|
block:
|
|
let res = MetricsHttpServerRef.new($metricsAddress, metricsPort)
|
|
if res.isErr():
|
|
error "Could not start metrics HTTP server", url = url,
|
|
error_msg = res.error()
|
|
return false
|
|
res.get()
|
|
vc.metricsServer = some(server)
|
|
try:
|
|
await server.start()
|
|
except MetricsError as exc:
|
|
error "Could not start metrics HTTP server", url = url,
|
|
error_msg = exc.msg, error_name = exc.name
|
|
return false
|
|
return true
|
|
|
|
proc shutdownMetrics(vc: ValidatorClientRef) {.async.} =
|
|
if vc.config.metricsEnabled:
|
|
if vc.metricsServer.isSome():
|
|
info "Shutting down metrics HTTP server"
|
|
await vc.metricsServer.get().close()
|
|
|
|
proc shutdownSlashingProtection(vc: ValidatorClientRef) =
|
|
info "Closing slashing protection", path = vc.config.validatorsDir()
|
|
vc.attachedValidators[].slashingProtection.close()
|
|
|
|
proc onSlotStart(vc: ValidatorClientRef, wallTime: BeaconTime,
|
|
lastSlot: Slot): Future[bool] {.async.} =
|
|
## Called at the beginning of a slot - usually every slot, but sometimes might
|
|
## skip a few in case we're running late.
|
|
## wallTime: current system time - we will strive to perform all duties up
|
|
## to this point in time
|
|
## lastSlot: the last slot that we successfully processed, so we know where to
|
|
## start work from - there might be jumps if processing is delayed
|
|
|
|
let
|
|
# The slot we should be at, according to the clock
|
|
beaconTime = wallTime
|
|
wallSlot = wallTime.toSlot()
|
|
|
|
let
|
|
# If everything was working perfectly, the slot that we should be processing
|
|
expectedSlot = lastSlot + 1
|
|
delay = wallTime - expectedSlot.start_beacon_time()
|
|
|
|
if checkIfShouldStopAtEpoch(wallSlot.slot, vc.config.stopAtEpoch):
|
|
return true
|
|
|
|
if len(vc.beaconNodes) > 1:
|
|
let
|
|
counts = vc.getNodeCounts()
|
|
# Good nodes are nodes which can be used for ALL the requests.
|
|
goodNodes = counts.data[int(RestBeaconNodeStatus.Synced)]
|
|
# Viable nodes are nodes which can be used only SOME of the requests.
|
|
viableNodes = counts.data[int(RestBeaconNodeStatus.OptSynced)] +
|
|
counts.data[int(RestBeaconNodeStatus.NotSynced)] +
|
|
counts.data[int(RestBeaconNodeStatus.Compatible)]
|
|
# Bad nodes are nodes which can't be used at all.
|
|
badNodes = counts.data[int(RestBeaconNodeStatus.Offline)] +
|
|
counts.data[int(RestBeaconNodeStatus.Online)] +
|
|
counts.data[int(RestBeaconNodeStatus.Incompatible)]
|
|
info "Slot start",
|
|
slot = shortLog(wallSlot.slot),
|
|
attestationIn = vc.getDurationToNextAttestation(wallSlot.slot),
|
|
blockIn = vc.getDurationToNextBlock(wallSlot.slot),
|
|
validators = vc.attachedValidators[].count(),
|
|
good_nodes = goodNodes, viable_nodes = viableNodes, bad_nodes = badNodes,
|
|
delay = shortLog(delay)
|
|
else:
|
|
info "Slot start",
|
|
slot = shortLog(wallSlot.slot),
|
|
attestationIn = vc.getDurationToNextAttestation(wallSlot.slot),
|
|
blockIn = vc.getDurationToNextBlock(wallSlot.slot),
|
|
validators = vc.attachedValidators[].count(),
|
|
node_status = $vc.beaconNodes[0].status,
|
|
delay = shortLog(delay)
|
|
|
|
return false
|
|
|
|
proc new*(T: type ValidatorClientRef,
|
|
config: ValidatorClientConf,
|
|
rng: ref HmacDrbgContext): ValidatorClientRef =
|
|
let beaconNodes =
|
|
block:
|
|
var servers: seq[BeaconNodeServerRef]
|
|
for index, url in config.beaconNodes.pairs():
|
|
let res = BeaconNodeServerRef.init(url, index)
|
|
if res.isErr():
|
|
warn "Unable to initialize remote beacon node",
|
|
url = $url, error = res.error()
|
|
else:
|
|
debug "Beacon node was initialized", node = res.get()
|
|
servers.add(res.get())
|
|
let missingRoles = getMissingRoles(servers)
|
|
if len(missingRoles) != 0:
|
|
fatal "Beacon nodes do not use all required roles",
|
|
missing_roles = $missingRoles, nodes_count = len(servers)
|
|
quit 1
|
|
servers
|
|
|
|
if len(beaconNodes) == 0:
|
|
# This should not happen, thanks to defaults in `conf.nim`
|
|
fatal "Not enough beacon nodes in command line"
|
|
quit 1
|
|
|
|
when declared(waitSignal):
|
|
ValidatorClientRef(
|
|
rng: rng,
|
|
config: config,
|
|
beaconNodes: beaconNodes,
|
|
graffitiBytes: config.graffiti.get(defaultGraffitiBytes()),
|
|
nodesAvailable: newAsyncEvent(),
|
|
forksAvailable: newAsyncEvent(),
|
|
doppelExit: newAsyncEvent(),
|
|
indicesAvailable: newAsyncEvent(),
|
|
dynamicFeeRecipientsStore: newClone(DynamicFeeRecipientsStore.init()),
|
|
sigintHandleFut: waitSignal(SIGINT),
|
|
sigtermHandleFut: waitSignal(SIGTERM),
|
|
keystoreCache: KeystoreCacheRef.init()
|
|
)
|
|
else:
|
|
ValidatorClientRef(
|
|
rng: rng,
|
|
config: config,
|
|
beaconNodes: beaconNodes,
|
|
graffitiBytes: config.graffiti.get(defaultGraffitiBytes()),
|
|
nodesAvailable: newAsyncEvent(),
|
|
forksAvailable: newAsyncEvent(),
|
|
indicesAvailable: newAsyncEvent(),
|
|
doppelExit: newAsyncEvent(),
|
|
dynamicFeeRecipientsStore: newClone(DynamicFeeRecipientsStore.init()),
|
|
sigintHandleFut: newFuture[void]("sigint_placeholder"),
|
|
sigtermHandleFut: newFuture[void]("sigterm_placeholder"),
|
|
keystoreCache: KeystoreCacheRef.init()
|
|
)
|
|
|
|
proc asyncInit(vc: ValidatorClientRef): Future[ValidatorClientRef] {.async.} =
|
|
notice "Launching validator client", version = fullVersionStr,
|
|
cmdParams = commandLineParams(),
|
|
config = vc.config,
|
|
beacon_nodes_count = len(vc.beaconNodes)
|
|
|
|
vc.beaconGenesis = await vc.initGenesis()
|
|
info "Genesis information", genesis_time = vc.beaconGenesis.genesis_time,
|
|
genesis_fork_version = vc.beaconGenesis.genesis_fork_version,
|
|
genesis_root = vc.beaconGenesis.genesis_validators_root
|
|
|
|
vc.beaconClock = await vc.initClock()
|
|
|
|
if not(await initMetrics(vc)):
|
|
raise newException(ValidatorClientError,
|
|
"Could not initialize metrics server")
|
|
|
|
info "Initializing slashing protection", path = vc.config.validatorsDir()
|
|
|
|
let
|
|
slashingProtectionDB =
|
|
SlashingProtectionDB.init(
|
|
vc.beaconGenesis.genesis_validators_root,
|
|
vc.config.validatorsDir(), "slashing_protection")
|
|
validatorPool = newClone(ValidatorPool.init(
|
|
slashingProtectionDB, vc.config.doppelgangerDetection))
|
|
|
|
vc.attachedValidators = validatorPool
|
|
|
|
if not(await initValidators(vc)):
|
|
await vc.shutdownMetrics()
|
|
raise newException(ValidatorClientError,
|
|
"Could not initialize local validators")
|
|
|
|
let
|
|
keymanagerInitResult = initKeymanagerServer(vc.config, nil)
|
|
|
|
try:
|
|
vc.fallbackService = await FallbackServiceRef.init(vc)
|
|
vc.forkService = await ForkServiceRef.init(vc)
|
|
vc.dutiesService = await DutiesServiceRef.init(vc)
|
|
vc.doppelgangerService = await DoppelgangerServiceRef.init(vc)
|
|
vc.attestationService = await AttestationServiceRef.init(vc)
|
|
vc.syncCommitteeService = await SyncCommitteeServiceRef.init(vc)
|
|
vc.keymanagerServer = keymanagerInitResult.server
|
|
if vc.keymanagerServer != nil:
|
|
vc.keymanagerHost = newClone KeymanagerHost.init(
|
|
validatorPool,
|
|
vc.rng,
|
|
keymanagerInitResult.token,
|
|
vc.config.validatorsDir,
|
|
vc.config.secretsDir,
|
|
vc.config.defaultFeeRecipient,
|
|
vc.config.suggestedGasLimit,
|
|
nil,
|
|
vc.beaconClock.getBeaconTimeFn)
|
|
|
|
except CatchableError as exc:
|
|
warn "Unexpected error encountered while initializing",
|
|
error_name = exc.name, error_msg = exc.msg
|
|
await vc.shutdownMetrics()
|
|
vc.shutdownSlashingProtection()
|
|
except CancelledError:
|
|
debug "Initialization process interrupted"
|
|
await vc.shutdownMetrics()
|
|
vc.shutdownSlashingProtection()
|
|
return
|
|
|
|
return vc
|
|
|
|
proc asyncRun*(vc: ValidatorClientRef) {.async.} =
|
|
vc.fallbackService.start()
|
|
vc.forkService.start()
|
|
vc.dutiesService.start()
|
|
vc.doppelgangerService.start()
|
|
vc.attestationService.start()
|
|
vc.syncCommitteeService.start()
|
|
|
|
if not isNil(vc.keymanagerServer):
|
|
doAssert vc.keymanagerHost != nil
|
|
vc.keymanagerServer.router.installKeymanagerHandlers(vc.keymanagerHost[])
|
|
vc.keymanagerServer.start()
|
|
|
|
var doppelEventFut = vc.doppelExit.wait()
|
|
try:
|
|
vc.runSlotLoopFut = runSlotLoop(vc, vc.beaconClock.now(), onSlotStart)
|
|
vc.runKeystoreCachePruningLoopFut =
|
|
runKeystorecachePruningLoop(vc.keystoreCache)
|
|
discard await race(vc.runSlotLoopFut, doppelEventFut)
|
|
if not(vc.runSlotLoopFut.finished()):
|
|
notice "Received shutdown event, exiting"
|
|
except CancelledError:
|
|
debug "Main loop interrupted"
|
|
except CatchableError as exc:
|
|
debug "Main loop failed with an error", err_name = $exc.name,
|
|
err_msg = $exc.msg
|
|
|
|
await vc.shutdownMetrics()
|
|
vc.shutdownSlashingProtection()
|
|
|
|
if doppelEventFut.finished:
|
|
# Critically, database has been shut down - the rest doesn't matter, we need
|
|
# to stop as soon as possible
|
|
# TODO we need to actually quit _before_ any other async tasks have had the
|
|
# chance to happen
|
|
quitDoppelganger()
|
|
|
|
debug "Stopping main processing loop"
|
|
var pending: seq[Future[void]]
|
|
if not(vc.runSlotLoopFut.finished()):
|
|
pending.add(vc.runSlotLoopFut.cancelAndWait())
|
|
if not(vc.runKeystoreCachePruningLoopFut.finished()):
|
|
pending.add(vc.runKeystoreCachePruningLoopFut.cancelAndWait())
|
|
if not(doppelEventFut.finished()):
|
|
pending.add(doppelEventFut.cancelAndWait())
|
|
debug "Stopping running services"
|
|
pending.add(vc.fallbackService.stop())
|
|
pending.add(vc.forkService.stop())
|
|
pending.add(vc.dutiesService.stop())
|
|
pending.add(vc.doppelgangerService.stop())
|
|
pending.add(vc.attestationService.stop())
|
|
pending.add(vc.syncCommitteeService.stop())
|
|
if not isNil(vc.keymanagerServer):
|
|
pending.add(vc.keymanagerServer.stop())
|
|
|
|
await allFutures(pending)
|
|
|
|
template runWithSignals(vc: ValidatorClientRef, body: untyped): bool =
|
|
let future = body
|
|
discard await race(future, vc.sigintHandleFut, vc.sigtermHandleFut)
|
|
if future.finished():
|
|
if future.failed() or future.cancelled():
|
|
let exc = future.readError()
|
|
debug "Validator client initialization failed", err_name = $exc.name,
|
|
err_msg = $exc.msg
|
|
var pending: seq[Future[void]]
|
|
if not(vc.sigintHandleFut.finished()):
|
|
pending.add(cancelAndWait(vc.sigintHandleFut))
|
|
if not(vc.sigtermHandleFut.finished()):
|
|
pending.add(cancelAndWait(vc.sigtermHandleFut))
|
|
await allFutures(pending)
|
|
false
|
|
else:
|
|
true
|
|
else:
|
|
let signal = if vc.sigintHandleFut.finished(): "SIGINT" else: "SIGTERM"
|
|
info "Got interrupt, trying to shutdown gracefully", signal = signal
|
|
var pending = @[cancelAndWait(future)]
|
|
if not(vc.sigintHandleFut.finished()):
|
|
pending.add(cancelAndWait(vc.sigintHandleFut))
|
|
if not(vc.sigtermHandleFut.finished()):
|
|
pending.add(cancelAndWait(vc.sigtermHandleFut))
|
|
await allFutures(pending)
|
|
false
|
|
|
|
proc runValidatorClient*(config: ValidatorClientConf,
|
|
rng: ref HmacDrbgContext) {.async.} =
|
|
let vc = ValidatorClientRef.new(config, rng)
|
|
if not vc.runWithSignals(asyncInit vc):
|
|
return
|
|
if not vc.runWithSignals(asyncRun vc):
|
|
return
|
|
|
|
programMain:
|
|
let
|
|
config = makeBannerAndConfig("Nimbus validator client " & fullVersionStr,
|
|
ValidatorClientConf)
|
|
|
|
# Single RNG instance for the application - will be seeded on construction
|
|
# and avoid using system resources (such as urandom) after that
|
|
rng = crypto.newRng()
|
|
|
|
setupLogging(config.logLevel, config.logStdout, config.logFile)
|
|
waitFor runValidatorClient(config, rng)
|