806 lines
27 KiB
Bash
Executable File
806 lines
27 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
|
|
# Copyright (c) 2020-2022 Status Research & Development GmbH. Licensed under
|
|
# either of:
|
|
# - Apache License, version 2.0
|
|
# - MIT license
|
|
# at your option. This file may not be copied, modified, or distributed except
|
|
# according to those terms.
|
|
|
|
# Mostly a duplication of "tests/simulation/{start.sh,run_node.sh}", but with a focus on
|
|
# replicating testnets as closely as possible, which means following the Docker execution labyrinth.
|
|
|
|
set -e
|
|
|
|
cd "$(dirname "${BASH_SOURCE[0]}")"/..
|
|
|
|
# OS detection
|
|
OS="linux"
|
|
if uname | grep -qi darwin; then
|
|
OS="macos"
|
|
elif uname | grep -qiE "mingw|msys"; then
|
|
OS="windows"
|
|
fi
|
|
|
|
# architecture detection
|
|
ARCH="$(uname -m)"
|
|
|
|
####################
|
|
# argument parsing #
|
|
####################
|
|
|
|
GETOPT_BINARY="getopt"
|
|
if [[ "${OS}" == "macos" ]]; then
|
|
GETOPT_BINARY=$(find /opt/homebrew/opt/gnu-getopt/bin/getopt /usr/local/opt/gnu-getopt/bin/getopt 2> /dev/null || true)
|
|
[[ -f "$GETOPT_BINARY" ]] || { echo "GNU getopt not installed. Please run 'brew install gnu-getopt'. Aborting."; exit 1; }
|
|
fi
|
|
|
|
! ${GETOPT_BINARY} --test > /dev/null
|
|
if [[ ${PIPESTATUS[0]} != 4 ]]; then
|
|
# shellcheck disable=2016
|
|
echo '`getopt --test` failed in this environment.'
|
|
exit 1
|
|
fi
|
|
|
|
CURL_BINARY="$(command -v curl)" || { echo "Curl not installed. Aborting."; exit 1; }
|
|
JQ_BINARY="$(command -v jq)" || { echo "Jq not installed. Aborting."; exit 1; }
|
|
|
|
OPTS="ht:n:d:g"
|
|
LONGOPTS="help,preset:,nodes:,data-dir:,remote-validators-count:,threshold:,remote-signers:,with-ganache,stop-at-epoch:,disable-htop,disable-vc,enable-logtrace,log-level:,base-port:,base-rest-port:,base-metrics-port:,reuse-existing-data-dir,reuse-binaries,timeout:,kill-old-processes,eth2-docker-image:,lighthouse-vc-nodes:"
|
|
|
|
# default values
|
|
NUM_NODES="10"
|
|
DATA_DIR="local_testnet_data"
|
|
USE_HTOP="1"
|
|
USE_VC="1"
|
|
LIGHTHOUSE_VC_NODES="0"
|
|
USE_GANACHE="0"
|
|
LOG_LEVEL="DEBUG; TRACE:networking"
|
|
BASE_PORT="9000"
|
|
BASE_REMOTE_SIGNER_PORT="6000"
|
|
BASE_METRICS_PORT="8008"
|
|
BASE_REST_PORT="7500"
|
|
REUSE_EXISTING_DATA_DIR="0"
|
|
REUSE_BINARIES="0"
|
|
ENABLE_LOGTRACE="0"
|
|
STOP_AT_EPOCH_FLAG=""
|
|
TIMEOUT_DURATION="0"
|
|
CONST_PRESET="mainnet"
|
|
KILL_OLD_PROCESSES="0"
|
|
ETH2_DOCKER_IMAGE=""
|
|
REMOTE_SIGNER_NODES=0
|
|
REMOTE_SIGNER_THRESHOLD=1
|
|
REMOTE_VALIDATORS_COUNT=0
|
|
LC_NODES=1
|
|
|
|
print_help() {
|
|
cat <<EOF
|
|
Usage: $(basename "$0") [OPTIONS] -- [BEACON NODE OPTIONS]
|
|
E.g.: $(basename "$0") --nodes ${NUM_NODES} --stop-at-epoch 5 --data-dir "${DATA_DIR}" # defaults
|
|
CI run: $(basename "$0") --disable-htop -- --verify-finalization
|
|
|
|
-h, --help show this help message
|
|
-n, --nodes number of nodes to launch (default: ${NUM_NODES})
|
|
-g, --with-ganache simulate a genesis event based on a deposit contract
|
|
-s, --stop-at-epoch stop simulation at epoch (default: infinite)
|
|
-d, --data-dir directory where all the node data and logs will end up
|
|
(default: "${DATA_DIR}")
|
|
--preset Const preset to be (default: mainnet)
|
|
--base-port bootstrap node's Eth2 traffic port (default: ${BASE_PORT})
|
|
--base-rest-port bootstrap node's REST port (default: ${BASE_REST_PORT})
|
|
--base-metrics-port bootstrap node's metrics server port (default: ${BASE_METRICS_PORT})
|
|
--disable-htop don't use "htop" to see the nimbus_beacon_node processes
|
|
--disable-vc don't use validator client binaries for validators
|
|
(by default validators are split 50/50 between beacon nodes
|
|
and validator clients, with all beacon nodes being paired up
|
|
with a corresponding validator client)
|
|
--lighthouse-vc-nodes number of Lighthouse VC nodes (assigned before Nimbus VC nodes, default: ${LIGHTHOUSE_VC_NODES})
|
|
--enable-logtrace display logtrace analysis
|
|
--log-level set the log level (default: "${LOG_LEVEL}")
|
|
--reuse-existing-data-dir instead of deleting and recreating the data dir, keep it and reuse everything we can from it
|
|
--reuse-binaries don't (re)build the binaries we need and don't delete them at the end (speeds up testing)
|
|
--timeout timeout in seconds (default: ${TIMEOUT_DURATION} - no timeout)
|
|
--kill-old-processes if any process is found listening on a port we use, kill it (default: disabled)
|
|
--eth2-docker-image use docker image instead of compiling the beacon node
|
|
--remote-validators-count number of remote validators which will be generated
|
|
--threshold used by a threshold secret sharing mechanism and determine how many shares are need to
|
|
restore signature of the original secret key
|
|
--remote-signers number of remote signing nodes
|
|
--light-clients number of light clients
|
|
EOF
|
|
}
|
|
|
|
! PARSED=$(${GETOPT_BINARY} --options=${OPTS} --longoptions=${LONGOPTS} --name "$0" -- "$@")
|
|
if [[ ${PIPESTATUS[0]} != 0 ]]; then
|
|
# getopt has complained about wrong arguments to stdout
|
|
exit 1
|
|
fi
|
|
|
|
# read getopt's output this way to handle the quoting right
|
|
eval set -- "$PARSED"
|
|
while true; do
|
|
case "$1" in
|
|
-h|--help)
|
|
print_help
|
|
exit
|
|
;;
|
|
-n|--nodes)
|
|
NUM_NODES="$2"
|
|
shift 2
|
|
;;
|
|
--remote-signers)
|
|
REMOTE_SIGNER_NODES=$2
|
|
shift 2
|
|
;;
|
|
--remote-validators-count)
|
|
REMOTE_VALIDATORS_COUNT=$2
|
|
shift 2
|
|
;;
|
|
--threshold)
|
|
REMOTE_SIGNER_THRESHOLD=$2
|
|
shift 2
|
|
;;
|
|
-d|--data-dir)
|
|
DATA_DIR="$2"
|
|
shift 2
|
|
;;
|
|
-g|--with-ganache)
|
|
USE_GANACHE="1"
|
|
shift
|
|
;;
|
|
--preset)
|
|
CONST_PRESET="$2"
|
|
shift 2
|
|
;;
|
|
--stop-at-epoch)
|
|
STOP_AT_EPOCH_FLAG="--stop-at-epoch=$2"
|
|
shift 2
|
|
;;
|
|
--disable-htop)
|
|
USE_HTOP="0"
|
|
shift
|
|
;;
|
|
--disable-vc)
|
|
USE_VC="0"
|
|
shift
|
|
;;
|
|
--enable-logtrace)
|
|
ENABLE_LOGTRACE="1"
|
|
shift
|
|
;;
|
|
--log-level)
|
|
LOG_LEVEL="$2"
|
|
shift 2
|
|
;;
|
|
--base-port)
|
|
BASE_PORT="$2"
|
|
shift 2
|
|
;;
|
|
--base-rest-port)
|
|
BASE_REST_PORT="$2"
|
|
shift 2
|
|
;;
|
|
--base-metrics-port)
|
|
BASE_METRICS_PORT="$2"
|
|
shift 2
|
|
;;
|
|
--reuse-existing-data-dir)
|
|
REUSE_EXISTING_DATA_DIR="1"
|
|
shift
|
|
;;
|
|
--reuse-binaries)
|
|
REUSE_BINARIES="1"
|
|
shift
|
|
;;
|
|
--timeout)
|
|
TIMEOUT_DURATION="$2"
|
|
shift 2
|
|
;;
|
|
--kill-old-processes)
|
|
KILL_OLD_PROCESSES="1"
|
|
shift
|
|
;;
|
|
--eth2-docker-image)
|
|
ETH2_DOCKER_IMAGE="$2"
|
|
shift 2
|
|
# TODO The validator client is still not being shipped with
|
|
# our docker images, so we must disable it:
|
|
echo "warning: --eth-docker-image implies --disable-vc"
|
|
USE_VC="0"
|
|
;;
|
|
--lighthouse-vc-nodes)
|
|
LIGHTHOUSE_VC_NODES="$2"
|
|
shift 2
|
|
;;
|
|
--light-clients)
|
|
LC_NODES="$2"
|
|
shift 2
|
|
;;
|
|
--)
|
|
shift
|
|
break
|
|
;;
|
|
*)
|
|
echo "argument parsing error"
|
|
print_help
|
|
exit 1
|
|
esac
|
|
done
|
|
|
|
# when sourcing env.sh, it will try to execute $@, so empty it
|
|
EXTRA_ARGS="$@"
|
|
if [[ $# != 0 ]]; then
|
|
shift $#
|
|
fi
|
|
|
|
if [[ "$REUSE_EXISTING_DATA_DIR" == "0" ]]; then
|
|
rm -rf "${DATA_DIR}"
|
|
fi
|
|
|
|
if [[ "${LIGHTHOUSE_VC_NODES}" != "0" && "${CONST_PRESET}" != "mainnet" ]]; then
|
|
echo "The prebuilt Lighthouse binary we're using only supports mainnet. Aborting."
|
|
exit 1
|
|
fi
|
|
|
|
scripts/makedir.sh "${DATA_DIR}"
|
|
|
|
VALIDATORS_DIR="${DATA_DIR}/validators"
|
|
scripts/makedir.sh "${VALIDATORS_DIR}"
|
|
|
|
SECRETS_DIR="${DATA_DIR}/secrets"
|
|
scripts/makedir.sh "${SECRETS_DIR}"
|
|
|
|
USER_VALIDATORS=8
|
|
TOTAL_VALIDATORS=128
|
|
|
|
# "Make" binary
|
|
if [[ "${OS}" == "windows" ]]; then
|
|
MAKE="mingw32-make"
|
|
else
|
|
MAKE="make"
|
|
fi
|
|
|
|
# number of CPU cores
|
|
if uname | grep -qi darwin; then
|
|
NPROC="$(sysctl -n hw.logicalcpu)"
|
|
else
|
|
NPROC="$(nproc)"
|
|
fi
|
|
|
|
# kill lingering processes from a previous run
|
|
if [[ "${OS}" != "windows" ]]; then
|
|
which lsof &>/dev/null || \
|
|
{ echo "'lsof' not installed and we need it to check for ports already in use. Aborting."; exit 1; }
|
|
|
|
for NUM_NODE in $(seq 0 $(( NUM_NODES - 1 ))); do
|
|
for PORT in $(( BASE_PORT + NUM_NODE )) $(( BASE_METRICS_PORT + NUM_NODE )) $(( BASE_REST_PORT + NUM_NODE )); do
|
|
for PID in $(lsof -n -i tcp:${PORT} -sTCP:LISTEN -t); do
|
|
echo -n "Found old process listening on port ${PORT}, with PID ${PID}. "
|
|
if [[ "${KILL_OLD_PROCESSES}" == "1" ]]; then
|
|
echo "Killing it."
|
|
kill -9 ${PID} || true
|
|
else
|
|
echo "Aborting."
|
|
exit 1
|
|
fi
|
|
done
|
|
done
|
|
done
|
|
fi
|
|
|
|
# Download the Lighthouse binary.
|
|
LH_VERSION="2.1.3"
|
|
LH_ARCH="${ARCH}"
|
|
if [[ "${LH_ARCH}" == "arm64" ]]; then
|
|
LH_ARCH="aarch64"
|
|
fi
|
|
|
|
case "${OS}" in
|
|
linux)
|
|
LH_TARBALL="lighthouse-v${LH_VERSION}-${LH_ARCH}-unknown-linux-gnu-portable.tar.gz"
|
|
;;
|
|
macos)
|
|
LH_TARBALL="lighthouse-v${LH_VERSION}-${LH_ARCH}-apple-darwin-portable.tar.gz"
|
|
;;
|
|
windows)
|
|
LH_TARBALL="lighthouse-v${LH_VERSION}-${LH_ARCH}-windows-portable.tar.gz"
|
|
;;
|
|
esac
|
|
LH_URL="https://github.com/sigp/lighthouse/releases/download/v${LH_VERSION}/${LH_TARBALL}"
|
|
LH_BINARY="lighthouse-${LH_VERSION}"
|
|
|
|
if [[ "${USE_VC}" == "1" && "${LIGHTHOUSE_VC_NODES}" != "0" && ! -e "build/${LH_BINARY}" ]]; then
|
|
pushd "build" >/dev/null
|
|
"${CURL_BINARY}" -sSLO "${LH_URL}"
|
|
tar -xzf "${LH_TARBALL}" # contains just one file named "lighthouse"
|
|
rm lighthouse-* # deletes both the tarball and old binary versions
|
|
mv lighthouse "${LH_BINARY}"
|
|
popd >/dev/null
|
|
fi
|
|
|
|
# Build the binaries
|
|
BINARIES="deposit_contract"
|
|
|
|
if [ "$REMOTE_SIGNER_NODES" -ge "0" ]; then
|
|
BINARIES="${BINARIES} nimbus_signing_node"
|
|
fi
|
|
|
|
if [[ "${USE_VC}" == "1" ]]; then
|
|
BINARIES="${BINARIES} nimbus_validator_client"
|
|
fi
|
|
|
|
if [ "$LC_NODES" -ge "1" ]; then
|
|
BINARIES="${BINARIES} nimbus_light_client"
|
|
fi
|
|
|
|
if [[ "$ENABLE_LOGTRACE" == "1" ]]; then
|
|
BINARIES="${BINARIES} logtrace"
|
|
fi
|
|
|
|
if [[ -n "${ETH2_DOCKER_IMAGE}" ]]; then
|
|
DATA_DIR_FULL_PATH="$(cd "${DATA_DIR}"; pwd)"
|
|
# CONTAINER_DATA_DIR must be used everywhere where paths are supplied to BEACON_NODE_COMMAND executions.
|
|
# We'll use the CONTAINER_ prefix throughout the file to indicate such paths.
|
|
CONTAINER_DATA_DIR="/home/user/nimbus-eth2/testnet"
|
|
BEACON_NODE_COMMAND="docker run -v /etc/passwd:/etc/passwd -u $(id -u):$(id -g) --net=host -v ${DATA_DIR_FULL_PATH}:${CONTAINER_DATA_DIR}:rw $ETH2_DOCKER_IMAGE"
|
|
else
|
|
# When docker is not used CONTAINER_DATA_DIR is just an alias for DATA_DIR
|
|
CONTAINER_DATA_DIR="${DATA_DIR}"
|
|
BEACON_NODE_COMMAND="./build/nimbus_beacon_node"
|
|
BINARIES="${BINARIES} nimbus_beacon_node"
|
|
fi
|
|
|
|
BINARIES_MISSING="0"
|
|
for BINARY in ${BINARIES}; do
|
|
if [[ ! -e "build/${BINARY}" ]]; then
|
|
BINARIES_MISSING="1"
|
|
break
|
|
fi
|
|
done
|
|
if [[ "${REUSE_BINARIES}" == "0" || "${BINARIES_MISSING}" == "1" ]]; then
|
|
${MAKE} -j ${NPROC} LOG_LEVEL=TRACE NIMFLAGS="${NIMFLAGS} -d:local_testnet -d:const_preset=${CONST_PRESET}" ${BINARIES}
|
|
fi
|
|
|
|
# Kill child processes on Ctrl-C/SIGTERM/exit, passing the PID of this shell
|
|
# instance as the parent and the target process name as a pattern to the
|
|
# "pkill" command.
|
|
cleanup() {
|
|
pkill -f -P $$ nimbus_beacon_node &>/dev/null || true
|
|
pkill -f -P $$ nimbus_validator_client &>/dev/null || true
|
|
pkill -f -P $$ nimbus_signing_node &>/dev/null || true
|
|
pkill -f -P $$ nimbus_light_client &>/dev/null || true
|
|
pkill -f -P $$ ${LH_BINARY} &>/dev/null || true
|
|
sleep 2
|
|
pkill -f -9 -P $$ nimbus_beacon_node &>/dev/null || true
|
|
pkill -f -9 -P $$ nimbus_validator_client &>/dev/null || true
|
|
pkill -f -9 -P $$ nimbus_signing_node &>/dev/null || true
|
|
pkill -f -9 -P $$ nimbus_light_client &>/dev/null || true
|
|
pkill -f -9 -P $$ ${LH_BINARY} &>/dev/null || true
|
|
|
|
# Delete all binaries we just built, because these are unusable outside this
|
|
# local testnet.
|
|
if [[ "${REUSE_BINARIES}" == "0" ]]; then
|
|
for BINARY in ${BINARIES}; do
|
|
rm -f build/${BINARY}
|
|
done
|
|
fi
|
|
|
|
if [[ -n "$ETH2_DOCKER_IMAGE" ]]; then
|
|
docker rm $(docker stop $(docker ps -a -q --filter ancestor=$ETH2_DOCKER_IMAGE --format="{{.ID}}"))
|
|
fi
|
|
}
|
|
trap 'cleanup' SIGINT SIGTERM EXIT
|
|
|
|
# timeout - implemented with a background job
|
|
timeout_reached() {
|
|
echo -e "\nTimeout reached. Aborting.\n"
|
|
cleanup
|
|
}
|
|
trap 'timeout_reached' SIGALRM
|
|
|
|
if [[ "${TIMEOUT_DURATION}" != "0" ]]; then
|
|
export PARENT_PID=$$
|
|
( sleep ${TIMEOUT_DURATION} && kill -ALRM ${PARENT_PID} ) 2>/dev/null & WATCHER_PID=$!
|
|
fi
|
|
|
|
REMOTE_URLS=""
|
|
|
|
for NUM_REMOTE in $(seq 0 $(( REMOTE_SIGNER_NODES - 1 ))); do
|
|
REMOTE_PORT=$(( BASE_REMOTE_SIGNER_PORT + NUM_REMOTE ))
|
|
REMOTE_URLS="${REMOTE_URLS} --remote-signer=http://127.0.0.1:${REMOTE_PORT}"
|
|
done
|
|
|
|
# deposit and testnet creation
|
|
PIDS=""
|
|
WEB3_ARG=""
|
|
BOOTSTRAP_TIMEOUT=30 # in seconds
|
|
DEPOSIT_CONTRACT_ADDRESS="0x0000000000000000000000000000000000000000"
|
|
DEPOSIT_CONTRACT_BLOCK="0x0000000000000000000000000000000000000000000000000000000000000000"
|
|
RUNTIME_CONFIG_FILE="${DATA_DIR}/config.yaml"
|
|
NUM_JOBS=${NUM_NODES}
|
|
|
|
DEPOSITS_FILE="${DATA_DIR}/deposits.json"
|
|
CONTAINER_DEPOSITS_FILE="${CONTAINER_DATA_DIR}/deposits.json"
|
|
|
|
if [[ "$REUSE_EXISTING_DATA_DIR" == "0" ]]; then
|
|
./build/deposit_contract generateSimulationDeposits \
|
|
--count=${TOTAL_VALIDATORS} \
|
|
--out-validators-dir="${VALIDATORS_DIR}" \
|
|
--out-secrets-dir="${SECRETS_DIR}" \
|
|
--out-deposits-file="${DEPOSITS_FILE}" \
|
|
--threshold=${REMOTE_SIGNER_THRESHOLD} \
|
|
--remote-validators-count=${REMOTE_VALIDATORS_COUNT} \
|
|
${REMOTE_URLS}
|
|
fi
|
|
|
|
if [[ $USE_GANACHE == "0" ]]; then
|
|
GENESIS_OFFSET=30
|
|
BOOTSTRAP_IP="127.0.0.1"
|
|
|
|
$BEACON_NODE_COMMAND createTestnet \
|
|
--data-dir="${CONTAINER_DATA_DIR}" \
|
|
--deposits-file="${CONTAINER_DEPOSITS_FILE}" \
|
|
--total-validators=${TOTAL_VALIDATORS} \
|
|
--output-genesis="${CONTAINER_DATA_DIR}/genesis.ssz" \
|
|
--output-bootstrap-file="${CONTAINER_DATA_DIR}/bootstrap_nodes.txt" \
|
|
--bootstrap-address=${BOOTSTRAP_IP} \
|
|
--bootstrap-port=${BASE_PORT} \
|
|
--netkey-file=network_key.json \
|
|
--insecure-netkey-password=true \
|
|
--genesis-offset=${GENESIS_OFFSET} # Delay in seconds
|
|
|
|
else
|
|
echo "Launching ganache"
|
|
ganache-cli --blockTime 17 --gasLimit 100000000 -e 100000 --verbose > "${DATA_DIR}/log_ganache.txt" 2>&1 &
|
|
PIDS="${PIDS},$!"
|
|
|
|
WEB3_ARG="--web3-url=ws://localhost:8545"
|
|
|
|
echo "Deploying deposit contract"
|
|
DEPLOY_CMD_OUTPUT=$(./build/deposit_contract deploy $WEB3_ARG)
|
|
# https://stackoverflow.com/questions/918886/how-do-i-split-a-string-on-a-delimiter-in-bash
|
|
OUTPUT_PIECES=(${DEPLOY_CMD_OUTPUT//;/ })
|
|
DEPOSIT_CONTRACT_ADDRESS=${OUTPUT_PIECES[0]}
|
|
DEPOSIT_CONTRACT_BLOCK=${OUTPUT_PIECES[1]}
|
|
|
|
echo Contract deployed at "$DEPOSIT_CONTRACT_ADDRESS":"$DEPOSIT_CONTRACT_BLOCK"
|
|
|
|
MIN_DELAY=1
|
|
MAX_DELAY=5
|
|
|
|
BOOTSTRAP_TIMEOUT=$(( MAX_DELAY * TOTAL_VALIDATORS ))
|
|
|
|
./build/deposit_contract sendDeposits \
|
|
--deposits-file="${DEPOSITS_FILE}" \
|
|
--min-delay=$MIN_DELAY --max-delay=$MAX_DELAY \
|
|
$WEB3_ARG \
|
|
--deposit-contract=${DEPOSIT_CONTRACT_ADDRESS} > "${DATA_DIR}/log_deposit_maker.txt" 2>&1 &
|
|
|
|
PIDS="${PIDS},$!"
|
|
fi
|
|
|
|
./scripts/make_prometheus_config.sh \
|
|
--nodes ${NUM_NODES} \
|
|
--base-metrics-port ${BASE_METRICS_PORT} \
|
|
--config-file "${DATA_DIR}/prometheus.yml" || true # TODO: this currently fails on macOS,
|
|
# but it can be considered non-critical
|
|
echo Wrote $RUNTIME_CONFIG_FILE:
|
|
|
|
# TODO the runtime config file should be used during deposit generation as well!
|
|
tee "$RUNTIME_CONFIG_FILE" <<EOF
|
|
PRESET_BASE: ${CONST_PRESET}
|
|
MIN_GENESIS_ACTIVE_VALIDATOR_COUNT: ${TOTAL_VALIDATORS}
|
|
MIN_GENESIS_TIME: 0
|
|
GENESIS_DELAY: 10
|
|
DEPOSIT_CONTRACT_ADDRESS: ${DEPOSIT_CONTRACT_ADDRESS}
|
|
ETH1_FOLLOW_DISTANCE: 1
|
|
ALTAIR_FORK_EPOCH: 1
|
|
BELLATRIX_FORK_EPOCH: 2
|
|
EOF
|
|
|
|
if [[ "${LIGHTHOUSE_VC_NODES}" != "0" ]]; then
|
|
# I don't know what this is, but Lighthouse wants it, so we recreate it from
|
|
# Lighthouse's own local testnet.
|
|
echo 0 > "${DATA_DIR}/deploy_block.txt"
|
|
|
|
# Lighthouse wants all these variables here. Copying them from "beacon_chain/spec/presets.nim".
|
|
# Note: our parser can't handle quotes around numerical values.
|
|
cat >> "$RUNTIME_CONFIG_FILE" <<EOF
|
|
GENESIS_FORK_VERSION: 0x00000000
|
|
ALTAIR_FORK_VERSION: 0x01000000
|
|
SECONDS_PER_SLOT: 12
|
|
SECONDS_PER_ETH1_BLOCK: 14
|
|
MIN_VALIDATOR_WITHDRAWABILITY_DELAY: 256
|
|
SHARD_COMMITTEE_PERIOD: 256
|
|
INACTIVITY_SCORE_BIAS: 4
|
|
INACTIVITY_SCORE_RECOVERY_RATE: 16
|
|
EJECTION_BALANCE: 16000000000
|
|
MIN_PER_EPOCH_CHURN_LIMIT: 4
|
|
CHURN_LIMIT_QUOTIENT: 65536
|
|
DEPOSIT_CHAIN_ID: 1
|
|
DEPOSIT_NETWORK_ID: 1
|
|
EOF
|
|
fi
|
|
|
|
dump_logs() {
|
|
LOG_LINES=20
|
|
for LOG in "${DATA_DIR}"/log*.txt; do
|
|
echo "Last ${LOG_LINES} lines of ${LOG}:"
|
|
tail -n ${LOG_LINES} "${LOG}"
|
|
echo "======"
|
|
done
|
|
}
|
|
|
|
dump_logtrace() {
|
|
if [[ "$ENABLE_LOGTRACE" == "1" ]]; then
|
|
find "${DATA_DIR}" -maxdepth 1 -type f -regex '.*/log[0-9]+.txt' | sed -e"s/${DATA_DIR}\//--nodes=/" | sort | xargs ./build/logtrace aggasr --log-dir="${DATA_DIR}" || true
|
|
fi
|
|
}
|
|
|
|
NODES_WITH_VALIDATORS=${NODES_WITH_VALIDATORS:-4}
|
|
BOOTSTRAP_NODE=0
|
|
SYSTEM_VALIDATORS=$(( TOTAL_VALIDATORS - USER_VALIDATORS ))
|
|
VALIDATORS_PER_NODE=$(( SYSTEM_VALIDATORS / NODES_WITH_VALIDATORS ))
|
|
if [[ "${USE_VC}" == "1" ]]; then
|
|
# if using validator client binaries in addition to beacon nodes we will
|
|
# split the keys for this instance in half between the BN and the VC
|
|
# and the validators for the BNs will be from the first half of all validators
|
|
VALIDATORS_PER_NODE=$((VALIDATORS_PER_NODE / 2 ))
|
|
NUM_JOBS=$((NUM_JOBS * 2 ))
|
|
fi
|
|
|
|
if [ "$REMOTE_SIGNER_NODES" -ge "0" ]; then
|
|
NUM_JOBS=$((NUM_JOBS + REMOTE_SIGNER_NODES ))
|
|
fi
|
|
|
|
if [ "$LC_NODES" -ge "1" ]; then
|
|
NUM_JOBS=$((NUM_JOBS + LC_NODES ))
|
|
fi
|
|
|
|
VALIDATORS_PER_VALIDATOR=$(( (SYSTEM_VALIDATORS / NODES_WITH_VALIDATORS) / 2 ))
|
|
VALIDATOR_OFFSET=$((SYSTEM_VALIDATORS / 2))
|
|
|
|
BOOTSTRAP_ENR="${DATA_DIR}/node${BOOTSTRAP_NODE}/beacon_node.enr"
|
|
CONTAINER_BOOTSTRAP_ENR="${CONTAINER_DATA_DIR}/node${BOOTSTRAP_NODE}/beacon_node.enr"
|
|
|
|
CONTAINER_NETWORK_KEYFILE="network_key.json"
|
|
|
|
for NUM_NODE in $(seq 0 $(( NUM_NODES - 1 ))); do
|
|
# Copy validators to individual nodes.
|
|
# The first $NODES_WITH_VALIDATORS nodes split them equally between them,
|
|
# after skipping the first $USER_VALIDATORS.
|
|
NODE_DATA_DIR="${DATA_DIR}/node${NUM_NODE}"
|
|
rm -rf "${NODE_DATA_DIR}"
|
|
scripts/makedir.sh "${NODE_DATA_DIR}" 2>&1
|
|
scripts/makedir.sh "${NODE_DATA_DIR}/validators" 2>&1
|
|
scripts/makedir.sh "${NODE_DATA_DIR}/secrets" 2>&1
|
|
|
|
if [[ $NUM_NODE -lt $NODES_WITH_VALIDATORS ]]; then
|
|
if [[ "${USE_VC}" == "1" ]]; then
|
|
VALIDATOR_DATA_DIR="${DATA_DIR}/validator${NUM_NODE}"
|
|
rm -rf "${VALIDATOR_DATA_DIR}"
|
|
scripts/makedir.sh "${VALIDATOR_DATA_DIR}" 2>&1
|
|
scripts/makedir.sh "${VALIDATOR_DATA_DIR}/validators" 2>&1
|
|
scripts/makedir.sh "${VALIDATOR_DATA_DIR}/secrets" 2>&1
|
|
for VALIDATOR in $(ls "${VALIDATORS_DIR}" | tail -n +$(( $USER_VALIDATORS + ($VALIDATORS_PER_VALIDATOR * $NUM_NODE) + 1 + $VALIDATOR_OFFSET )) | head -n $VALIDATORS_PER_VALIDATOR); do
|
|
if [[ -f "${VALIDATORS_DIR}/${VALIDATOR}/keystore.json" ]]; then
|
|
cp -a "${VALIDATORS_DIR}/${VALIDATOR}" "${VALIDATOR_DATA_DIR}/validators/" 2>&1
|
|
cp -a "${SECRETS_DIR}/${VALIDATOR}" "${VALIDATOR_DATA_DIR}/secrets/" 2>&1
|
|
else
|
|
# TODO: validators support remote signers
|
|
cp -a "${VALIDATORS_DIR}/${VALIDATOR}" "${NODE_DATA_DIR}/validators/" 2>&1
|
|
fi
|
|
done
|
|
if [[ "${OS}" == "Windows_NT" ]]; then
|
|
find "${VALIDATOR_DATA_DIR}" -type f \( -iname "*.json" -o ! -iname "*.*" \) -exec icacls "{}" /inheritance:r /grant:r ${USERDOMAIN}\\${USERNAME}:\(F\) \;
|
|
fi
|
|
fi
|
|
for VALIDATOR in $(ls "${VALIDATORS_DIR}" | tail -n +$(( $USER_VALIDATORS + ($VALIDATORS_PER_NODE * $NUM_NODE) + 1 )) | head -n $VALIDATORS_PER_NODE); do
|
|
cp -a "${VALIDATORS_DIR}/${VALIDATOR}" "${NODE_DATA_DIR}/validators/" 2>&1
|
|
if [[ -f "${VALIDATORS_DIR}/${VALIDATOR}/keystore.json" ]]; then
|
|
# Only remote key stores doesn't have a secret
|
|
cp -a "${SECRETS_DIR}/${VALIDATOR}" "${NODE_DATA_DIR}/secrets/" 2>&1
|
|
fi
|
|
done
|
|
if [[ "${OS}" == "Windows_NT" ]]; then
|
|
find "${NODE_DATA_DIR}" -type f \( -iname "*.json" -o ! -iname "*.*" \) -exec icacls "{}" /inheritance:r /grant:r ${USERDOMAIN}\\${USERNAME}:\(F\) \;
|
|
fi
|
|
fi
|
|
done
|
|
|
|
CLI_CONF_FILE="$CONTAINER_DATA_DIR/config.toml"
|
|
|
|
cat > "$CLI_CONF_FILE" <<END_CLI_CONFIG
|
|
non-interactive = true
|
|
nat = "extip:127.0.0.1"
|
|
network = "${CONTAINER_DATA_DIR}"
|
|
log-level = "${LOG_LEVEL}"
|
|
log-format = "json"
|
|
rest = true
|
|
rest-address = "127.0.0.1"
|
|
metrics = true
|
|
metrics-address = "127.0.0.1"
|
|
END_CLI_CONFIG
|
|
|
|
for NUM_REMOTE in $(seq 0 $(( REMOTE_SIGNER_NODES - 1 ))); do
|
|
./build/nimbus_signing_node \
|
|
--validators-dir="${DATA_DIR}/validators_shares/${NUM_REMOTE}" \
|
|
--secrets-dir="${DATA_DIR}/secrets_shares/${NUM_REMOTE}" \
|
|
--bind-port=$(( BASE_REMOTE_SIGNER_PORT + NUM_REMOTE )) \
|
|
> "${DATA_DIR}/log_remote_signer_${NUM_REMOTE}.txt" &
|
|
done
|
|
|
|
# give each node time to load keys
|
|
sleep 10
|
|
|
|
for NUM_NODE in $(seq 0 $(( NUM_NODES - 1 ))); do
|
|
NODE_DATA_DIR="${DATA_DIR}/node${NUM_NODE}"
|
|
CONTAINER_NODE_DATA_DIR="${CONTAINER_DATA_DIR}/node${NUM_NODE}"
|
|
VALIDATOR_DATA_DIR="${DATA_DIR}/validator${NUM_NODE}"
|
|
if [[ ${NUM_NODE} == ${BOOTSTRAP_NODE} ]]; then
|
|
# Due to star topology, the bootstrap node must relay all attestations,
|
|
# even if it itself is not interested. --subscribe-all-subnets could be
|
|
# removed by switching to a fully-connected topology.
|
|
BOOTSTRAP_ARG="--netkey-file=${CONTAINER_NETWORK_KEYFILE} --insecure-netkey-password=true --subscribe-all-subnets"
|
|
else
|
|
BOOTSTRAP_ARG="--bootstrap-file=${CONTAINER_BOOTSTRAP_ENR}"
|
|
|
|
if [[ "${CONST_PRESET}" == "minimal" ]]; then
|
|
# The fast epoch and slot times in the minimal config might cause the
|
|
# mesh to break down due to re-subscriptions happening within the prune
|
|
# backoff time
|
|
BOOTSTRAP_ARG="${BOOTSTRAP_ARG} --subscribe-all-subnets"
|
|
fi
|
|
|
|
# Wait for the master node to write out its address file
|
|
START_TIMESTAMP=$(date +%s)
|
|
while [[ ! -f "${BOOTSTRAP_ENR}" ]]; do
|
|
sleep 0.1
|
|
NOW_TIMESTAMP=$(date +%s)
|
|
if [[ "$(( NOW_TIMESTAMP - START_TIMESTAMP - GENESIS_OFFSET ))" -ge "$BOOTSTRAP_TIMEOUT" ]]; then
|
|
echo "Bootstrap node failed to start in ${BOOTSTRAP_TIMEOUT} seconds. Aborting."
|
|
dump_logs
|
|
exit 1
|
|
fi
|
|
done
|
|
fi
|
|
|
|
${BEACON_NODE_COMMAND} \
|
|
--config-file="${CLI_CONF_FILE}" \
|
|
--tcp-port=$(( BASE_PORT + NUM_NODE )) \
|
|
--udp-port=$(( BASE_PORT + NUM_NODE )) \
|
|
--max-peers=$(( NUM_NODES + LC_NODES - 1 )) \
|
|
--data-dir="${CONTAINER_NODE_DATA_DIR}" \
|
|
${BOOTSTRAP_ARG} \
|
|
${WEB3_ARG} \
|
|
${STOP_AT_EPOCH_FLAG} \
|
|
--rest-port="$(( BASE_REST_PORT + NUM_NODE ))" \
|
|
--metrics-port="$(( BASE_METRICS_PORT + NUM_NODE ))" \
|
|
--light-client-enable=on \
|
|
--serve-light-client-data=on \
|
|
--import-light-client-data=only-new \
|
|
${EXTRA_ARGS} \
|
|
&> "${DATA_DIR}/log${NUM_NODE}.txt" &
|
|
|
|
PIDS="${PIDS},$!"
|
|
|
|
if [[ "${USE_VC}" == "1" ]]; then
|
|
if [[ "${LIGHTHOUSE_VC_NODES}" -gt "${NUM_NODE}" ]]; then
|
|
# Lighthouse needs a different keystore filename for its auto-discovery process.
|
|
for D in "${VALIDATOR_DATA_DIR}/validators"/0x*; do
|
|
if [[ -e "${D}/keystore.json" ]]; then
|
|
mv "${D}/keystore.json" "${D}/voting-keystore.json"
|
|
fi
|
|
done
|
|
|
|
./build/${LH_BINARY} vc \
|
|
--debug-level "debug" \
|
|
--logfile-max-number 0 \
|
|
--log-format "JSON" \
|
|
--validators-dir "${VALIDATOR_DATA_DIR}" \
|
|
--secrets-dir "${VALIDATOR_DATA_DIR}/secrets" \
|
|
--beacon-nodes "http://127.0.0.1:$((BASE_REST_PORT + NUM_NODE))" \
|
|
--testnet-dir "${DATA_DIR}" \
|
|
--init-slashing-protection \
|
|
&> "${DATA_DIR}/log_val${NUM_NODE}.txt" &
|
|
# No "--stop-at-epoch" equivalent here, so we let these VC processes be
|
|
# killed the ugly way, when the script exits.
|
|
else
|
|
./build/nimbus_validator_client \
|
|
--log-level="${LOG_LEVEL}" \
|
|
${STOP_AT_EPOCH_FLAG} \
|
|
--data-dir="${VALIDATOR_DATA_DIR}" \
|
|
--beacon-node="http://127.0.0.1:$((BASE_REST_PORT + NUM_NODE))" \
|
|
&> "${DATA_DIR}/log_val${NUM_NODE}.txt" &
|
|
PIDS="${PIDS},$!"
|
|
fi
|
|
fi
|
|
done
|
|
|
|
# light clients
|
|
if [ "$LC_NODES" -ge "1" ]; then
|
|
echo "Waiting for Altair finalization"
|
|
ALTAIR_FORK_EPOCH="$(
|
|
"${CURL_BINARY}" -s "http://localhost:${BASE_REST_PORT}/eth/v1/config/spec" | \
|
|
"${JQ_BINARY}" -r '.data.ALTAIR_FORK_EPOCH')"
|
|
while :; do
|
|
CURRENT_FORK_EPOCH="$(
|
|
"${CURL_BINARY}" -s "http://localhost:${BASE_REST_PORT}/eth/v1/beacon/states/finalized/fork" | \
|
|
"${JQ_BINARY}" -r '.data.epoch')"
|
|
if [ "${CURRENT_FORK_EPOCH}" -ge "${ALTAIR_FORK_EPOCH}" ]; then
|
|
break
|
|
fi
|
|
sleep 1
|
|
done
|
|
|
|
echo "Altair finalized, launching $LC_NODES light client(s)"
|
|
LC_BOOTSTRAP_NODE="$(
|
|
"${CURL_BINARY}" -s "http://localhost:${BASE_REST_PORT}/eth/v1/node/identity" | \
|
|
"${JQ_BINARY}" -r '.data.enr')"
|
|
LC_TRUSTED_BLOCK_ROOT="$(
|
|
"${CURL_BINARY}" -s "http://localhost:${BASE_REST_PORT}/eth/v1/beacon/headers/finalized" | \
|
|
"${JQ_BINARY}" -r '.data.root')"
|
|
for NUM_LC in $(seq 0 $(( LC_NODES - 1 ))); do
|
|
./build/nimbus_light_client \
|
|
--log-level="${LOG_LEVEL}" \
|
|
--log-format="json" \
|
|
--network="${CONTAINER_DATA_DIR}" \
|
|
--bootstrap-node="${LC_BOOTSTRAP_NODE}" \
|
|
--tcp-port=$(( BASE_PORT + NUM_NODES + NUM_LC )) \
|
|
--udp-port=$(( BASE_PORT + NUM_NODES + NUM_LC )) \
|
|
--max-peers=$(( NUM_NODES + LC_NODES - 1 )) \
|
|
--nat="extip:127.0.0.1" \
|
|
--trusted-block-root="${LC_TRUSTED_BLOCK_ROOT}" \
|
|
${STOP_AT_EPOCH_FLAG} \
|
|
&> "${DATA_DIR}/log_lc${NUM_LC}.txt" &
|
|
PIDS="${PIDS},$!"
|
|
done
|
|
fi
|
|
|
|
# give the regular nodes time to crash
|
|
sleep 5
|
|
BG_JOBS="$(jobs | wc -l | tr -d ' ')"
|
|
if [[ "${TIMEOUT_DURATION}" != "0" ]]; then
|
|
BG_JOBS=$(( BG_JOBS - 1 )) # minus the timeout bg job
|
|
fi
|
|
if [[ "$BG_JOBS" != "$NUM_JOBS" ]]; then
|
|
echo "$(( NUM_JOBS - BG_JOBS )) nimbus_beacon_node/nimbus_validator_client/nimbus_light_client instance(s) exited early. Aborting."
|
|
dump_logs
|
|
dump_logtrace
|
|
exit 1
|
|
fi
|
|
|
|
# launch "htop" or wait for background jobs
|
|
if [[ "$USE_HTOP" == "1" ]]; then
|
|
htop -p "$PIDS"
|
|
# Cleanup is done when this script exists, since we listen to the EXIT signal.
|
|
else
|
|
FAILED=0
|
|
for PID in $(echo "$PIDS" | tr ',' ' '); do
|
|
wait "$PID" || FAILED="$(( FAILED += 1 ))"
|
|
done
|
|
if [[ "$FAILED" != "0" ]]; then
|
|
echo "${FAILED} child processes had non-zero exit codes (or exited early)."
|
|
dump_logs
|
|
dump_logtrace
|
|
if [[ "${TIMEOUT_DURATION}" != "0" ]]; then
|
|
if uname | grep -qiE "mingw|msys"; then
|
|
taskkill //F //PID ${WATCHER_PID}
|
|
else
|
|
pkill -HUP -P ${WATCHER_PID}
|
|
fi
|
|
fi
|
|
exit 1
|
|
fi
|
|
fi
|
|
|
|
dump_logtrace
|
|
|
|
if [[ "${TIMEOUT_DURATION}" != "0" ]]; then
|
|
if uname | grep -qiE "mingw|msys"; then
|
|
taskkill //F //PID ${WATCHER_PID}
|
|
else
|
|
pkill -HUP -P ${WATCHER_PID}
|
|
fi
|
|
fi
|