# Nimbus # Copyright (c) 2021 Status Research & Development GmbH # Licensed and distributed under either of # * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT). # * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0). # at your option. This file may not be copied, modified, or distributed except according to those terms. import os, std/sequtils, unittest2, testutils, confutils, chronos, stew/byteutils, eth/p2p/discoveryv5/random2, eth/keys, ../../nimbus/rpc/[hexstrings, rpc_types], ../rpc/portal_rpc_client, ../rpc/eth_rpc_client, ../data/[history_data_seeding, history_data_parser], ../seed_db type FutureCallback[A] = proc (): Future[A] {.gcsafe, raises: [Defect].} CheckCallback[A] = proc (a: A): bool {.gcsafe, raises: [Defect].} PortalTestnetConf* = object nodeCount* {. defaultValue: 17 desc: "Number of nodes to test" name: "node-count" .}: int rpcAddress* {. desc: "Listening address of the JSON-RPC service for all nodes" defaultValue: "127.0.0.1" name: "rpc-address" }: string baseRpcPort* {. defaultValue: 7000 desc: "Port of the JSON-RPC service of the bootstrap (first) node" name: "base-rpc-port" .}: uint16 proc connectToRpcServers(config: PortalTestnetConf): Future[seq[RpcClient]] {.async.} = var clients: seq[RpcClient] for i in 0.. numRetries: # if we reached max number of retries fail let msg = "Call failed with msg: " & exc.msg & ", for node with idx: " & $nodeIdx raise newException(ValueError, msg) inc tries # wait before new retry await sleepAsync(currentDuration) currentDuration = currentDuration * 2 # Sometimes we need to wait till data will be propagated over the network. # To avoid long sleeps, this combinator can be used to retry some calls until # success or until some condition hold (or both) proc retryUntil[A]( f: FutureCallback[A], c: CheckCallback[A], checkFailMessage: string, nodeIdx: int): Future[A] = # some reasonable limits, which will cause waits as: 1, 2, 4, 8, 16, 32 seconds return withRetries(f, c, 6, seconds(1), checkFailMessage, nodeIdx) # Note: # When doing json-rpc requests following `RpcPostError` can occur: # "Failed to send POST Request with JSON-RPC." when a `HttpClientRequestRef` # POST request is send in the json-rpc http client. # This error is raised when the httpclient hits error: # "Could not send request headers", which in its turn is caused by the # "Incomplete data sent or received" in `AsyncStream`, which is caused by # `ECONNRESET` or `EPIPE` error (see `isConnResetError()`) on the TCP stream. # This can occur when the server side closes the connection, which happens after # a `httpHeadersTimeout` of default 10 seconds (set on `HttpServerRef.new()`). # In order to avoid here hitting this timeout a `close()` is done after each # json-rpc call. Because the first json-rpc call opens up the connection, and it # remains open until a close() (or timeout). No need to do another connect # before any new call as the proc `connectToRpcServers` doesn't actually connect # to servers, as client.connect doesn't do that. It just sets the `httpAddress`. # Yes, this client json rpc API couldn't be more confusing. # Could also just retry each call on failure, which would set up a new # connection. # We are kind of abusing the unittest2 here to run json rpc tests against other # processes. Needs to be compiled with `-d:unittest2DisableParamFiltering` or # the confutils cli will not work. procSuite "Portal testnet tests": let config = PortalTestnetConf.load() let rng = newRng() asyncTest "Discv5 - Random node lookup from each node": let clients = await connectToRpcServers(config) var nodeInfos: seq[NodeInfo] for client in clients: let nodeInfo = await client.discv5_nodeInfo() await client.close() nodeInfos.add(nodeInfo) # Kick off the network by trying to add all records to each node. # These nodes are also set as seen, so they get passed along on findNode # requests. # Note: The amount of Records added here can be less but then the # probability that all nodes will still be reached needs to be calculated. # Note 2: One could also ping all nodes but that is much slower and more # error prone for client in clients: discard await client.discv5_addEnrs(nodeInfos.map( proc(x: NodeInfo): Record = x.nodeENR)) await client.close() for client in clients: let routingTableInfo = await client.discv5_routingTableInfo() await client.close() var start: seq[NodeId] let nodes = foldl(routingTableInfo.buckets, a & b, start) # A node will have at least the first bucket filled. One could increase # this based on the probability that x amount of nodes fit in the buckets. check nodes.len >= (min(config.nodeCount - 1, 16)) # grab a random node its `NodeInfo` and lookup that node from all nodes. let randomNodeInfo = sample(rng[], nodeInfos) for client in clients: var enr: Record enr = await client.discv5_lookupEnr(randomNodeInfo.nodeId) check enr == randomNodeInfo.nodeENR await client.close() asyncTest "Portal State - Random node lookup from each node": let clients = await connectToRpcServers(config) var nodeInfos: seq[NodeInfo] for client in clients: let nodeInfo = await client.portal_state_nodeInfo() await client.close() nodeInfos.add(nodeInfo) for client in clients: discard await client.portal_state_addEnrs(nodeInfos.map( proc(x: NodeInfo): Record = x.nodeENR)) await client.close() for client in clients: let routingTableInfo = await client.portal_state_routingTableInfo() await client.close() var start: seq[NodeId] let nodes = foldl(routingTableInfo.buckets, a & b, start) check nodes.len >= (min(config.nodeCount - 1, 16)) # grab a random node its `NodeInfo` and lookup that node from all nodes. let randomNodeInfo = sample(rng[], nodeInfos) for client in clients: var enr: Record try: enr = await client.portal_state_lookupEnr(randomNodeInfo.nodeId) except CatchableError as e: echo e.msg # TODO: For state network this occasionally fails. It might be because the # distance function is not used in all locations, or perhaps it just # doesn't converge to the target always with this distance function. To be # further investigated. skip() # check enr == randomNodeInfo.nodeENR await client.close() asyncTest "Portal History - Random node lookup from each node": let clients = await connectToRpcServers(config) var nodeInfos: seq[NodeInfo] for client in clients: let nodeInfo = await client.portal_history_nodeInfo() await client.close() nodeInfos.add(nodeInfo) for client in clients: discard await client.portal_history_addEnrs(nodeInfos.map( proc(x: NodeInfo): Record = x.nodeENR)) await client.close() for client in clients: let routingTableInfo = await client.portal_history_routingTableInfo() await client.close() var start: seq[NodeId] let nodes = foldl(routingTableInfo.buckets, a & b, start) check nodes.len >= (min(config.nodeCount - 1, 16)) # grab a random node its `NodeInfo` and lookup that node from all nodes. let randomNodeInfo = sample(rng[], nodeInfos) for client in clients: var enr: Record enr = await client.portal_history_lookupEnr(randomNodeInfo.nodeId) await client.close() check enr == randomNodeInfo.nodeENR asyncTest "Portal History - Propagate blocks and do content lookups": let clients = await connectToRpcServers(config) var nodeInfos: seq[NodeInfo] for client in clients: let nodeInfo = await client.portal_history_nodeInfo() await client.close() nodeInfos.add(nodeInfo) # const dataFileEpoch = "./fluffy/scripts/eth-epoch-accumulator.json" # check (await clients[0].portal_history_propagateEpochAccumulator(dataFileEpoch)) # await clients[0].close() # await sleepAsync(60.seconds) const dataFile = "./fluffy/tests/blocks/mainnet_blocks_1000001_1000010.json" check (await clients[0].portal_history_propagateHeaders(dataFile)) await clients[0].close() # Short sleep between propagation of block headers and propagation of block # bodies and receipts as the latter two require the first for validation. await sleepAsync(5.seconds) # This will fill the first node its db with blocks from the data file. Next, # this node wil offer all these blocks their headers one by one. check (await clients[0].portal_history_propagate(dataFile)) await clients[0].close() let blockData = readJsonType(dataFile, BlockDataTable) check blockData.isOk() for i, client in clients: # Note: Once there is the Canonical Indices Network, we don't need to # access this file anymore here for the block hashes. for hash in blockData.get().blockHashes(): # Note: More flexible approach instead of generic retries could be to # add a json-rpc debug proc that returns whether the offer queue is empty or # not. And then poll every node until all nodes have an empty queue. let content = await retryUntil( proc (): Future[Option[BlockObject]] {.async.} = try: let res = await client.eth_getBlockByHash(hash.ethHashStr(), false) await client.close() return res except CatchableError as exc: await client.close() raise exc , proc (mc: Option[BlockObject]): bool = return mc.isSome(), "Did not receive expected Block with hash " & hash.data.toHex(), i ) check content.isSome() let blockObj = content.get() check blockObj.hash.get() == hash for tx in blockObj.transactions: var txObj: TransactionObject tx.fromJson("tx", txObj) check txObj.blockHash.get() == hash let filterOptions = FilterOptions( blockHash: some(hash) ) let logs = await retryUntil( proc (): Future[seq[FilterLog]] {.async.} = try: let res = await client.eth_getLogs(filterOptions) await client.close() return res except CatchableError as exc: await client.close() raise exc , proc (mc: seq[FilterLog]): bool = return true, "", i ) for l in logs: check: l.blockHash == some(hash) # TODO: Check ommersHash, need the headers and not just the hashes # for uncle in blockObj.uncles: # discard await client.close() asyncTest "Portal History - Propagate content from seed db": let clients = await connectToRpcServers(config) var nodeInfos: seq[NodeInfo] for client in clients: let nodeInfo = await client.portal_history_nodeInfo() await client.close() nodeInfos.add(nodeInfo) const dataPath = "./fluffy/tests/blocks/mainnet_blocks_1000011_1000030.json" # path for temporary db, separate dir is used as sqlite usually also creates # wal files, and we do not want for those to linger in filesystem const tempDbPath = "./fluffy/tests/blocks/tempDir/mainnet_blocks_1000011_1000030.sqlite3" let (dbFile, dbName) = getDbBasePathAndName(tempDbPath).unsafeGet() let blockData = readJsonType(dataPath, BlockDataTable) check blockData.isOk() let bd = blockData.get() createDir(dbFile) let db = SeedDb.new(path = dbFile, name = dbName) try: let lastNodeIdx = len(nodeInfos) - 1 # populate temp database from json file for t in blocksContent(bd, false): db.put(t[0], t[1], t[2]) # store content in node0 database check (await clients[0].portal_history_storeContentInNodeRange(tempDbPath, 100, 0)) await clients[0].close() # offer content to node 1..63 for i in 1..lastNodeIdx: let receipientId = nodeInfos[i].nodeId let offerResponse = await retryUntil( proc (): Future[int] {.async.} = try: let res = await clients[0].portal_history_offerContentInNodeRange(tempDbPath, receipientId, 64, 0) await clients[0].close() return res except CatchableError as exc: await clients[0].close() raise exc , proc (os: int): bool = return true, "Offer failed", i ) check: offerResponse > 0 for i, client in clients: # Note: Once there is the Canonical Indices Network, we don't need to # access this file anymore here for the block hashes. for hash in bd.blockHashes(): let content = await retryUntil( proc (): Future[Option[BlockObject]] {.async.} = try: let res = await client.eth_getBlockByHash(hash.ethHashStr(), false) await client.close() return res except CatchableError as exc: await client.close() raise exc , proc (mc: Option[BlockObject]): bool = return mc.isSome(), "Did not receive expected Block with hash " & hash.data.toHex(), i ) check content.isSome() let blockObj = content.get() check blockObj.hash.get() == hash for tx in blockObj.transactions: var txObj: TransactionObject tx.fromJson("tx", txObj) check txObj.blockHash.get() == hash await client.close() finally: db.close() removeDir(dbFile) asyncTest "Portal History - Propagate content from seed db in depth first fashion": # Skipping this test as it is flawed considering block headers should be # offered before bodies and receipts. # TODO: Split this up and activate test skip() # let clients = await connectToRpcServers(config) # var nodeInfos: seq[NodeInfo] # for client in clients: # let nodeInfo = await client.portal_history_nodeInfo() # await client.close() # nodeInfos.add(nodeInfo) # # different set of data for each test as tests are statefull so previously propagated # # block are already in the network # const dataPath = "./fluffy/tests/blocks/mainnet_blocks_1000040_1000050.json" # # path for temporary db, separate dir is used as sqlite usually also creates # # wal files, and we do not want for those to linger in filesystem # const tempDbPath = "./fluffy/tests/blocks/tempDir/mainnet_blocks_1000040_100050.sqlite3" # let (dbFile, dbName) = getDbBasePathAndName(tempDbPath).unsafeGet() # let blockData = readJsonType(dataPath, BlockDataTable) # check blockData.isOk() # let bd = blockData.get() # createDir(dbFile) # let db = SeedDb.new(path = dbFile, name = dbName) # try: # # populate temp database from json file # for t in blocksContent(bd, false): # db.put(t[0], t[1], t[2]) # check (await clients[0].portal_history_depthContentPropagate(tempDbPath, 64)) # await clients[0].close() # for i, client in clients: # # Note: Once there is the Canonical Indices Network, we don't need to # # access this file anymore here for the block hashes. # for hash in bd.blockHashes(): # let content = await retryUntil( # proc (): Future[Option[BlockObject]] {.async.} = # try: # let res = await client.eth_getBlockByHash(hash.ethHashStr(), false) # await client.close() # return res # except CatchableError as exc: # await client.close() # raise exc # , # proc (mc: Option[BlockObject]): bool = return mc.isSome(), # "Did not receive expected Block with hash " & hash.data.toHex(), # i # ) # check content.isSome() # let blockObj = content.get() # check blockObj.hash.get() == hash # for tx in blockObj.transactions: # var txObj: TransactionObject # tx.fromJson("tx", txObj) # check txObj.blockHash.get() == hash # await client.close() # finally: # db.close() # removeDir(dbFile)