mirror of https://github.com/status-im/nim-eth.git
Address review feedback
This commit is contained in:
parent
6117fed595
commit
9e5cf2086c
|
@ -1,12 +1,20 @@
|
||||||
import
|
import
|
||||||
chronos/asyncfutures2, chronicles
|
chronos/[asyncfutures2, asyncloop], chronicles
|
||||||
|
|
||||||
|
proc catchOrQuit(error: Exception) =
|
||||||
|
if error of CatchableError:
|
||||||
|
trace "Async operation ended with a recoverable error", err = error.msg
|
||||||
|
else:
|
||||||
|
fatal "Fatal exception reached", err = error.msg
|
||||||
|
quit 1
|
||||||
|
|
||||||
proc traceAsyncErrors*(fut: FutureBase) =
|
proc traceAsyncErrors*(fut: FutureBase) =
|
||||||
fut.addCallback do (arg: pointer):
|
fut.addCallback do (arg: pointer):
|
||||||
if not fut.error.isNil:
|
if not fut.error.isNil:
|
||||||
if fut.error[] of CatchableError:
|
catchOrQuit fut.error[]
|
||||||
trace "Async operation ended with a recoverable error", err = fut.error.msg
|
|
||||||
else:
|
|
||||||
fatal "Fatal exception reached", err = fut.error.msg
|
|
||||||
quit 1
|
|
||||||
|
|
||||||
|
template traceAwaitErrors*(fut: FutureBase) =
|
||||||
|
let f = fut
|
||||||
|
yield f
|
||||||
|
if not f.error.isNil:
|
||||||
|
catchOrQuit f.error[]
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
import
|
import
|
||||||
macros, tables, algorithm, deques, hashes, options, typetraits,
|
macros, tables, algorithm, deques, hashes, options, typetraits,
|
||||||
std_shims/macros_shim, chronicles, nimcrypto, chronos, eth/[rlp, common, keys],
|
std_shims/macros_shim, chronicles, nimcrypto, chronos, eth/[rlp, common, keys, async_utils],
|
||||||
private/p2p_types, kademlia, auth, rlpxcrypt, enode
|
private/p2p_types, kademlia, auth, rlpxcrypt, enode
|
||||||
|
|
||||||
when useSnappy:
|
when useSnappy:
|
||||||
|
@ -503,6 +503,12 @@ proc nextMsg*(peer: Peer, MsgType: type): Future[MsgType] =
|
||||||
newFuture result
|
newFuture result
|
||||||
peer.awaitedMessages[wantedId] = result
|
peer.awaitedMessages[wantedId] = result
|
||||||
|
|
||||||
|
# Known fatal errors are handled inside dispatchMessages.
|
||||||
|
# Errors we are currently unaware of are caught in the dispatchMessages
|
||||||
|
# callback. There they will be logged if CatchableError and quit on Defect.
|
||||||
|
# Non fatal errors such as the current CatchableError could be moved and
|
||||||
|
# handled a layer lower for clarity (and consistency), as also the actual
|
||||||
|
# message handler code as the TODO mentions already.
|
||||||
proc dispatchMessages*(peer: Peer) {.async.} =
|
proc dispatchMessages*(peer: Peer) {.async.} =
|
||||||
while true:
|
while true:
|
||||||
var msgId: int
|
var msgId: int
|
||||||
|
@ -510,10 +516,10 @@ proc dispatchMessages*(peer: Peer) {.async.} =
|
||||||
try:
|
try:
|
||||||
(msgId, msgData) = await peer.recvMsg()
|
(msgId, msgData) = await peer.recvMsg()
|
||||||
except TransportIncompleteError:
|
except TransportIncompleteError:
|
||||||
trace "Connection dropped in dispatchMessages"
|
trace "Connection dropped, ending dispatchMessages loop", peer
|
||||||
# This can happen during the rlpx connection setup or at any point after.
|
# This can happen during the rlpx connection setup or at any point after.
|
||||||
# Because this code does not know, a disconnect needs to be done.
|
# Because this code does not know, a disconnect needs to be done.
|
||||||
asyncDiscard peer.disconnect(ClientQuitting)
|
await peer.disconnect(ClientQuitting)
|
||||||
return
|
return
|
||||||
|
|
||||||
if msgId == 1: # p2p.disconnect
|
if msgId == 1: # p2p.disconnect
|
||||||
|
@ -525,8 +531,9 @@ proc dispatchMessages*(peer: Peer) {.async.} =
|
||||||
try:
|
try:
|
||||||
await peer.invokeThunk(msgId, msgData)
|
await peer.invokeThunk(msgId, msgData)
|
||||||
except RlpError:
|
except RlpError:
|
||||||
debug "ending dispatchMessages loop", peer, err = getCurrentExceptionMsg()
|
debug "RlpError, ending dispatchMessages loop", peer,
|
||||||
await peer.disconnect(BreachOfProtocol)
|
err = getCurrentExceptionMsg()
|
||||||
|
await peer.disconnect(BreachOfProtocol, true)
|
||||||
return
|
return
|
||||||
except CatchableError:
|
except CatchableError:
|
||||||
warn "Error while handling RLPx message", peer,
|
warn "Error while handling RLPx message", peer,
|
||||||
|
@ -545,8 +552,10 @@ proc dispatchMessages*(peer: Peer) {.async.} =
|
||||||
# TODO: Handling errors here must be investigated more carefully.
|
# TODO: Handling errors here must be investigated more carefully.
|
||||||
# They also are supposed to be handled at the call-site where
|
# They also are supposed to be handled at the call-site where
|
||||||
# `nextMsg` is used.
|
# `nextMsg` is used.
|
||||||
debug "nextMsg resolver failed", err = getCurrentExceptionMsg()
|
debug "nextMsg resolver failed, ending dispatchMessages loop", peer,
|
||||||
raise
|
err = getCurrentExceptionMsg()
|
||||||
|
await peer.disconnect(BreachOfProtocol, true)
|
||||||
|
return
|
||||||
peer.awaitedMessages[msgId] = nil
|
peer.awaitedMessages[msgId] = nil
|
||||||
|
|
||||||
macro p2pProtocolImpl(name: static[string],
|
macro p2pProtocolImpl(name: static[string],
|
||||||
|
@ -1167,16 +1176,16 @@ proc disconnect*(peer: Peer, reason: DisconnectionReason, notifyOtherPeer = fals
|
||||||
yield fut
|
yield fut
|
||||||
if fut.failed:
|
if fut.failed:
|
||||||
debug "Failed to delived disconnect message", peer
|
debug "Failed to delived disconnect message", peer
|
||||||
try:
|
|
||||||
if not peer.dispatcher.isNil:
|
if not peer.dispatcher.isNil:
|
||||||
await callDisconnectHandlers(peer, reason)
|
# In case of `CatchableError` in any of the handlers, this will be logged.
|
||||||
except:
|
# Other handlers will still execute.
|
||||||
error "Exception in callDisconnectHandlers()",
|
# In case of `Defect` in any of the handlers, program will quit.
|
||||||
err = getCurrentExceptionMsg()
|
traceAwaitErrors callDisconnectHandlers(peer, reason)
|
||||||
finally:
|
|
||||||
logDisconnectedPeer peer
|
logDisconnectedPeer peer
|
||||||
peer.connectionState = Disconnected
|
peer.connectionState = Disconnected
|
||||||
removePeer(peer.network, peer)
|
removePeer(peer.network, peer)
|
||||||
|
|
||||||
proc validatePubKeyInHello(msg: devp2p.hello, pubKey: PublicKey): bool =
|
proc validatePubKeyInHello(msg: devp2p.hello, pubKey: PublicKey): bool =
|
||||||
var pk: PublicKey
|
var pk: PublicKey
|
||||||
|
@ -1237,7 +1246,7 @@ proc postHelloSteps(peer: Peer, h: devp2p.hello) {.async.} =
|
||||||
if messageProcessingLoop.failed:
|
if messageProcessingLoop.failed:
|
||||||
debug "Ending dispatchMessages loop", peer,
|
debug "Ending dispatchMessages loop", peer,
|
||||||
err = messageProcessingLoop.error.msg
|
err = messageProcessingLoop.error.msg
|
||||||
asyncDiscard peer.disconnect(ClientQuitting)
|
traceAsyncErrors peer.disconnect(ClientQuitting)
|
||||||
|
|
||||||
# The handshake may involve multiple async steps, so we wait
|
# The handshake may involve multiple async steps, so we wait
|
||||||
# here for all of them to finish.
|
# here for all of them to finish.
|
||||||
|
|
|
@ -0,0 +1,34 @@
|
||||||
|
import
|
||||||
|
unittest, chronos, eth/[keys, p2p], eth/p2p/[discovery, enode]
|
||||||
|
|
||||||
|
var nextPort = 30303
|
||||||
|
|
||||||
|
proc localAddress(port: int): Address =
|
||||||
|
let port = Port(port)
|
||||||
|
result = Address(udpPort: port, tcpPort: port, ip: parseIpAddress("127.0.0.1"))
|
||||||
|
|
||||||
|
proc startDiscoveryNode(privKey: PrivateKey, address: Address,
|
||||||
|
bootnodes: seq[ENode]): Future[DiscoveryProtocol] {.async.} =
|
||||||
|
result = newDiscoveryProtocol(privKey, address, bootnodes)
|
||||||
|
result.open()
|
||||||
|
await result.bootstrap()
|
||||||
|
|
||||||
|
proc setupBootNode*(): Future[ENode] {.async.} =
|
||||||
|
let
|
||||||
|
bootNodeKey = newPrivateKey()
|
||||||
|
bootNodeAddr = localAddress(30301)
|
||||||
|
bootNode = await startDiscoveryNode(bootNodeKey, bootNodeAddr, @[])
|
||||||
|
result = initENode(bootNodeKey.getPublicKey, bootNodeAddr)
|
||||||
|
|
||||||
|
proc setupTestNode*(capabilities: varargs[ProtocolInfo, `protocolInfo`]): EthereumNode =
|
||||||
|
let keys1 = newKeyPair()
|
||||||
|
result = newEthereumNode(keys1, localAddress(nextPort), 1, nil,
|
||||||
|
addAllCapabilities = false)
|
||||||
|
nextPort.inc
|
||||||
|
for capability in capabilities:
|
||||||
|
result.addCapability capability
|
||||||
|
|
||||||
|
template asyncTest*(name, body: untyped) =
|
||||||
|
test name:
|
||||||
|
proc scenario {.async.} = body
|
||||||
|
waitFor scenario()
|
|
@ -8,31 +8,8 @@
|
||||||
# MIT license (LICENSE-MIT)
|
# MIT license (LICENSE-MIT)
|
||||||
|
|
||||||
import
|
import
|
||||||
unittest, tables, chronos, eth/[keys, p2p], eth/p2p/[discovery, enode]
|
unittest, tables, chronos, eth/p2p,
|
||||||
|
./p2p_test_helper
|
||||||
var nextPort = 30303
|
|
||||||
|
|
||||||
proc localAddress(port: int): Address =
|
|
||||||
let port = Port(port)
|
|
||||||
result = Address(udpPort: port, tcpPort: port, ip: parseIpAddress("127.0.0.1"))
|
|
||||||
|
|
||||||
proc startDiscoveryNode(privKey: PrivateKey, address: Address,
|
|
||||||
bootnodes: seq[ENode]): Future[DiscoveryProtocol] {.async.} =
|
|
||||||
result = newDiscoveryProtocol(privKey, address, bootnodes)
|
|
||||||
result.open()
|
|
||||||
await result.bootstrap()
|
|
||||||
|
|
||||||
proc setupBootNode(): Future[ENode] {.async.} =
|
|
||||||
let
|
|
||||||
bootNodeKey = newPrivateKey()
|
|
||||||
bootNodeAddr = localAddress(30301)
|
|
||||||
bootNode = await startDiscoveryNode(bootNodeKey, bootNodeAddr, @[])
|
|
||||||
result = initENode(bootNodeKey.getPublicKey, bootNodeAddr)
|
|
||||||
|
|
||||||
template asyncTest(name, body: untyped) =
|
|
||||||
test name:
|
|
||||||
proc scenario {.async.} = body
|
|
||||||
waitFor scenario()
|
|
||||||
|
|
||||||
type
|
type
|
||||||
network = ref object
|
network = ref object
|
||||||
|
@ -48,7 +25,7 @@ p2pProtocol abc(version = 1,
|
||||||
onPeerDisconnected do (peer: Peer, reason: DisconnectionReason) {.gcsafe.}:
|
onPeerDisconnected do (peer: Peer, reason: DisconnectionReason) {.gcsafe.}:
|
||||||
peer.networkState.count -= 1
|
peer.networkState.count -= 1
|
||||||
if true:
|
if true:
|
||||||
raise newException(UnsupportedProtocol, "Fake abc exception")
|
raise newException(CatchableError, "Fake abc exception")
|
||||||
|
|
||||||
p2pProtocol xyz(version = 1,
|
p2pProtocol xyz(version = 1,
|
||||||
shortName = "xyz",
|
shortName = "xyz",
|
||||||
|
@ -60,21 +37,13 @@ p2pProtocol xyz(version = 1,
|
||||||
onPeerDisconnected do (peer: Peer, reason: DisconnectionReason) {.gcsafe.}:
|
onPeerDisconnected do (peer: Peer, reason: DisconnectionReason) {.gcsafe.}:
|
||||||
peer.networkState.count -= 1
|
peer.networkState.count -= 1
|
||||||
if true:
|
if true:
|
||||||
raise newException(UnsupportedProtocol, "Fake xyz exception")
|
raise newException(CatchableError, "Fake xyz exception")
|
||||||
|
|
||||||
proc prepTestNode(): EthereumNode =
|
suite "Testing protocol handlers":
|
||||||
let keys1 = newKeyPair()
|
|
||||||
result = newEthereumNode(keys1, localAddress(nextPort), 1, nil,
|
|
||||||
addAllCapabilities = false)
|
|
||||||
nextPort.inc
|
|
||||||
result.addCapability abc
|
|
||||||
result.addCapability xyz
|
|
||||||
|
|
||||||
suite "Failing handlers":
|
|
||||||
asyncTest "Failing disconnect handler":
|
asyncTest "Failing disconnect handler":
|
||||||
let bootENode = waitFor setupBootNode()
|
let bootENode = waitFor setupBootNode()
|
||||||
var node1 = prepTestNode()
|
var node1 = setupTestNode(abc, xyz)
|
||||||
var node2 = prepTestNode()
|
var node2 = setupTestNode(abc, xyz)
|
||||||
# node2 listening and node1 not, to avoid many incoming vs outgoing
|
# node2 listening and node1 not, to avoid many incoming vs outgoing
|
||||||
var node1Connected = node1.connectToNetwork(@[bootENode], false, true)
|
var node1Connected = node1.connectToNetwork(@[bootENode], false, true)
|
||||||
var node2Connected = node2.connectToNetwork(@[bootENode], true, true)
|
var node2Connected = node2.connectToNetwork(@[bootENode], true, true)
|
||||||
|
|
|
@ -8,52 +8,18 @@
|
||||||
# MIT license (LICENSE-MIT)
|
# MIT license (LICENSE-MIT)
|
||||||
|
|
||||||
import
|
import
|
||||||
sequtils, options, unittest, tables, chronos, eth/[rlp, keys, p2p],
|
sequtils, options, unittest, tables, chronos, eth/[keys, p2p],
|
||||||
eth/p2p/rlpx_protocols/[whisper_protocol], eth/p2p/[discovery, enode]
|
eth/p2p/rlpx_protocols/[whisper_protocol],
|
||||||
|
./p2p_test_helper
|
||||||
const
|
|
||||||
useCompression = defined(useSnappy)
|
|
||||||
|
|
||||||
var nextPort = 30303
|
|
||||||
|
|
||||||
proc localAddress(port: int): Address =
|
|
||||||
let port = Port(port)
|
|
||||||
result = Address(udpPort: port, tcpPort: port, ip: parseIpAddress("127.0.0.1"))
|
|
||||||
|
|
||||||
proc startDiscoveryNode(privKey: PrivateKey, address: Address,
|
|
||||||
bootnodes: seq[ENode]): Future[DiscoveryProtocol] {.async.} =
|
|
||||||
result = newDiscoveryProtocol(privKey, address, bootnodes)
|
|
||||||
result.open()
|
|
||||||
await result.bootstrap()
|
|
||||||
|
|
||||||
proc setupBootNode(): Future[ENode] {.async.} =
|
|
||||||
let
|
|
||||||
bootNodeKey = newPrivateKey()
|
|
||||||
bootNodeAddr = localAddress(30301)
|
|
||||||
bootNode = await startDiscoveryNode(bootNodeKey, bootNodeAddr, @[])
|
|
||||||
result = initENode(bootNodeKey.getPublicKey, bootNodeAddr)
|
|
||||||
|
|
||||||
template asyncTest(name, body: untyped) =
|
|
||||||
test name:
|
|
||||||
proc scenario {.async.} = body
|
|
||||||
waitFor scenario()
|
|
||||||
|
|
||||||
proc resetMessageQueues(nodes: varargs[EthereumNode]) =
|
proc resetMessageQueues(nodes: varargs[EthereumNode]) =
|
||||||
for node in nodes:
|
for node in nodes:
|
||||||
node.resetMessageQueue()
|
node.resetMessageQueue()
|
||||||
|
|
||||||
proc prepTestNode(): EthereumNode =
|
|
||||||
let keys1 = newKeyPair()
|
|
||||||
result = newEthereumNode(keys1, localAddress(nextPort), 1, nil,
|
|
||||||
addAllCapabilities = false,
|
|
||||||
useCompression = useCompression)
|
|
||||||
nextPort.inc
|
|
||||||
result.addCapability Whisper
|
|
||||||
|
|
||||||
let bootENode = waitFor setupBootNode()
|
let bootENode = waitFor setupBootNode()
|
||||||
|
|
||||||
var node1 = prepTestNode()
|
var node1 = setupTestNode(Whisper)
|
||||||
var node2 = prepTestNode()
|
var node2 = setupTestNode(Whisper)
|
||||||
# node2 listening and node1 not, to avoid many incoming vs outgoing
|
# node2 listening and node1 not, to avoid many incoming vs outgoing
|
||||||
var node1Connected = node1.connectToNetwork(@[bootENode], false, true)
|
var node1Connected = node1.connectToNetwork(@[bootENode], false, true)
|
||||||
var node2Connected = node2.connectToNetwork(@[bootENode], true, true)
|
var node2Connected = node2.connectToNetwork(@[bootENode], true, true)
|
||||||
|
@ -352,7 +318,7 @@ suite "Whisper connections":
|
||||||
node1.unsubscribeFilter(filter) == true
|
node1.unsubscribeFilter(filter) == true
|
||||||
|
|
||||||
test "Light node posting":
|
test "Light node posting":
|
||||||
var ln1 = prepTestNode()
|
var ln1 = setupTestNode(Whisper)
|
||||||
ln1.setLightNode(true)
|
ln1.setLightNode(true)
|
||||||
|
|
||||||
# not listening, so will only connect to others that are listening (node2)
|
# not listening, so will only connect to others that are listening (node2)
|
||||||
|
@ -373,8 +339,8 @@ suite "Whisper connections":
|
||||||
ln1.protocolState(Whisper).queue.items.len == 0
|
ln1.protocolState(Whisper).queue.items.len == 0
|
||||||
|
|
||||||
test "Connect two light nodes":
|
test "Connect two light nodes":
|
||||||
var ln1 = prepTestNode()
|
var ln1 = setupTestNode(Whisper)
|
||||||
var ln2 = prepTestNode()
|
var ln2 = setupTestNode(Whisper)
|
||||||
|
|
||||||
ln1.setLightNode(true)
|
ln1.setLightNode(true)
|
||||||
ln2.setLightNode(true)
|
ln2.setLightNode(true)
|
||||||
|
|
Loading…
Reference in New Issue