mirror of https://github.com/waku-org/nwaku.git
549 lines
23 KiB
Nim
549 lines
23 KiB
Nim
{.used.}
|
||
|
||
import
|
||
std/[options, sequtils],
|
||
stew/shims/net as stewNet,
|
||
testutils/unittests,
|
||
chronicles,
|
||
chronos,
|
||
json_rpc/rpcserver,
|
||
json_rpc/rpcclient,
|
||
eth/keys,
|
||
eth/common/eth_types,
|
||
libp2p/[builders, switch, multiaddress],
|
||
libp2p/protobuf/minprotobuf,
|
||
libp2p/stream/[bufferstream, connection],
|
||
libp2p/crypto/crypto,
|
||
libp2p/protocols/pubsub/pubsub,
|
||
libp2p/protocols/pubsub/rpc/message,
|
||
libp2p/peerid
|
||
import
|
||
../../waku/common/sqlite,
|
||
../../waku/v2/node/peer_manager/peer_manager,
|
||
../../waku/v2/node/peer_manager/peer_store/waku_peer_storage,
|
||
../../waku/v2/node/waku_node,
|
||
../../waku/v2/protocol/waku_relay,
|
||
../../waku/v2/protocol/waku_store,
|
||
../../waku/v2/protocol/waku_filter,
|
||
../../waku/v2/protocol/waku_lightpush,
|
||
../../waku/v2/protocol/waku_peer_exchange,
|
||
../../waku/v2/protocol/waku_swap/waku_swap,
|
||
./testlib/common,
|
||
./testlib/testutils,
|
||
./testlib/waku2
|
||
|
||
procSuite "Peer Manager":
|
||
asyncTest "Peer dialing works":
|
||
# Create 2 nodes
|
||
let nodes = toSeq(0..<2).mapIt(WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0)))
|
||
|
||
await allFutures(nodes.mapIt(it.start()))
|
||
await allFutures(nodes.mapIt(it.mountRelay()))
|
||
|
||
# Dial node2 from node1
|
||
let conn = (await nodes[0].peerManager.dialPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec)).get()
|
||
|
||
# Check connection
|
||
check:
|
||
conn.activity
|
||
conn.peerId == nodes[1].peerInfo.peerId
|
||
|
||
# Check that node2 is being managed in node1
|
||
check:
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[1].peerInfo.peerId)
|
||
|
||
# Check connectedness
|
||
check:
|
||
nodes[0].peerManager.peerStore.connectedness(nodes[1].peerInfo.peerId) == Connectedness.Connected
|
||
|
||
await allFutures(nodes.mapIt(it.stop()))
|
||
|
||
asyncTest "Dialing fails gracefully":
|
||
# Create 2 nodes
|
||
let nodes = toSeq(0..<2).mapIt(WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0)))
|
||
|
||
await nodes[0].start()
|
||
await nodes[0].mountRelay()
|
||
|
||
# Purposefully don't start node2
|
||
|
||
# Dial node2 from node1
|
||
let connOpt = await nodes[1].peerManager.dialPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec, 2.seconds)
|
||
|
||
# Check connection failed gracefully
|
||
check:
|
||
connOpt.isNone()
|
||
|
||
await nodes[0].stop()
|
||
|
||
asyncTest "Adding, selecting and filtering peers work":
|
||
let
|
||
node = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0))
|
||
|
||
# Create filter peer
|
||
filterLoc = MultiAddress.init("/ip4/127.0.0.1/tcp/0").tryGet()
|
||
filterPeer = PeerInfo.new(generateEcdsaKey(), @[filterLoc])
|
||
# Create swap peer
|
||
swapLoc = MultiAddress.init("/ip4/127.0.0.2/tcp/2").tryGet()
|
||
swapPeer = PeerInfo.new(generateEcdsaKey(), @[swapLoc])
|
||
# Create store peer
|
||
storeLoc = MultiAddress.init("/ip4/127.0.0.3/tcp/4").tryGet()
|
||
storePeer = PeerInfo.new(generateEcdsaKey(), @[storeLoc])
|
||
|
||
await node.start()
|
||
|
||
await node.mountFilterClient()
|
||
await node.mountSwap()
|
||
node.mountStoreClient()
|
||
|
||
node.wakuSwap.setPeer(swapPeer.toRemotePeerInfo())
|
||
|
||
node.setStorePeer(storePeer.toRemotePeerInfo())
|
||
node.setFilterPeer(filterPeer.toRemotePeerInfo())
|
||
|
||
# Check peers were successfully added to peer manager
|
||
check:
|
||
node.peerManager.peerStore.peers().len == 3
|
||
node.peerManager.peerStore.peers(WakuFilterCodec).allIt(it.peerId == filterPeer.peerId and
|
||
it.addrs.contains(filterLoc) and
|
||
it.protos.contains(WakuFilterCodec))
|
||
node.peerManager.peerStore.peers(WakuSwapCodec).allIt(it.peerId == swapPeer.peerId and
|
||
it.addrs.contains(swapLoc) and
|
||
it.protos.contains(WakuSwapCodec))
|
||
node.peerManager.peerStore.peers(WakuStoreCodec).allIt(it.peerId == storePeer.peerId and
|
||
it.addrs.contains(storeLoc) and
|
||
it.protos.contains(WakuStoreCodec))
|
||
|
||
await node.stop()
|
||
|
||
|
||
asyncTest "Peer manager keeps track of connections":
|
||
# Create 2 nodes
|
||
let nodes = toSeq(0..<2).mapIt(WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0)))
|
||
|
||
await nodes[0].start()
|
||
# Do not start node2
|
||
|
||
await allFutures(nodes.mapIt(it.mountRelay()))
|
||
|
||
# Test default connectedness for new peers
|
||
nodes[0].peerManager.addPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec)
|
||
check:
|
||
# No information about node2's connectedness
|
||
nodes[0].peerManager.peerStore.connectedness(nodes[1].peerInfo.peerId) == NotConnected
|
||
|
||
# Purposefully don't start node2
|
||
# Attempt dialing node2 from node1
|
||
discard await nodes[0].peerManager.dialPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec, 2.seconds)
|
||
check:
|
||
# Cannot connect to node2
|
||
nodes[0].peerManager.peerStore.connectedness(nodes[1].peerInfo.peerId) == CannotConnect
|
||
|
||
# Successful connection
|
||
await nodes[1].start()
|
||
discard await nodes[0].peerManager.dialPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec, 2.seconds)
|
||
check:
|
||
# Currently connected to node2
|
||
nodes[0].peerManager.peerStore.connectedness(nodes[1].peerInfo.peerId) == Connected
|
||
|
||
# Stop node. Gracefully disconnect from all peers.
|
||
await nodes[0].stop()
|
||
check:
|
||
# Not currently connected to node2, but had recent, successful connection.
|
||
nodes[0].peerManager.peerStore.connectedness(nodes[1].peerInfo.peerId) == CanConnect
|
||
|
||
await nodes[1].stop()
|
||
|
||
asyncTest "Peer manager updates failed peers correctly":
|
||
# Create 2 nodes
|
||
let nodes = toSeq(0..<2).mapIt(WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0)))
|
||
|
||
await nodes[0].start()
|
||
await nodes[0].mountRelay()
|
||
nodes[0].peerManager.addPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec)
|
||
|
||
# Set a low backoff to speed up test: 2, 4, 8, 16
|
||
nodes[0].peerManager.initialBackoffInSec = 2
|
||
nodes[0].peerManager.backoffFactor = 2
|
||
|
||
# node2 is not started, so dialing will fail
|
||
let conn1 = await nodes[0].peerManager.dialPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec, 1.seconds)
|
||
check:
|
||
# Cannot connect to node2
|
||
nodes[0].peerManager.peerStore.connectedness(nodes[1].peerInfo.peerId) == CannotConnect
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[1].peerInfo.peerId] == CannotConnect
|
||
nodes[0].peerManager.peerStore[NumberFailedConnBook][nodes[1].peerInfo.peerId] == 1
|
||
|
||
# And the connection failed
|
||
conn1.isNone() == true
|
||
|
||
# Right after failing there is a backoff period
|
||
nodes[0].peerManager.peerStore.canBeConnected(
|
||
nodes[1].peerInfo.peerId,
|
||
nodes[0].peerManager.initialBackoffInSec,
|
||
nodes[0].peerManager.backoffFactor) == false
|
||
|
||
# We wait the first backoff period
|
||
await sleepAsync(2100.milliseconds)
|
||
|
||
# And backoff period is over
|
||
check:
|
||
nodes[0].peerManager.peerStore.canBeConnected(
|
||
nodes[1].peerInfo.peerId,
|
||
nodes[0].peerManager.initialBackoffInSec,
|
||
nodes[0].peerManager.backoffFactor) == true
|
||
|
||
await nodes[1].start()
|
||
await nodes[1].mountRelay()
|
||
|
||
# Now we can connect and failed count is reset
|
||
let conn2 = await nodes[0].peerManager.dialPeer(nodes[1].peerInfo.toRemotePeerInfo(), WakuRelayCodec, 1.seconds)
|
||
check:
|
||
conn2.isNone() == false
|
||
nodes[0].peerManager.peerStore[NumberFailedConnBook][nodes[1].peerInfo.peerId] == 0
|
||
|
||
await allFutures(nodes.mapIt(it.stop()))
|
||
|
||
asyncTest "Peer manager can use persistent storage and survive restarts":
|
||
let
|
||
database = SqliteDatabase.new(":memory:")[]
|
||
storage = WakuPeerStorage.new(database)[]
|
||
node1 = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0), peerStorage = storage)
|
||
node2 = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0))
|
||
peerInfo2 = node2.switch.peerInfo
|
||
|
||
await node1.start()
|
||
await node2.start()
|
||
|
||
await node1.mountRelay()
|
||
await node2.mountRelay()
|
||
|
||
discard await node1.peerManager.dialPeer(peerInfo2.toRemotePeerInfo(), WakuRelayCodec, 2.seconds)
|
||
check:
|
||
# Currently connected to node2
|
||
node1.peerManager.peerStore.peers().len == 1
|
||
node1.peerManager.peerStore.peers().anyIt(it.peerId == peerInfo2.peerId)
|
||
node1.peerManager.peerStore.connectedness(peerInfo2.peerId) == Connected
|
||
|
||
# Simulate restart by initialising a new node using the same storage
|
||
let
|
||
node3 = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0), peerStorage = storage)
|
||
|
||
await node3.start()
|
||
check:
|
||
# Node2 has been loaded after "restart", but we have not yet reconnected
|
||
node3.peerManager.peerStore.peers().len == 1
|
||
node3.peerManager.peerStore.peers().anyIt(it.peerId == peerInfo2.peerId)
|
||
node3.peerManager.peerStore.connectedness(peerInfo2.peerId) == NotConnected
|
||
|
||
await node3.mountRelay() # This should trigger a reconnect
|
||
|
||
check:
|
||
# Reconnected to node2 after "restart"
|
||
node3.peerManager.peerStore.peers().len == 1
|
||
node3.peerManager.peerStore.peers().anyIt(it.peerId == peerInfo2.peerId)
|
||
node3.peerManager.peerStore.connectedness(peerInfo2.peerId) == Connected
|
||
|
||
await allFutures([node1.stop(), node2.stop(), node3.stop()])
|
||
|
||
# TODO: nwaku/issues/1377
|
||
xasyncTest "Peer manager support multiple protocol IDs when reconnecting to peers":
|
||
let
|
||
database = SqliteDatabase.new(":memory:")[]
|
||
storage = WakuPeerStorage.new(database)[]
|
||
node1 = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0), peerStorage = storage)
|
||
node2 = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0))
|
||
peerInfo2 = node2.switch.peerInfo
|
||
betaCodec = "/vac/waku/relay/2.0.0-beta2"
|
||
stableCodec = "/vac/waku/relay/2.0.0"
|
||
|
||
await node1.start()
|
||
await node2.start()
|
||
|
||
await node1.mountRelay()
|
||
node1.wakuRelay.codec = betaCodec
|
||
await node2.mountRelay()
|
||
node2.wakuRelay.codec = betaCodec
|
||
|
||
discard await node1.peerManager.dialPeer(peerInfo2.toRemotePeerInfo(), node2.wakuRelay.codec, 2.seconds)
|
||
check:
|
||
# Currently connected to node2
|
||
node1.peerManager.peerStore.peers().len == 1
|
||
node1.peerManager.peerStore.peers().anyIt(it.peerId == peerInfo2.peerId)
|
||
node1.peerManager.peerStore.peers().anyIt(it.protos.contains(node2.wakuRelay.codec))
|
||
node1.peerManager.peerStore.connectedness(peerInfo2.peerId) == Connected
|
||
|
||
# Simulate restart by initialising a new node using the same storage
|
||
let
|
||
node3 = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0), peerStorage = storage)
|
||
|
||
await node3.mountRelay()
|
||
node3.wakuRelay.codec = stableCodec
|
||
check:
|
||
# Node 2 and 3 have differing codecs
|
||
node2.wakuRelay.codec == betaCodec
|
||
node3.wakuRelay.codec == stableCodec
|
||
# Node2 has been loaded after "restart", but we have not yet reconnected
|
||
node3.peerManager.peerStore.peers().len == 1
|
||
node3.peerManager.peerStore.peers().anyIt(it.peerId == peerInfo2.peerId)
|
||
node3.peerManager.peerStore.peers().anyIt(it.protos.contains(betaCodec))
|
||
node3.peerManager.peerStore.connectedness(peerInfo2.peerId) == NotConnected
|
||
|
||
await node3.start() # This should trigger a reconnect
|
||
|
||
check:
|
||
# Reconnected to node2 after "restart"
|
||
node3.peerManager.peerStore.peers().len == 1
|
||
node3.peerManager.peerStore.peers().anyIt(it.peerId == peerInfo2.peerId)
|
||
node3.peerManager.peerStore.peers().anyIt(it.protos.contains(betaCodec))
|
||
node3.peerManager.peerStore.peers().anyIt(it.protos.contains(stableCodec))
|
||
node3.peerManager.peerStore.connectedness(peerInfo2.peerId) == Connected
|
||
|
||
await allFutures([node1.stop(), node2.stop(), node3.stop()])
|
||
|
||
asyncTest "Peer manager connects to all peers supporting a given protocol":
|
||
# Create 4 nodes
|
||
let nodes = toSeq(0..<4).mapIt(WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0)))
|
||
|
||
# Start them
|
||
await allFutures(nodes.mapIt(it.start()))
|
||
await allFutures(nodes.mapIt(it.mountRelay()))
|
||
|
||
# Get all peer infos
|
||
let peerInfos = nodes.mapIt(it.switch.peerInfo.toRemotePeerInfo())
|
||
|
||
# Add all peers (but self) to node 0
|
||
nodes[0].peerManager.addPeer(peerInfos[1], WakuRelayCodec)
|
||
nodes[0].peerManager.addPeer(peerInfos[2], WakuRelayCodec)
|
||
nodes[0].peerManager.addPeer(peerInfos[3], WakuRelayCodec)
|
||
|
||
# Attempt to connect to all known peers supporting a given protocol
|
||
await nodes[0].peerManager.reconnectPeers(WakuRelayCodec, protocolMatcher(WakuRelayCodec))
|
||
|
||
check:
|
||
# Peerstore track all three peers
|
||
nodes[0].peerManager.peerStore.peers().len == 3
|
||
|
||
# All peer ids are correct
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[1].switch.peerInfo.peerId)
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[2].switch.peerInfo.peerId)
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[3].switch.peerInfo.peerId)
|
||
|
||
# All peers support the relay protocol
|
||
nodes[0].peerManager.peerStore[ProtoBook][nodes[1].switch.peerInfo.peerId].contains(WakuRelayCodec)
|
||
nodes[0].peerManager.peerStore[ProtoBook][nodes[2].switch.peerInfo.peerId].contains(WakuRelayCodec)
|
||
nodes[0].peerManager.peerStore[ProtoBook][nodes[3].switch.peerInfo.peerId].contains(WakuRelayCodec)
|
||
|
||
# All peers are connected
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[1].switch.peerInfo.peerId] == Connected
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[2].switch.peerInfo.peerId] == Connected
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[3].switch.peerInfo.peerId] == Connected
|
||
|
||
await allFutures(nodes.mapIt(it.stop()))
|
||
|
||
asyncTest "Peer store keeps track of incoming connections":
|
||
# Create 4 nodes
|
||
let nodes = toSeq(0..<4).mapIt(WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0)))
|
||
|
||
# Start them
|
||
await allFutures(nodes.mapIt(it.start()))
|
||
await allFutures(nodes.mapIt(it.mountRelay()))
|
||
|
||
# Get all peer infos
|
||
let peerInfos = nodes.mapIt(it.switch.peerInfo.toRemotePeerInfo())
|
||
|
||
# all nodes connect to peer 0
|
||
discard await nodes[1].peerManager.dialPeer(peerInfos[0], WakuRelayCodec, 2.seconds)
|
||
discard await nodes[2].peerManager.dialPeer(peerInfos[0], WakuRelayCodec, 2.seconds)
|
||
discard await nodes[3].peerManager.dialPeer(peerInfos[0], WakuRelayCodec, 2.seconds)
|
||
|
||
check:
|
||
# Peerstore track all three peers
|
||
nodes[0].peerManager.peerStore.peers().len == 3
|
||
|
||
# Inbound/Outbound number of peers match
|
||
nodes[0].peerManager.peerStore.getPeersByDirection(Inbound).len == 3
|
||
nodes[0].peerManager.peerStore.getPeersByDirection(Outbound).len == 0
|
||
nodes[1].peerManager.peerStore.getPeersByDirection(Inbound).len == 0
|
||
nodes[1].peerManager.peerStore.getPeersByDirection(Outbound).len == 1
|
||
nodes[2].peerManager.peerStore.getPeersByDirection(Inbound).len == 0
|
||
nodes[2].peerManager.peerStore.getPeersByDirection(Outbound).len == 1
|
||
nodes[3].peerManager.peerStore.getPeersByDirection(Inbound).len == 0
|
||
nodes[3].peerManager.peerStore.getPeersByDirection(Outbound).len == 1
|
||
|
||
# All peer ids are correct
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[1].switch.peerInfo.peerId)
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[2].switch.peerInfo.peerId)
|
||
nodes[0].peerManager.peerStore.peers().anyIt(it.peerId == nodes[3].switch.peerInfo.peerId)
|
||
|
||
# All peers support the relay protocol
|
||
nodes[0].peerManager.peerStore[ProtoBook][nodes[1].switch.peerInfo.peerId].contains(WakuRelayCodec)
|
||
nodes[0].peerManager.peerStore[ProtoBook][nodes[2].switch.peerInfo.peerId].contains(WakuRelayCodec)
|
||
nodes[0].peerManager.peerStore[ProtoBook][nodes[3].switch.peerInfo.peerId].contains(WakuRelayCodec)
|
||
|
||
# All peers are connected
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[1].switch.peerInfo.peerId] == Connected
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[2].switch.peerInfo.peerId] == Connected
|
||
nodes[0].peerManager.peerStore[ConnectionBook][nodes[3].switch.peerInfo.peerId] == Connected
|
||
|
||
# All peers are Inbound in peer 0
|
||
nodes[0].peerManager.peerStore[DirectionBook][nodes[1].switch.peerInfo.peerId] == Inbound
|
||
nodes[0].peerManager.peerStore[DirectionBook][nodes[2].switch.peerInfo.peerId] == Inbound
|
||
nodes[0].peerManager.peerStore[DirectionBook][nodes[3].switch.peerInfo.peerId] == Inbound
|
||
|
||
# All peers have an Outbound connection with peer 0
|
||
nodes[1].peerManager.peerStore[DirectionBook][nodes[0].switch.peerInfo.peerId] == Outbound
|
||
nodes[2].peerManager.peerStore[DirectionBook][nodes[0].switch.peerInfo.peerId] == Outbound
|
||
nodes[3].peerManager.peerStore[DirectionBook][nodes[0].switch.peerInfo.peerId] == Outbound
|
||
|
||
await allFutures(nodes.mapIt(it.stop()))
|
||
|
||
asyncTest "Peer store addServicePeer() stores service peers":
|
||
# Valid peer id missing the last digit
|
||
let basePeerId = "16Uiu2HAm7QGEZKujdSbbo1aaQyfDPQ6Bw3ybQnj6fruH5Dxwd7D"
|
||
|
||
let
|
||
node = WakuNode.new(generateSecp256k1Key(), ValidIpAddress.init("0.0.0.0"), Port(0))
|
||
peer1 = parseRemotePeerInfo("/ip4/0.0.0.0/tcp/30300/p2p/" & basePeerId & "1")
|
||
peer2 = parseRemotePeerInfo("/ip4/0.0.0.0/tcp/30301/p2p/" & basePeerId & "2")
|
||
peer3 = parseRemotePeerInfo("/ip4/0.0.0.0/tcp/30302/p2p/" & basePeerId & "3")
|
||
peer4 = parseRemotePeerInfo("/ip4/0.0.0.0/tcp/30303/p2p/" & basePeerId & "4")
|
||
peer5 = parseRemotePeerInfo("/ip4/0.0.0.0/tcp/30303/p2p/" & basePeerId & "5")
|
||
|
||
# service peers
|
||
node.peerManager.addServicePeer(peer1, WakuStoreCodec)
|
||
node.peerManager.addServicePeer(peer2, WakuFilterCodec)
|
||
node.peerManager.addServicePeer(peer3, WakuLightPushCodec)
|
||
node.peerManager.addServicePeer(peer4, WakuPeerExchangeCodec)
|
||
|
||
# relay peers (should not be added)
|
||
node.peerManager.addServicePeer(peer5, WakuRelayCodec)
|
||
|
||
# all peers are stored in the peerstore
|
||
check:
|
||
node.peerManager.peerStore.peers().anyIt(it.peerId == peer1.peerId)
|
||
node.peerManager.peerStore.peers().anyIt(it.peerId == peer2.peerId)
|
||
node.peerManager.peerStore.peers().anyIt(it.peerId == peer3.peerId)
|
||
node.peerManager.peerStore.peers().anyIt(it.peerId == peer4.peerId)
|
||
|
||
# but the relay peer is not
|
||
node.peerManager.peerStore.peers().anyIt(it.peerId == peer5.peerId) == false
|
||
|
||
# all service peers are added to its service slot
|
||
check:
|
||
node.peerManager.serviceSlots[WakuStoreCodec].peerId == peer1.peerId
|
||
node.peerManager.serviceSlots[WakuFilterCodec].peerId == peer2.peerId
|
||
node.peerManager.serviceSlots[WakuLightPushCodec].peerId == peer3.peerId
|
||
node.peerManager.serviceSlots[WakuPeerExchangeCodec].peerId == peer4.peerId
|
||
|
||
# but the relay peer is not
|
||
node.peerManager.serviceSlots.hasKey(WakuRelayCodec) == false
|
||
|
||
test "selectPeer() returns the correct peer":
|
||
# Valid peer id missing the last digit
|
||
let basePeerId = "16Uiu2HAm7QGEZKujdSbbo1aaQyfDPQ6Bw3ybQnj6fruH5Dxwd7D"
|
||
|
||
# Create peer manager
|
||
let pm = PeerManager.new(
|
||
switch = SwitchBuilder.new().withRng(rng).withMplex().withNoise().build(),
|
||
storage = nil)
|
||
|
||
# Create 3 peer infos
|
||
let peers = toSeq(1..3).mapIt(parseRemotePeerInfo("/ip4/0.0.0.0/tcp/30300/p2p/" & basePeerId & $it))
|
||
|
||
# Add a peer[0] to the peerstore
|
||
pm.peerStore[AddressBook][peers[0].peerId] = peers[0].addrs
|
||
pm.peerStore[ProtoBook][peers[0].peerId] = @[WakuRelayCodec, WakuStoreCodec, WakuFilterCodec]
|
||
|
||
# When no service peers, we get one from the peerstore
|
||
let selectedPeer1 = pm.selectPeer(WakuStoreCodec)
|
||
check:
|
||
selectedPeer1.isSome() == true
|
||
selectedPeer1.get().peerId == peers[0].peerId
|
||
|
||
# Same for other protocol
|
||
let selectedPeer2 = pm.selectPeer(WakuFilterCodec)
|
||
check:
|
||
selectedPeer2.isSome() == true
|
||
selectedPeer2.get().peerId == peers[0].peerId
|
||
|
||
# And return none if we dont have any peer for that protocol
|
||
let selectedPeer3 = pm.selectPeer(WakuLightPushCodec)
|
||
check:
|
||
selectedPeer3.isSome() == false
|
||
|
||
# Now we add service peers for different protocols peer[1..3]
|
||
pm.addServicePeer(peers[1], WakuStoreCodec)
|
||
pm.addServicePeer(peers[2], WakuLightPushCodec)
|
||
|
||
# We no longer get one from the peerstore. Slots are being used instead.
|
||
let selectedPeer4 = pm.selectPeer(WakuStoreCodec)
|
||
check:
|
||
selectedPeer4.isSome() == true
|
||
selectedPeer4.get().peerId == peers[1].peerId
|
||
|
||
let selectedPeer5 = pm.selectPeer(WakuLightPushCodec)
|
||
check:
|
||
selectedPeer5.isSome() == true
|
||
selectedPeer5.get().peerId == peers[2].peerId
|
||
|
||
test "peer manager cant have more max connections than peerstore size":
|
||
# Peerstore size can't be smaller than max connections
|
||
let peerStoreSize = 5
|
||
let maxConnections = 10
|
||
|
||
expect(Defect):
|
||
let pm = PeerManager.new(
|
||
switch = SwitchBuilder.new().withRng(rng).withMplex().withNoise()
|
||
.withPeerStore(peerStoreSize)
|
||
.withMaxConnections(maxConnections)
|
||
.build(),
|
||
storage = nil)
|
||
|
||
test "prunePeerStore() correctly removes peers to match max quota":
|
||
# Create peer manager
|
||
let pm = PeerManager.new(
|
||
switch = SwitchBuilder.new().withRng(rng).withMplex().withNoise()
|
||
.withPeerStore(10)
|
||
.withMaxConnections(5)
|
||
.build(),
|
||
maxFailedAttempts = 1,
|
||
storage = nil)
|
||
|
||
# Create 15 peers and add them to the peerstore
|
||
let peers = toSeq(1..15).mapIt(parseRemotePeerInfo("/ip4/0.0.0.0/tcp/0/p2p/" & $PeerId.random().get()))
|
||
for p in peers: pm.addPeer(p, "")
|
||
|
||
# Check that we have 15 peers in the peerstore
|
||
check:
|
||
pm.peerStore.peers.len == 15
|
||
|
||
# fake that some peers failed to connected
|
||
pm.peerStore[NumberFailedConnBook][peers[0].peerId] = 2
|
||
pm.peerStore[NumberFailedConnBook][peers[1].peerId] = 2
|
||
pm.peerStore[NumberFailedConnBook][peers[2].peerId] = 2
|
||
|
||
# fake that some peers are connected
|
||
pm.peerStore[ConnectionBook][peers[5].peerId] = Connected
|
||
pm.peerStore[ConnectionBook][peers[8].peerId] = Connected
|
||
pm.peerStore[ConnectionBook][peers[10].peerId] = Connected
|
||
pm.peerStore[ConnectionBook][peers[12].peerId] = Connected
|
||
|
||
# Prune the peerstore
|
||
pm.prunePeerStore()
|
||
|
||
check:
|
||
# ensure peerstore was pruned
|
||
pm.peerStore.peers.len == 10
|
||
|
||
# ensure connected peers were not pruned
|
||
pm.peerStore.peers.anyIt(it.peerId == peers[5].peerId)
|
||
pm.peerStore.peers.anyIt(it.peerId == peers[8].peerId)
|
||
pm.peerStore.peers.anyIt(it.peerId == peers[10].peerId)
|
||
pm.peerStore.peers.anyIt(it.peerId == peers[12].peerId)
|
||
|
||
# ensure peers that failed were the first to be pruned
|
||
not pm.peerStore.peers.anyIt(it.peerId == peers[0].peerId)
|
||
not pm.peerStore.peers.anyIt(it.peerId == peers[1].peerId)
|
||
not pm.peerStore.peers.anyIt(it.peerId == peers[2].peerId)
|