mirror of
https://github.com/logos-messaging/logos-delivery-interop-tests.git
synced 2026-04-17 14:33:49 +00:00
147 lines
6.8 KiB
Python
147 lines
6.8 KiB
Python
import pytest
|
|
from tenacity import retry, stop_after_delay, wait_fixed
|
|
|
|
from src.env_vars import NODE_1, NODE_2
|
|
from src.libs.common import delay
|
|
from src.libs.custom_logger import get_custom_logger
|
|
from src.node.waku_node import peer_info2id, peer_info2multiaddr, multiaddr2id
|
|
from src.steps.relay import StepsRelay
|
|
from src.steps.store import StepsStore
|
|
from src.test_data import VALID_PUBSUB_TOPICS
|
|
|
|
logger = get_custom_logger(__name__)
|
|
|
|
|
|
class TestPeerStore(StepsRelay, StepsStore):
|
|
def test_get_peers(self):
|
|
self.setup_main_nodes()
|
|
self.setup_optional_nodes()
|
|
self.ensure_relay_subscriptions_on_nodes(self.main_nodes + self.optional_nodes, VALID_PUBSUB_TOPICS)
|
|
nodes = [self.node1, self.node2]
|
|
nodes.extend(self.optional_nodes)
|
|
delay(10)
|
|
ids = []
|
|
for i, node in enumerate(nodes):
|
|
node_id = node.get_id()
|
|
logger.debug(f"Node {i} peer ID {node_id}")
|
|
ids.append(node_id)
|
|
|
|
# Discv5 discovery is eventually-consistent; poll until every node's peer
|
|
# store reflects the expected mesh, or fail after the timeout.
|
|
@retry(stop=stop_after_delay(60), wait=wait_fixed(2), reraise=True)
|
|
def check_peer_stores():
|
|
for i in range(5):
|
|
others = []
|
|
for peer_info in nodes[i].get_peers():
|
|
logger.debug(f"Node {i} peer info {peer_info}")
|
|
peer_id = peer_info2id(peer_info, nodes[i].is_nwaku())
|
|
others.append(peer_id)
|
|
|
|
assert (i == 0 and len(others) == 4) or (i > 0 and len(others) >= 1), f"Some nodes missing in the peer store of Node ID {ids[i]}"
|
|
|
|
check_peer_stores()
|
|
|
|
def test_add_peers(self):
|
|
self.setup_main_nodes()
|
|
self.setup_optional_nodes()
|
|
self.ensure_relay_subscriptions_on_nodes(self.main_nodes + self.optional_nodes, VALID_PUBSUB_TOPICS)
|
|
nodes = [self.node1, self.node2]
|
|
nodes.extend(self.optional_nodes)
|
|
delay(10)
|
|
|
|
# Discv5 discovery is eventually-consistent; poll node1 and node2's peer
|
|
# stores until the union contains >=5 multiaddrs, or fail after the timeout.
|
|
@retry(stop=stop_after_delay(60), wait=wait_fixed(2), reraise=True)
|
|
def collect_multiaddrs():
|
|
peers_multiaddr = set()
|
|
for i in range(2):
|
|
for peer_info in nodes[i].get_peers():
|
|
multiaddr = peer_info2multiaddr(peer_info, nodes[i].is_nwaku())
|
|
peers_multiaddr.add(multiaddr)
|
|
|
|
assert len(peers_multiaddr) >= 5, f"At least 5 multi addresses are expected, got {len(peers_multiaddr)}: {peers_multiaddr}"
|
|
return peers_multiaddr
|
|
|
|
peers_multiaddr = collect_multiaddrs()
|
|
|
|
# Group multiaddrs by peer ID. libp2p identify can leak observed (ephemeral
|
|
# source) addresses into the peer store alongside the real listen address;
|
|
# those observed addrs are unreachable when dialed back, so we try every
|
|
# known address for a peer until one succeeds.
|
|
addrs_by_peer = {}
|
|
for peer in peers_multiaddr:
|
|
addrs_by_peer.setdefault(multiaddr2id(peer), []).append(peer)
|
|
|
|
# For each of nodes 2-5, add every other peer via the add_peers API.
|
|
for i in range(1, 5):
|
|
self_id = nodes[i].get_id()
|
|
for peer_id, addrs in addrs_by_peer.items():
|
|
if peer_id == self_id:
|
|
continue
|
|
last_err = None
|
|
for addr in addrs:
|
|
try:
|
|
if nodes[i].is_nwaku():
|
|
nodes[i].add_peers([addr])
|
|
else:
|
|
peer_info = {"multiaddr": addr, "shards": [0], "protocols": ["/vac/waku/relay/2.0.0"]}
|
|
nodes[i].add_peers(peer_info)
|
|
last_err = None
|
|
break
|
|
except Exception as ex:
|
|
last_err = ex
|
|
logger.warning(f"Node {i} failed to add peer {peer_id} via {addr}: {ex}")
|
|
if last_err is not None:
|
|
logger.error(f"Node {i} could not add peer {peer_id} via any known address")
|
|
raise last_err
|
|
|
|
@pytest.mark.skip(reason="waiting for https://github.com/waku-org/nwaku/issues/1549 resolution")
|
|
def test_get_peers_two_protocols(self):
|
|
self.setup_first_publishing_node(store="true", relay="true")
|
|
self.setup_first_store_node(store="true", relay="false")
|
|
self.ensure_relay_subscriptions_on_nodes([self.publishing_node1, self.store_node1], VALID_PUBSUB_TOPICS)
|
|
delay(1)
|
|
node1_peers = self.publishing_node1.get_peers()
|
|
node2_peers = self.store_node1.get_peers()
|
|
logger.debug(f"Node 1 connected peers {node1_peers}")
|
|
logger.debug(f"Node 2 connected peers {node2_peers}")
|
|
|
|
assert len(node1_peers) == 2 and len(node2_peers) == 2, f"Some nodes and/or their services are missing"
|
|
|
|
@pytest.mark.skip(reason="pending on https://github.com/waku-org/nwaku/issues/2792")
|
|
def test_use_persistent_storage_survive_restart(self):
|
|
self.setup_first_relay_node(peer_persistence="true")
|
|
self.setup_second_relay_node()
|
|
self.ensure_relay_subscriptions_on_nodes(self.main_nodes, VALID_PUBSUB_TOPICS)
|
|
delay(1)
|
|
node1_peers = self.node1.get_peers()
|
|
node2_peers = self.node2.get_peers()
|
|
node1_id = self.node1.get_id()
|
|
node2_id = self.node2.get_id()
|
|
assert node1_id == peer_info2id(node2_peers[0], self.node2.is_nwaku())
|
|
assert node2_id == peer_info2id(node1_peers[0], self.node1.is_nwaku())
|
|
|
|
# Node 3 takes over Node 1
|
|
self.setup_third_relay_node(peer_persistence="true")
|
|
self.node1.kill()
|
|
node2_peers = self.node2.get_peers()
|
|
node3_peers = self.node3.get_peers()
|
|
assert node1_id == peer_info2id(node2_peers[0], self.node2.is_nwaku())
|
|
assert node2_id == peer_info2id(node3_peers[0], self.node3.is_nwaku())
|
|
|
|
@pytest.mark.skip(reason="waiting for https://github.com/waku-org/nwaku/issues/2592 resolution")
|
|
def test_peer_store_content_after_node2_restarts(self):
|
|
self.setup_first_relay_node()
|
|
self.setup_second_relay_node()
|
|
self.ensure_relay_subscriptions_on_nodes(self.main_nodes, VALID_PUBSUB_TOPICS)
|
|
delay(1)
|
|
node1_peers = self.node1.get_peers()
|
|
node2_peers = self.node2.get_peers()
|
|
assert len(node1_peers) == len(node2_peers), "Nodes should have each other in the peer store"
|
|
self.node2.restart()
|
|
self.node2.ensure_ready()
|
|
delay(1)
|
|
node1_peers = self.node1.get_peers()
|
|
node2_peers = self.node2.get_peers()
|
|
assert len(node1_peers) == len(node2_peers), "Nodes should have each other in the peer store"
|