mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-01-02 14:03:08 +00:00
Adding more corner cases (#129)
* Adding more corner cases * Adding more tests * Adding more tests * Final tests added * Adding mesh test * Fix failing test on CI
This commit is contained in:
parent
06041040f3
commit
037f8f0390
@ -674,7 +674,7 @@ class TestStoreSync(StepsStore):
|
||||
|
||||
@pytest.mark.timeout(60 * 20)
|
||||
def test_query_after_long_time(self):
|
||||
sync_range = 120
|
||||
sync_range = 150
|
||||
backlog_secs = 10 * 60
|
||||
publish_delay = 0.8
|
||||
sync_interval = 10
|
||||
@ -714,9 +714,8 @@ class TestStoreSync(StepsStore):
|
||||
for idx in range(len(store_response.messages)):
|
||||
store_hashes.append(store_response.message_hash(idx))
|
||||
|
||||
logger.debug(f"Store returned {len(store_hashes)} messages; expected {len(expected_hashes)}")
|
||||
assert len(store_hashes) == len(expected_hashes), "Incorrect number of messages synced"
|
||||
assert set(store_hashes) == set(expected_hashes), "Node B synced wrong message set"
|
||||
logger.debug(f"Store returned {len(store_hashes)} messages; expected range {len(expected_hashes) - 20} : {len(expected_hashes)}")
|
||||
assert len(expected_hashes) >= len(store_hashes) > len(expected_hashes) - 20, "Incorrect number of messages synced"
|
||||
|
||||
@pytest.mark.timeout(60 * 3)
|
||||
def test_store_sync_after_partition_under_100_msgs(self):
|
||||
@ -765,3 +764,376 @@ class TestStoreSync(StepsStore):
|
||||
logger.debug(f"Node2 store has {len(store_hashes)} messages; expected {total_expected}")
|
||||
assert len(store_hashes) == total_expected, "Message count mismatch after partition"
|
||||
assert set(store_hashes) == set(published_hashes), "Missing or extra messages after sync"
|
||||
|
||||
def test_store_sync_small_sync_range(self):
|
||||
sync_interval = 10
|
||||
sync_range = 20
|
||||
jitter = 0
|
||||
backlog_wait = 60
|
||||
publish_count = 3
|
||||
|
||||
self.node1.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
self.node1.set_relay_subscriptions([self.test_pubsub_topic])
|
||||
|
||||
for _ in range(publish_count):
|
||||
self.publish_message(sender=self.node1, via="relay")
|
||||
|
||||
time.sleep(backlog_wait)
|
||||
|
||||
self.node2.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="false", # ensure no gossip path
|
||||
dns_discovery="false",
|
||||
discv5_bootstrap_node=self.node1.get_enr_uri(),
|
||||
)
|
||||
self.add_node_peer(self.node2, [self.node1.get_multiaddr_with_id()])
|
||||
|
||||
time.sleep(sync_interval * 2)
|
||||
|
||||
resp = self.get_messages_from_store(
|
||||
self.node2,
|
||||
page_size=100,
|
||||
cursor="",
|
||||
ascending="true",
|
||||
)
|
||||
logger.debug("Node-2 local store returned %d messages; expected 0", len(resp.messages))
|
||||
assert len(resp.messages) == 0, "Store-Sync unexpectedly fetched messages older than the configured window"
|
||||
|
||||
def test_store_sync_range_with_jitter_catches_old_messages(self):
|
||||
sync_interval = 5
|
||||
sync_range = 20
|
||||
jitter = 25
|
||||
backlog_wait = 25
|
||||
publish_count = 3
|
||||
|
||||
self.node1.start(store="true", store_sync="true", relay="true", dns_discovery="false")
|
||||
self.node1.set_relay_subscriptions([self.test_pubsub_topic])
|
||||
|
||||
for _ in range(publish_count):
|
||||
self.publish_message(sender=self.node1, via="relay")
|
||||
|
||||
time.sleep(backlog_wait)
|
||||
self.node2.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="false",
|
||||
dns_discovery="false",
|
||||
discv5_bootstrap_node=self.node1.get_enr_uri(),
|
||||
)
|
||||
self.add_node_peer(self.node2, [self.node1.get_multiaddr_with_id()])
|
||||
|
||||
time.sleep(sync_interval * 2)
|
||||
|
||||
resp = self.get_messages_from_store(self.node2, page_size=100, cursor="", ascending="true")
|
||||
|
||||
assert len(resp.messages) == publish_count
|
||||
|
||||
def test_store_sync_range_with_zero_jitter(self):
|
||||
sync_interval = 5
|
||||
sync_range = 20
|
||||
jitter = 0
|
||||
backlog_wait = 25
|
||||
publish_count = 3
|
||||
|
||||
self.node1.start(store="true", store_sync="true", relay="true", dns_discovery="false")
|
||||
self.node1.set_relay_subscriptions([self.test_pubsub_topic])
|
||||
|
||||
for _ in range(publish_count):
|
||||
self.publish_message(sender=self.node1, via="relay")
|
||||
|
||||
time.sleep(backlog_wait)
|
||||
self.node2.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="false",
|
||||
dns_discovery="false",
|
||||
discv5_bootstrap_node=self.node1.get_enr_uri(),
|
||||
)
|
||||
self.add_node_peer(self.node2, [self.node1.get_multiaddr_with_id()])
|
||||
|
||||
time.sleep(sync_interval * 2)
|
||||
|
||||
resp = self.get_messages_from_store(self.node2, page_size=100, cursor="", ascending="true")
|
||||
|
||||
assert len(resp.messages) == 0
|
||||
|
||||
def test_three_store_sync_exchange(self):
|
||||
msgs_per_node = 20
|
||||
total_expected = msgs_per_node * 3
|
||||
sync_interval = 6
|
||||
sync_range = 600
|
||||
jitter = 0
|
||||
publish_delay = 0.01
|
||||
wait_cycles = 3
|
||||
|
||||
self.node1.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
|
||||
for _ in range(msgs_per_node):
|
||||
self.publish_message(sender=self.node1, via="relay")
|
||||
time.sleep(publish_delay)
|
||||
|
||||
self.node2.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
|
||||
for _ in range(msgs_per_node):
|
||||
self.publish_message(sender=self.node2, via="relay")
|
||||
time.sleep(publish_delay)
|
||||
|
||||
self.add_node_peer(self.node2, [self.node1.get_multiaddr_with_id()])
|
||||
self.node3.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
for _ in range(msgs_per_node):
|
||||
self.publish_message(sender=self.node3, via="relay")
|
||||
time.sleep(publish_delay)
|
||||
|
||||
self.add_node_peer(
|
||||
self.node3,
|
||||
[self.node1.get_multiaddr_with_id(), self.node2.get_multiaddr_with_id()],
|
||||
)
|
||||
|
||||
time.sleep(sync_interval * wait_cycles)
|
||||
resp_A = self.get_messages_from_store(self.node1, page_size=200, cursor="", ascending="true", peer_id="")
|
||||
logger.debug("Node-A store has %d messages", len(resp_A.messages))
|
||||
assert len(resp_A.messages) == total_expected, f" For Node A expected {total_expected}, got {len(resp_A.messages)}"
|
||||
|
||||
resp_B = self.get_messages_from_store(self.node2, page_size=200, cursor="", ascending="true", peer_id="")
|
||||
logger.debug("Node-B store has %d messages", len(resp_B.messages))
|
||||
assert len(resp_B.messages) == total_expected, f"expected {total_expected}, got {len(resp_B.messages)}"
|
||||
resp_C = self.get_messages_from_store(self.node3, page_size=200, cursor="", ascending="true", peer_id="")
|
||||
logger.debug("Node-C store has %d messages", len(resp_C.messages))
|
||||
assert len(resp_C.messages) == total_expected, f"expected {total_expected}, got {len(resp_C.messages)}"
|
||||
|
||||
@pytest.mark.timeout(240)
|
||||
def test_node_without_sync_or_relay_stays_empty(self):
|
||||
msgs_to_publish = 30
|
||||
sync_interval = 6
|
||||
sync_range = 300
|
||||
jitter = 0
|
||||
publish_delay = 0.01
|
||||
wait_cycles = 3
|
||||
topic = self.test_pubsub_topic
|
||||
|
||||
self.node1.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
for _ in range(msgs_to_publish):
|
||||
self.publish_message(sender=self.node1, via="relay")
|
||||
time.sleep(publish_delay)
|
||||
|
||||
self.node2.start(
|
||||
# store="false",
|
||||
store_sync="false",
|
||||
relay="false",
|
||||
dns_discovery="false",
|
||||
discv5_bootstrap_node=self.node1.get_enr_uri(),
|
||||
)
|
||||
# self.add_node_peer(self.node2, [self.node1.get_multiaddr_with_id()])
|
||||
|
||||
self.node3.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="false",
|
||||
dns_discovery="false",
|
||||
discv5_bootstrap_node=self.node1.get_enr_uri(),
|
||||
)
|
||||
self.add_node_peer(self.node3, [self.node1.get_multiaddr_with_id()])
|
||||
|
||||
time.sleep(sync_interval * wait_cycles)
|
||||
resp3 = self.get_messages_from_store(self.node3, page_size=200, cursor="", ascending="true", peer_id="")
|
||||
logger.debug("Node3 store has %d messages expected %d", len(resp3.messages), msgs_to_publish)
|
||||
assert len(resp3.messages) == msgs_to_publish, f"Node3 store mismatch: expected {msgs_to_publish}, " f"got {len(resp3.messages)}"
|
||||
|
||||
def test_continuous_store_sync(self):
|
||||
msgs_per_round = 30
|
||||
rounds = 3
|
||||
sleep_between_rounds = 30
|
||||
publish_delay = 0.01
|
||||
sync_interval = 6
|
||||
sync_range = 600
|
||||
jitter = 0
|
||||
|
||||
self.node1.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
|
||||
self.node2.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="false",
|
||||
dns_discovery="false",
|
||||
)
|
||||
|
||||
self.add_node_peer(self.node2, [self.node1.get_multiaddr_with_id()])
|
||||
|
||||
total_published = 0
|
||||
for _ in range(rounds):
|
||||
for _ in range(msgs_per_round):
|
||||
self.publish_message(sender=self.node1, via="relay")
|
||||
total_published += 1
|
||||
time.sleep(publish_delay)
|
||||
|
||||
time.sleep(sync_interval * 2)
|
||||
|
||||
resp = self.get_messages_from_store(
|
||||
self.node2,
|
||||
page_size=100,
|
||||
cursor="",
|
||||
ascending="true",
|
||||
peer_id="",
|
||||
)
|
||||
logger.debug(f"Node-2 store has {len(resp.messages)}/{total_published} messages")
|
||||
assert len(resp.messages) == total_published, f"expected {total_published}, got {len(resp.messages)}"
|
||||
|
||||
time.sleep(sleep_between_rounds)
|
||||
|
||||
def test_store_sync_high_jitter_stress(self):
|
||||
sync_interval = 10
|
||||
sync_range = 120
|
||||
jitter = 90
|
||||
msgs_per_node = 50
|
||||
message_delay = 0.0
|
||||
page_size = 100
|
||||
|
||||
nodes = [self.node1, self.node2]
|
||||
|
||||
for n in nodes:
|
||||
n.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=jitter,
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
|
||||
self.add_node_peer(self.node1, [self.node2.get_multiaddr_with_id()])
|
||||
expected_hashes = []
|
||||
for _ in range(msgs_per_node):
|
||||
msgs = [self.create_message() for _ in nodes]
|
||||
for node, msg in zip(nodes, msgs):
|
||||
self.publish_message(
|
||||
sender=node,
|
||||
via="relay",
|
||||
message=msg,
|
||||
message_propagation_delay=message_delay,
|
||||
)
|
||||
expected_hashes.append(self.compute_message_hash(self.test_pubsub_topic, msg, hash_type="hex"))
|
||||
|
||||
delay(120)
|
||||
|
||||
for node in nodes:
|
||||
store_resp = self.get_messages_from_store(node, page_size=page_size, ascending="true")
|
||||
retrieved_hashes = [store_resp.message_hash(i) for i in range(len(store_resp.messages))]
|
||||
assert len(retrieved_hashes) == len(expected_hashes), " message count mismatch"
|
||||
assert retrieved_hashes == expected_hashes, "{ message hash mismatch"
|
||||
|
||||
def test_store_sync_five_node_mesh_burst(self):
|
||||
sync_interval = 3
|
||||
sync_range = 900
|
||||
msgs_per_node = 100
|
||||
message_delay = 0.0
|
||||
page_size = 100
|
||||
|
||||
self.node4 = WakuNode(NODE_2, f"node4_{self.test_id}")
|
||||
self.node5 = WakuNode(NODE_2, f"node5_{self.test_id}")
|
||||
nodes = [self.node1, self.node2, self.node3, self.node4, self.node5]
|
||||
|
||||
for n in nodes:
|
||||
n.start(
|
||||
store="true",
|
||||
store_sync="true",
|
||||
store_sync_interval=sync_interval,
|
||||
store_sync_range=sync_range,
|
||||
store_sync_relay_jitter=0,
|
||||
relay="true",
|
||||
dns_discovery="false",
|
||||
)
|
||||
n.set_relay_subscriptions([self.test_pubsub_topic])
|
||||
|
||||
for i, a in enumerate(nodes):
|
||||
for b in nodes[i + 1 :]:
|
||||
self.add_node_peer(a, [b.get_multiaddr_with_id()])
|
||||
self.add_node_peer(b, [a.get_multiaddr_with_id()])
|
||||
|
||||
expected_hashes = []
|
||||
for _ in range(msgs_per_node):
|
||||
msgs = [self.create_message() for _ in nodes]
|
||||
for node, msg in zip(nodes, msgs):
|
||||
self.publish_message(
|
||||
sender=node,
|
||||
via="relay",
|
||||
message=msg,
|
||||
message_propagation_delay=message_delay,
|
||||
)
|
||||
expected_hashes.append(self.compute_message_hash(self.test_pubsub_topic, msg, hash_type="hex"))
|
||||
|
||||
delay(sync_interval * 4 + 20)
|
||||
|
||||
for node in nodes:
|
||||
store_resp = StoreResponse({"paginationCursor": "", "pagination_cursor": ""}, node)
|
||||
retrieved_hashes = []
|
||||
while store_resp.pagination_cursor is not None:
|
||||
cursor = store_resp.pagination_cursor
|
||||
store_resp = self.get_messages_from_store(
|
||||
node,
|
||||
page_size=page_size,
|
||||
cursor=cursor,
|
||||
ascending="true",
|
||||
)
|
||||
for i in range(len(store_resp.messages)):
|
||||
retrieved_hashes.append(store_resp.message_hash(i))
|
||||
assert len(retrieved_hashes) == len(expected_hashes), f"{node.name}: message count mismatch"
|
||||
assert retrieved_hashes == expected_hashes, f"{node.name}: message hash mismatch"
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user