1147 lines
34 KiB
Go
Raw Permalink Normal View History

2024-11-27 12:08:50 +02:00
package waku
import (
"context"
"errors"
"fmt"
2024-11-27 12:08:50 +02:00
"slices"
"testing"
"time"
"go.uber.org/zap"
2024-12-13 17:26:49 +01:00
"google.golang.org/protobuf/proto"
2024-11-27 12:08:50 +02:00
"github.com/cenkalti/backoff/v3"
"github.com/libp2p/go-libp2p/core/peer"
"github.com/logos-messaging/logos-messaging-go-bindings/waku/common"
2024-11-27 12:08:50 +02:00
ma "github.com/multiformats/go-multiaddr"
"github.com/stretchr/testify/require"
2025-12-16 03:06:15 +05:30
"github.com/logos-messaging/logos-messaging-go/waku/v2/protocol/pb"
"github.com/logos-messaging/logos-messaging-go/waku/v2/protocol/store"
2024-11-27 12:08:50 +02:00
)
// In order to run this test, you must run an nwaku node
//
// Using Docker:
//
// IP_ADDRESS=$(hostname -I | awk '{print $1}');
// docker run \
// -p 61000:61000/tcp -p 8000:8000/udp -p 8646:8646/tcp harbor.status.im/wakuorg/nwaku:v0.33.0 \
// --discv5-discovery=true --cluster-id=16 --log-level=DEBUG --shard=64 --tcp-port=61000 \
// --nat=extip:${IP_ADDRESS} --discv5-udp-port=8000 --rest-address=0.0.0.0 --store --rest-port=8646
func TestBasicWaku(t *testing.T) {
2025-03-18 11:13:42 +02:00
t.Skip("Skipping test as choosing this port will fail the CI")
2024-11-27 12:08:50 +02:00
extNodeRestPort := 8646
storeNodeInfo, err := GetNwakuInfo(nil, &extNodeRestPort)
require.NoError(t, err)
2024-12-04 17:41:28 +01:00
// ctx := context.Background()
nwakuConfig := common.WakuConfig{
Nodekey: "11d0dcea28e86f81937a3bd1163473c7fbc0a0db54fd72914849bc47bdf78710",
Relay: true,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
DnsDiscoveryUrl: "enrtree://AMOJVZX4V6EXP7NTJPMAYJYST2QP6AJXYW76IU6VGJS7UVSNDYZG4@boot.prod.status.nodes.status.im",
DnsDiscovery: true,
Discv5Discovery: true,
Staticnodes: []string{storeNodeInfo.ListenAddresses[0]},
ClusterID: 16,
Shards: []uint16{64},
}
storeNodeMa, err := ma.NewMultiaddr(storeNodeInfo.ListenAddresses[0])
require.NoError(t, err)
2025-02-11 16:21:14 +02:00
w, err := NewWakuNode(&nwakuConfig, "nwaku")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, w.Start())
2025-01-10 11:21:50 +01:00
enr, err := w.ENR()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NotNil(t, enr)
options := func(b *backoff.ExponentialBackOff) {
b.MaxElapsedTime = 30 * time.Second
}
// Sanity check, not great, but it's probably helpful
err = RetryWithBackOff(func() error {
2025-01-10 11:21:50 +01:00
numConnected, err := w.GetNumConnectedPeers()
2024-11-27 12:08:50 +02:00
if err != nil {
return err
}
// Have to be connected to at least 3 nodes: the static node, the bootstrap node, and one discovered node
if numConnected > 2 {
return nil
}
return errors.New("no peers discovered")
}, options)
require.NoError(t, err)
// Get local store node address
storeNode, err := peer.AddrInfoFromString(storeNodeInfo.ListenAddresses[0])
require.NoError(t, err)
2024-12-04 17:41:28 +01:00
/*
w.node.DialPeer(ctx, storeNode.Addrs[0], "")
2024-11-27 12:08:50 +02:00
2024-12-04 17:41:28 +01:00
w.StorenodeCycle.SetStorenodeConfigProvider(newTestStorenodeConfigProvider(*storeNode))
*/
2024-11-27 12:08:50 +02:00
// Check that we are indeed connected to the store node
2025-01-10 11:21:50 +01:00
connectedStoreNodes, err := w.GetPeerIDsByProtocol(store.StoreQueryID_v300)
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, slices.Contains(connectedStoreNodes, storeNode.ID), "nwaku should be connected to the store node")
// Disconnect from the store node
2025-01-10 11:21:50 +01:00
err = w.DisconnectPeerByID(storeNode.ID)
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
// Check that we are indeed disconnected
2025-01-10 11:21:50 +01:00
connectedStoreNodes, err = w.GetPeerIDsByProtocol(store.StoreQueryID_v300)
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
isDisconnected := !slices.Contains(connectedStoreNodes, storeNode.ID)
require.True(t, isDisconnected, "nwaku should be disconnected from the store node")
// Re-connect
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
2025-01-10 11:21:50 +01:00
err = w.Connect(ctx, storeNodeMa)
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
// Check that we are connected again
2025-01-10 11:21:50 +01:00
connectedStoreNodes, err = w.GetPeerIDsByProtocol(store.StoreQueryID_v300)
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, slices.Contains(connectedStoreNodes, storeNode.ID), "nwaku should be connected to the store node")
/* filter := &common.Filter{
PubsubTopic: w.cfg.DefaultShardPubsubTopic,
Messages: common.NewMemoryMessageStore(),
ContentTopics: common.NewTopicSetFromBytes([][]byte{{1, 2, 3, 4}}),
}
_, err = w.Subscribe(filter)
require.NoError(t, err)
msgTimestamp := w.timestamp()
contentTopic := maps.Keys(filter.ContentTopics)[0]
time.Sleep(2 * time.Second)
msgID, err := w.Send(w.cfg.DefaultShardPubsubTopic, &pb.WakuMessage{
Payload: []byte{1, 2, 3, 4, 5},
ContentTopic: contentTopic.ContentTopic(),
Version: proto.Uint32(0),
Timestamp: &msgTimestamp,
}, nil)
require.NoError(t, err)
require.NotEqual(t, msgID, "1")
time.Sleep(1 * time.Second)
messages := filter.Retrieve()
require.Len(t, messages, 1)
timestampInSeconds := msgTimestamp / int64(time.Second)
marginInSeconds := 20
options = func(b *backoff.ExponentialBackOff) {
b.MaxElapsedTime = 60 * time.Second
b.InitialInterval = 500 * time.Millisecond
}
err = RetryWithBackOff(func() error {
err := w.HistoryRetriever.Query(
context.Background(),
store.FilterCriteria{
ContentFilter: protocol.NewContentFilter(w.cfg.DefaultShardPubsubTopic, contentTopic.ContentTopic()),
TimeStart: proto.Int64((timestampInSeconds - int64(marginInSeconds)) * int64(time.Second)),
TimeEnd: proto.Int64((timestampInSeconds + int64(marginInSeconds)) * int64(time.Second)),
},
*storeNode,
10,
nil, false,
)
return err
// TODO-nwaku
if err != nil || envelopeCount == 0 {
// in case of failure extend timestamp margin up to 40secs
if marginInSeconds < 40 {
marginInSeconds += 5
}
return errors.New("no messages received from store node")
}
return nil
}, options)
require.NoError(t, err)
time.Sleep(10 * time.Second)
*/
require.NoError(t, w.Stop())
2025-02-11 16:21:14 +02:00
2024-11-27 12:08:50 +02:00
}
func TestPeerExchange(t *testing.T) {
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-11-27 12:08:50 +02:00
// start node that will be discovered by PeerExchange
discV5NodeWakuConfig := common.WakuConfig{
Relay: true,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
Discv5Discovery: true,
ClusterID: 16,
Shards: []uint16{64},
PeerExchange: false,
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-11-27 12:08:50 +02:00
}
2025-02-11 16:21:14 +02:00
discV5Node, err := NewWakuNode(&discV5NodeWakuConfig, "discV5Node")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, discV5Node.Start())
2025-01-10 11:21:50 +01:00
discV5NodePeerId, err := discV5Node.PeerID()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
2025-01-10 11:21:50 +01:00
discv5NodeEnr, err := discV5Node.ENR()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-11-27 12:08:50 +02:00
// start node which serves as PeerExchange server
pxServerWakuConfig := common.WakuConfig{
Relay: true,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
Discv5Discovery: true,
ClusterID: 16,
Shards: []uint16{64},
PeerExchange: true,
Discv5UdpPort: udpPort,
2024-11-27 12:08:50 +02:00
Discv5BootstrapNodes: []string{discv5NodeEnr.String()},
TcpPort: tcpPort,
2024-11-27 12:08:50 +02:00
}
2025-02-11 16:21:14 +02:00
pxServerNode, err := NewWakuNode(&pxServerWakuConfig, "pxServerNode")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, pxServerNode.Start())
// Adding an extra second to make sure PX cache is not empty
time.Sleep(2 * time.Second)
serverNodeMa, err := pxServerNode.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, serverNodeMa)
require.True(t, len(serverNodeMa) > 0)
2024-11-27 12:08:50 +02:00
// Sanity check, not great, but it's probably helpful
options := func(b *backoff.ExponentialBackOff) {
b.MaxElapsedTime = 30 * time.Second
}
// Check that pxServerNode has discV5Node in its Peer Store
err = RetryWithBackOff(func() error {
2025-01-10 11:21:50 +01:00
peers, err := pxServerNode.GetPeerIDsFromPeerStore()
2024-11-27 12:08:50 +02:00
if err != nil {
return err
}
if slices.Contains(peers, discV5NodePeerId) {
return nil
}
return errors.New("pxServer is missing the discv5 node in its peer store")
}, options)
require.NoError(t, err)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-11-27 12:08:50 +02:00
// start light node which uses PeerExchange to discover peers
pxClientWakuConfig := common.WakuConfig{
Relay: false,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
PeerExchange: true,
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-11-27 12:08:50 +02:00
PeerExchangeNode: serverNodeMa[0].String(),
}
2025-02-11 16:21:14 +02:00
lightNode, err := NewWakuNode(&pxClientWakuConfig, "lightNode")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, lightNode.Start())
2025-01-10 11:21:50 +01:00
pxServerPeerId, err := pxServerNode.PeerID()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
// Check that the light node discovered the discV5Node and has both nodes in its peer store
err = RetryWithBackOff(func() error {
2025-01-10 11:21:50 +01:00
peers, err := lightNode.GetPeerIDsFromPeerStore()
2024-11-27 12:08:50 +02:00
if err != nil {
return err
}
if slices.Contains(peers, discV5NodePeerId) && slices.Contains(peers, pxServerPeerId) {
return nil
}
return errors.New("lightnode is missing peers")
}, options)
require.NoError(t, err)
// Now perform the PX request manually to see if it also works
err = RetryWithBackOff(func() error {
2025-01-10 11:21:50 +01:00
numPeersReceived, err := lightNode.PeerExchangeRequest(1)
2024-11-27 12:08:50 +02:00
if err != nil {
return err
}
if numPeersReceived == 1 {
return nil
}
return errors.New("Peer Exchange is not returning peers")
}, options)
require.NoError(t, err)
// Stop nodes
require.NoError(t, lightNode.Stop())
require.NoError(t, pxServerNode.Stop())
require.NoError(t, discV5Node.Stop())
}
func TestDnsDiscover(t *testing.T) {
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-11-27 12:08:50 +02:00
nameserver := "8.8.8.8"
nodeWakuConfig := common.WakuConfig{
Relay: true,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-11-27 12:08:50 +02:00
}
2025-01-10 11:21:50 +01:00
2025-02-11 16:21:14 +02:00
node, err := NewWakuNode(&nodeWakuConfig, "node")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, node.Start())
sampleEnrTree := "enrtree://AMOJVZX4V6EXP7NTJPMAYJYST2QP6AJXYW76IU6VGJS7UVSNDYZG4@boot.prod.status.nodes.status.im"
ctx, cancel := context.WithTimeout(context.TODO(), requestTimeout)
defer cancel()
2025-01-10 11:21:50 +01:00
res, err := node.DnsDiscovery(ctx, sampleEnrTree, nameserver)
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, len(res) > 1, "multiple nodes should be returned from the DNS Discovery query")
// Stop nodes
require.NoError(t, node.Stop())
}
func TestDial(t *testing.T) {
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-11-27 12:08:50 +02:00
// start node that will initiate the dial
dialerNodeWakuConfig := common.WakuConfig{
Relay: true,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-11-27 12:08:50 +02:00
}
2025-02-11 16:21:14 +02:00
dialerNode, err := NewWakuNode(&dialerNodeWakuConfig, "dialerNode")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, dialerNode.Start())
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-11-27 12:08:50 +02:00
// start node that will receive the dial
receiverNodeWakuConfig := common.WakuConfig{
Relay: true,
2024-11-27 12:08:50 +02:00
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-11-27 12:08:50 +02:00
}
2025-01-10 11:21:50 +01:00
2025-02-11 16:21:14 +02:00
receiverNode, err := NewWakuNode(&receiverNodeWakuConfig, "receiverNode")
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.NoError(t, receiverNode.Start())
receiverMultiaddr, err := receiverNode.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, receiverMultiaddr)
require.True(t, len(receiverMultiaddr) > 0)
2024-11-27 12:08:50 +02:00
// Check that both nodes start with no connected peers
2025-01-10 11:21:50 +01:00
dialerPeerCount, err := dialerNode.GetNumConnectedPeers()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, dialerPeerCount == 0, "Dialer node should have no connected peers")
2025-01-10 11:21:50 +01:00
receiverPeerCount, err := receiverNode.GetNumConnectedPeers()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, receiverPeerCount == 0, "Receiver node should have no connected peers")
// Dial
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
2025-01-10 11:21:50 +01:00
err = dialerNode.Connect(ctx, receiverMultiaddr[0])
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
time.Sleep(1 * time.Second)
// Check that both nodes now have one connected peer
2025-01-10 11:21:50 +01:00
dialerPeerCount, err = dialerNode.GetNumConnectedPeers()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, dialerPeerCount == 1, "Dialer node should have 1 peer")
2025-01-10 11:21:50 +01:00
receiverPeerCount, err = receiverNode.GetNumConnectedPeers()
2024-11-27 12:08:50 +02:00
require.NoError(t, err)
require.True(t, receiverPeerCount == 1, "Receiver node should have 1 peer")
// Stop nodes
require.NoError(t, dialerNode.Stop())
require.NoError(t, receiverNode.Stop())
}
2024-12-13 17:26:49 +01:00
func TestRelay(t *testing.T) {
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-12-13 17:26:49 +01:00
// start node that will send the message
senderNodeWakuConfig := common.WakuConfig{
2024-12-13 17:26:49 +01:00
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-12-13 17:26:49 +01:00
}
2025-02-11 16:21:14 +02:00
senderNode, err := NewWakuNode(&senderNodeWakuConfig, "senderNode")
2024-12-13 17:26:49 +01:00
require.NoError(t, err)
require.NoError(t, senderNode.Start())
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
2024-12-13 17:26:49 +01:00
// start node that will receive the message
receiverNodeWakuConfig := common.WakuConfig{
2024-12-13 17:26:49 +01:00
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
2024-12-13 17:26:49 +01:00
}
2025-02-11 16:21:14 +02:00
receiverNode, err := NewWakuNode(&receiverNodeWakuConfig, "receiverNode")
2024-12-13 17:26:49 +01:00
require.NoError(t, err)
require.NoError(t, receiverNode.Start())
receiverMultiaddr, err := receiverNode.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, receiverMultiaddr)
require.True(t, len(receiverMultiaddr) > 0)
2024-12-13 17:26:49 +01:00
// Dial so they become peers
2025-01-10 11:21:50 +01:00
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
err = senderNode.Connect(ctx, receiverMultiaddr[0])
2024-12-13 17:26:49 +01:00
require.NoError(t, err)
time.Sleep(1 * time.Second)
// Check that both nodes now have one connected peer
2025-01-10 11:21:50 +01:00
senderPeerCount, err := senderNode.GetNumConnectedPeers()
2024-12-13 17:26:49 +01:00
require.NoError(t, err)
require.True(t, senderPeerCount == 1, "Dialer node should have 1 peer")
2025-01-10 11:21:50 +01:00
receiverPeerCount, err := receiverNode.GetNumConnectedPeers()
2024-12-13 17:26:49 +01:00
require.NoError(t, err)
require.True(t, receiverPeerCount == 1, "Receiver node should have 1 peer")
message := &pb.WakuMessage{
Payload: []byte{1, 2, 3, 4, 5, 6},
ContentTopic: "test-content-topic",
Version: proto.Uint32(0),
Timestamp: proto.Int64(time.Now().UnixNano()),
2024-12-13 17:26:49 +01:00
}
// send message
pubsubTopic := FormatWakuRelayTopic(senderNodeWakuConfig.ClusterID, senderNodeWakuConfig.Shards[0])
ctx2, cancel2 := context.WithTimeout(context.Background(), requestTimeout)
defer cancel2()
senderNode.RelayPublish(ctx2, message, pubsubTopic)
2024-12-13 17:26:49 +01:00
// Wait to receive message
select {
2025-01-10 11:21:50 +01:00
case envelope := <-receiverNode.MsgChan:
2024-12-13 17:26:49 +01:00
require.NotNil(t, envelope, "Envelope should be received")
require.Equal(t, message.Payload, envelope.Message().Payload, "Received payload should match")
require.Equal(t, message.ContentTopic, envelope.Message().ContentTopic, "Content topic should match")
case <-time.After(10 * time.Second):
t.Fatal("Timeout: No message received within 10 seconds")
}
// Stop nodes
require.NoError(t, senderNode.Stop())
require.NoError(t, receiverNode.Stop())
}
func TestTopicHealth(t *testing.T) {
clusterId := uint16(16)
shardId := uint16(64)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node1
wakuConfig1 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
node1, err := NewWakuNode(&wakuConfig1, "node1")
require.NoError(t, err)
require.NoError(t, node1.Start())
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node2
wakuConfig2 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
node2, err := NewWakuNode(&wakuConfig2, "node2")
require.NoError(t, err)
require.NoError(t, node2.Start())
multiaddr2, err := node2.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, multiaddr2)
require.True(t, len(multiaddr2) > 0)
// node1 dials node2 so they become peers
2025-01-10 11:21:50 +01:00
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
err = node1.Connect(ctx, multiaddr2[0])
require.NoError(t, err)
time.Sleep(1 * time.Second)
// Check that both nodes now have one connected peer
2025-01-10 11:21:50 +01:00
peerCount1, err := node1.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount1 == 1, "node1 should have 1 peer")
2025-01-10 11:21:50 +01:00
peerCount2, err := node2.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount2 == 1, "node2 should have 1 peer")
// Wait to receive topic health update
select {
2025-01-10 11:21:50 +01:00
case topicHealth := <-node2.TopicHealthChan:
require.NotNil(t, topicHealth, "topicHealth should be updated")
require.Equal(t, topicHealth.TopicHealth, "MinimallyHealthy", "Topic health should be MinimallyHealthy")
require.Equal(t, topicHealth.PubsubTopic, FormatWakuRelayTopic(clusterId, shardId), "PubsubTopic should match configured cluster and shard")
case <-time.After(10 * time.Second):
t.Fatal("Timeout: No topic health event received within 10 seconds")
}
// Stop nodes
require.NoError(t, node1.Stop())
require.NoError(t, node2.Stop())
}
func TestConnectionChange(t *testing.T) {
clusterId := uint16(16)
shardId := uint16(64)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node1
wakuConfig1 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
node1, err := NewWakuNode(&wakuConfig1, "node1")
require.NoError(t, err)
require.NoError(t, node1.Start())
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node2
wakuConfig2 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
node2, err := NewWakuNode(&wakuConfig2, "node2")
require.NoError(t, err)
require.NoError(t, node2.Start())
multiaddr2, err := node2.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, multiaddr2)
require.True(t, len(multiaddr2) > 0)
// node1 dials node2 so they become peers
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
err = node1.Connect(ctx, multiaddr2[0])
require.NoError(t, err)
time.Sleep(1 * time.Second)
// Check that both nodes now have one connected peer
peerCount1, err := node1.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount1 == 1, "node1 should have 1 peer")
peerCount2, err := node2.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount2 == 1, "node2 should have 1 peer")
peerId1, err := node1.PeerID()
require.NoError(t, err)
// Wait to receive connectionChange event
select {
case connectionChange := <-node2.ConnectionChangeChan:
require.NotNil(t, connectionChange, "connectionChange should be updated")
require.Equal(t, connectionChange.PeerEvent, "Joined", "connectionChange Joined event should be emitted")
require.Equal(t, connectionChange.PeerId, peerId1, "connectionChange event should contain node 1's peerId")
case <-time.After(10 * time.Second):
t.Fatal("Timeout: No connectionChange event received within 10 seconds")
}
// Disconnect from node1
err = node2.DisconnectPeerByID(peerId1)
require.NoError(t, err)
// Wait to receive connectionChange event
select {
case connectionChange := <-node2.ConnectionChangeChan:
require.NotNil(t, connectionChange, "connectionChange should be updated")
require.Equal(t, connectionChange.PeerEvent, "Left", "connectionChange Left event should be emitted")
require.Equal(t, connectionChange.PeerId, peerId1, "connectionChange event should contain node 1's peerId")
case <-time.After(10 * time.Second):
t.Fatal("Timeout: No connectionChange event received within 10 seconds")
}
// Stop nodes
require.NoError(t, node1.Stop())
require.NoError(t, node2.Stop())
}
func TestStore(t *testing.T) {
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node that will send the message
senderNodeWakuConfig := common.WakuConfig{
Relay: true,
Store: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
LegacyStore: false,
}
2025-02-11 16:21:14 +02:00
senderNode, err := NewWakuNode(&senderNodeWakuConfig, "senderNode")
require.NoError(t, err)
require.NoError(t, senderNode.Start())
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node that will receive the message
receiverNodeWakuConfig := common.WakuConfig{
Relay: true,
Store: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
LegacyStore: false,
}
2025-02-11 16:21:14 +02:00
receiverNode, err := NewWakuNode(&receiverNodeWakuConfig, "receiverNode")
require.NoError(t, err)
require.NoError(t, receiverNode.Start())
receiverMultiaddr, err := receiverNode.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, receiverMultiaddr)
require.True(t, len(receiverMultiaddr) > 0)
// Dial so they become peers
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
err = senderNode.Connect(ctx, receiverMultiaddr[0])
require.NoError(t, err)
time.Sleep(1 * time.Second)
// Check that both nodes now have one connected peer
senderPeerCount, err := senderNode.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, senderPeerCount == 1, "Dialer node should have 1 peer")
receiverPeerCount, err := receiverNode.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, receiverPeerCount == 1, "Receiver node should have 1 peer")
// Send 8 messages
numMessages := 8
paginationLimit := 5
timeStart := proto.Int64(time.Now().UnixNano())
hashes := []common.MessageHash{}
pubsubTopic := FormatWakuRelayTopic(senderNodeWakuConfig.ClusterID, senderNodeWakuConfig.Shards[0])
for i := 0; i < numMessages; i++ {
message := &pb.WakuMessage{
Payload: []byte{byte(i)}, // Include message number in payload
ContentTopic: "test-content-topic",
Version: proto.Uint32(0),
Timestamp: proto.Int64(time.Now().UnixNano()),
}
ctx2, cancel2 := context.WithTimeout(context.Background(), requestTimeout)
defer cancel2()
hash, err := senderNode.RelayPublish(ctx2, message, pubsubTopic)
require.NoError(t, err)
hashes = append(hashes, hash)
}
// Wait to receive all 8 messages
receivedCount := 0
receivedMessages := make(map[byte]bool)
// Use a timeout for the entire receive operation
timeoutChan := time.After(10 * time.Second)
for receivedCount < numMessages {
select {
case envelope := <-receiverNode.MsgChan:
require.NotNil(t, envelope, "Envelope should be received")
payload := envelope.Message().Payload
msgNum := payload[0]
// Check if we've already received this message number
if !receivedMessages[msgNum] {
receivedMessages[msgNum] = true
receivedCount++
}
require.Equal(t, "test-content-topic", envelope.Message().ContentTopic, "Content topic should match")
case <-timeoutChan:
t.Fatalf("Timeout: Only received %d messages out of 8 within 10 seconds", receivedCount)
}
}
// Verify we received all messages
for i := 0; i < numMessages; i++ {
require.True(t, receivedMessages[byte(i)], fmt.Sprintf("Message %d was not received", i))
}
// Now send store query
storeReq1 := common.StoreQueryRequest{
IncludeData: true,
ContentTopics: &[]string{"test-content-topic"},
PaginationLimit: proto.Uint64(uint64(paginationLimit)),
PaginationForward: true,
TimeStart: timeStart,
}
storeNodeAddrInfo, err := peer.AddrInfoFromString(receiverMultiaddr[0].String())
require.NoError(t, err)
ctx3, cancel3 := context.WithTimeout(context.Background(), requestTimeout)
defer cancel3()
res1, err := senderNode.StoreQuery(ctx3, &storeReq1, *storeNodeAddrInfo)
require.NoError(t, err)
storedMessages1 := *res1.Messages
for i := 0; i < paginationLimit; i++ {
require.True(t, storedMessages1[i].MessageHash == hashes[i], fmt.Sprintf("Stored message does not match received message for index %d", i))
}
// Now let's query the second page
storeReq2 := common.StoreQueryRequest{
IncludeData: true,
ContentTopics: &[]string{"test-content-topic"},
PaginationLimit: proto.Uint64(uint64(paginationLimit)),
PaginationForward: true,
TimeStart: timeStart,
PaginationCursor: &res1.PaginationCursor,
}
ctx4, cancel4 := context.WithTimeout(context.Background(), requestTimeout)
defer cancel4()
res2, err := senderNode.StoreQuery(ctx4, &storeReq2, *storeNodeAddrInfo)
require.NoError(t, err)
storedMessages2 := *res2.Messages
for i := 0; i < len(storedMessages2); i++ {
require.True(t, storedMessages2[i].MessageHash == hashes[i+paginationLimit], fmt.Sprintf("Stored message does not match received message for index %d", i))
}
// Now let's query for two specific message hashes
storeReq3 := common.StoreQueryRequest{
IncludeData: true,
ContentTopics: &[]string{"test-content-topic"},
MessageHashes: &[]common.MessageHash{hashes[0], hashes[2]},
}
ctx5, cancel5 := context.WithTimeout(context.Background(), requestTimeout)
defer cancel5()
res3, err := senderNode.StoreQuery(ctx5, &storeReq3, *storeNodeAddrInfo)
require.NoError(t, err)
storedMessages3 := *res3.Messages
require.True(t, storedMessages3[0].MessageHash == hashes[0], "Stored message does not match queried message")
require.True(t, storedMessages3[1].MessageHash == hashes[2], "Stored message does not match queried message")
// Stop nodes
require.NoError(t, senderNode.Stop())
require.NoError(t, receiverNode.Stop())
}
2025-02-05 18:16:57 +02:00
func TestParallelPings(t *testing.T) {
logger, err := zap.NewDevelopment()
require.NoError(t, err)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
// start node that will initiate the dial
dialerNodeWakuConfig := common.WakuConfig{
2025-02-05 18:16:57 +02:00
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
dialerNode, err := NewWakuNode(&dialerNodeWakuConfig, "dialerNode")
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
require.NoError(t, dialerNode.Start())
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
receiverNodeWakuConfig1 := common.WakuConfig{
2025-02-05 18:16:57 +02:00
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
receiverNode1, err := NewWakuNode(&receiverNodeWakuConfig1, "receiverNode1")
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
require.NoError(t, receiverNode1.Start())
receiverMultiaddr1, err := receiverNode1.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, receiverMultiaddr1)
require.True(t, len(receiverMultiaddr1) > 0)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
receiverNodeWakuConfig2 := common.WakuConfig{
2025-02-05 18:16:57 +02:00
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
receiverNode2, err := NewWakuNode(&receiverNodeWakuConfig2, "receiverNode2")
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
require.NoError(t, receiverNode2.Start())
receiverMultiaddr2, err := receiverNode2.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, receiverMultiaddr2)
require.True(t, len(receiverMultiaddr2) > 0)
2025-02-11 16:21:14 +02:00
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
receiverNodeWakuConfig3 := common.WakuConfig{
2025-02-05 18:16:57 +02:00
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: 16,
Shards: []uint16{64},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
2025-02-11 16:21:14 +02:00
receiverNode3, err := NewWakuNode(&receiverNodeWakuConfig3, "receiverNode3")
2025-02-05 18:16:57 +02:00
require.NoError(t, err)
require.NoError(t, receiverNode3.Start())
receiverMultiaddr3, err := receiverNode3.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, receiverMultiaddr3)
require.True(t, len(receiverMultiaddr3) > 0)
receiverNodes := []string{receiverMultiaddr1[0].String(), receiverMultiaddr2[0].String(), receiverMultiaddr3[0].String()}
// node.PingPeer(ctx, peerInfo)
for _, receiverNode := range receiverNodes {
addrInfo, err := peer.AddrInfoFromString(receiverNode)
require.NoError(t, err)
go func(peerInfo peer.AddrInfo) {
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
_, err := dialerNode.PingPeer(ctx, peerInfo)
if err != nil { // pinging storenodes might fail, but we don't care
logger.Warn("failed pinging node", zap.Stringer("peerId", addrInfo.ID), zap.Error(err))
}
}(*addrInfo)
}
options := func(b *backoff.ExponentialBackOff) {
b.MaxElapsedTime = 30 * time.Second
}
err = RetryWithBackOff(func() error {
dialerPeerCount, err := dialerNode.GetNumConnectedPeers()
if err != nil {
return err
}
if dialerPeerCount == 3 {
return nil
}
return fmt.Errorf("dialerNode should have 3 peers but it has %d", dialerPeerCount)
}, options)
require.NoError(t, err)
// Stop nodes
require.NoError(t, dialerNode.Stop())
}
func TestOnline(t *testing.T) {
clusterId := uint16(16)
shardId := uint16(64)
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node1
wakuConfig1 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
node1, err := NewWakuNode(&wakuConfig1, "node1")
require.NoError(t, err)
require.NoError(t, node1.Start())
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node2
wakuConfig2 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
node2, err := NewWakuNode(&wakuConfig2, "node2")
require.NoError(t, err)
require.NoError(t, node2.Start())
multiaddr2, err := node2.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, multiaddr2)
require.True(t, len(multiaddr2) > 0)
// node1 dials node2 so they become peers
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
err = node1.Connect(ctx, multiaddr2[0])
require.NoError(t, err)
time.Sleep(1 * time.Second)
// Check that both nodes now have one connected peer
peerCount1, err := node1.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount1 == 1, "node1 should have 1 peer")
peerCount2, err := node2.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount2 == 1, "node2 should have 1 peer")
isOnline, err := node1.IsOnline()
require.NoError(t, err)
require.True(t, isOnline, "node1 should be online")
// Stop nodes
require.NoError(t, node1.Stop())
require.NoError(t, node2.Stop())
}
func TestDisconnectAllPeers(t *testing.T) {
clusterId := uint16(16)
shardId := uint16(64)
tcpPort, udpPort, err := GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node1
wakuConfig1 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
node1, err := NewWakuNode(&wakuConfig1, "node1")
require.NoError(t, err)
require.NoError(t, node1.Start())
defer node1.Stop()
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node2
wakuConfig2 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
node2, err := NewWakuNode(&wakuConfig2, "node2")
require.NoError(t, err)
require.NoError(t, node2.Start())
defer node2.Stop()
multiaddr2, err := node2.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, multiaddr2)
require.True(t, len(multiaddr2) > 0)
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node3
wakuConfig3 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
node3, err := NewWakuNode(&wakuConfig3, "node3")
require.NoError(t, err)
require.NoError(t, node3.Start())
defer node3.Stop()
multiaddr3, err := node3.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, multiaddr3)
require.True(t, len(multiaddr3) > 0)
tcpPort, udpPort, err = GetFreePortIfNeeded(0, 0)
require.NoError(t, err)
// start node4
wakuConfig4 := common.WakuConfig{
Relay: true,
LogLevel: "DEBUG",
Discv5Discovery: false,
ClusterID: clusterId,
Shards: []uint16{shardId},
Discv5UdpPort: udpPort,
TcpPort: tcpPort,
}
node4, err := NewWakuNode(&wakuConfig4, "node4")
require.NoError(t, err)
require.NoError(t, node4.Start())
defer node4.Stop()
multiaddr4, err := node4.ListenAddresses()
require.NoError(t, err)
require.NotNil(t, multiaddr4)
require.True(t, len(multiaddr4) > 0)
to_dial := []ma.Multiaddr{multiaddr2[0], multiaddr3[0], multiaddr4[0]}
for _, addr := range to_dial {
ctx, cancel := context.WithTimeout(context.Background(), requestTimeout)
defer cancel()
err = node1.Connect(ctx, addr)
require.NoError(t, err)
}
time.Sleep(1 * time.Second)
peerCount1, err := node1.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount1 == 3, "node1 should have 3 peers")
err = node1.DisconnectAllPeers()
require.NoError(t, err)
peerCount1, err = node1.GetNumConnectedPeers()
require.NoError(t, err)
require.True(t, peerCount1 == 0, "node1 should have 0 peers")
}