2022-12-09 16:16:21 +00:00
|
|
|
package wakuv2
|
|
|
|
|
|
|
|
import (
|
2023-02-22 21:58:17 +00:00
|
|
|
"context"
|
|
|
|
"crypto/rand"
|
2022-12-09 16:16:21 +00:00
|
|
|
"errors"
|
2023-02-22 21:58:17 +00:00
|
|
|
"math/big"
|
|
|
|
"os"
|
2024-06-28 13:54:48 +00:00
|
|
|
"sync"
|
2022-12-09 16:16:21 +00:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2024-02-27 09:24:34 +00:00
|
|
|
"go.uber.org/zap"
|
|
|
|
|
2022-12-09 16:16:21 +00:00
|
|
|
"github.com/cenkalti/backoff/v3"
|
2023-10-12 19:21:49 +00:00
|
|
|
|
2024-06-11 07:45:01 +00:00
|
|
|
ethcommon "github.com/ethereum/go-ethereum/common"
|
2023-12-07 00:28:08 +00:00
|
|
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
|
|
|
"github.com/ethereum/go-ethereum/crypto"
|
|
|
|
ethdnsdisc "github.com/ethereum/go-ethereum/p2p/dnsdisc"
|
|
|
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
|
|
|
|
|
|
|
"github.com/stretchr/testify/require"
|
2023-10-12 19:21:49 +00:00
|
|
|
"golang.org/x/exp/maps"
|
2023-12-07 00:28:08 +00:00
|
|
|
"google.golang.org/protobuf/proto"
|
2023-10-12 19:21:49 +00:00
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
"github.com/waku-org/go-waku/waku/v2/dnsdisc"
|
2024-07-12 10:34:56 +00:00
|
|
|
wps "github.com/waku-org/go-waku/waku/v2/peerstore"
|
2024-06-14 12:41:45 +00:00
|
|
|
"github.com/waku-org/go-waku/waku/v2/protocol/filter"
|
|
|
|
"github.com/waku-org/go-waku/waku/v2/protocol/legacy_store"
|
|
|
|
"github.com/waku-org/go-waku/waku/v2/protocol/pb"
|
2024-06-19 10:40:52 +00:00
|
|
|
|
|
|
|
"github.com/status-im/status-go/appdatabase"
|
2024-06-28 13:54:48 +00:00
|
|
|
"github.com/status-im/status-go/connection"
|
2024-06-19 10:40:52 +00:00
|
|
|
"github.com/status-im/status-go/eth-node/types"
|
2024-07-15 15:25:12 +00:00
|
|
|
"github.com/status-im/status-go/protocol/common/shard"
|
2024-06-19 10:40:52 +00:00
|
|
|
"github.com/status-im/status-go/protocol/tt"
|
|
|
|
"github.com/status-im/status-go/t/helpers"
|
|
|
|
"github.com/status-im/status-go/wakuv2/common"
|
2022-12-09 16:16:21 +00:00
|
|
|
)
|
|
|
|
|
2024-07-11 14:37:16 +00:00
|
|
|
var testStoreENRBootstrap = "enrtree://AI4W5N5IFEUIHF5LESUAOSMV6TKWF2MB6GU2YK7PU4TYUGUNOCEPW@store.staging.status.nodes.status.im"
|
|
|
|
var testBootENRBootstrap = "enrtree://AMOJVZX4V6EXP7NTJPMAYJYST2QP6AJXYW76IU6VGJS7UVSNDYZG4@boot.staging.status.nodes.status.im"
|
2022-12-09 16:16:21 +00:00
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
func setDefaultConfig(config *Config, lightMode bool) {
|
|
|
|
config.ClusterID = 16
|
|
|
|
|
|
|
|
if lightMode {
|
|
|
|
config.EnablePeerExchangeClient = true
|
|
|
|
config.LightClient = true
|
|
|
|
config.EnableDiscV5 = false
|
|
|
|
} else {
|
|
|
|
config.EnableDiscV5 = true
|
|
|
|
config.EnablePeerExchangeServer = true
|
|
|
|
config.LightClient = false
|
|
|
|
config.EnablePeerExchangeClient = false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-12-09 16:16:21 +00:00
|
|
|
func TestDiscoveryV5(t *testing.T) {
|
|
|
|
config := &Config{}
|
2024-06-14 12:41:45 +00:00
|
|
|
setDefaultConfig(config, false)
|
2024-06-28 18:12:12 +00:00
|
|
|
config.DiscV5BootstrapNodes = []string{testStoreENRBootstrap}
|
2022-12-09 16:16:21 +00:00
|
|
|
config.DiscoveryLimit = 20
|
2024-06-14 12:41:45 +00:00
|
|
|
w, err := New(nil, "shards.staging", config, nil, nil, nil, nil, nil)
|
2022-12-09 16:16:21 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
require.NoError(t, w.Start())
|
|
|
|
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
|
|
|
if len(w.Peers()) == 0 {
|
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
require.NotEqual(t, 0, len(w.Peers()))
|
|
|
|
require.NoError(t, w.Stop())
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestRestartDiscoveryV5(t *testing.T) {
|
|
|
|
config := &Config{}
|
2024-06-14 12:41:45 +00:00
|
|
|
setDefaultConfig(config, false)
|
2022-12-09 16:16:21 +00:00
|
|
|
// Use wrong discv5 bootstrap address, to simulate being offline
|
|
|
|
config.DiscV5BootstrapNodes = []string{"enrtree://AOGECG2SPND25EEFMAJ5WF3KSGJNSGV356DSTL2YVLLZWIV6SAYBM@1.1.1.2"}
|
|
|
|
config.DiscoveryLimit = 20
|
|
|
|
config.UDPPort = 9002
|
2024-06-06 13:52:51 +00:00
|
|
|
config.ClusterID = 16
|
2024-05-15 23:15:22 +00:00
|
|
|
w, err := New(nil, "", config, nil, nil, nil, nil, nil)
|
2022-12-09 16:16:21 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
require.NoError(t, w.Start())
|
|
|
|
require.False(t, w.seededBootnodesForDiscV5)
|
|
|
|
|
|
|
|
options := func(b *backoff.ExponentialBackOff) {
|
|
|
|
b.MaxElapsedTime = 2 * time.Second
|
|
|
|
}
|
|
|
|
|
|
|
|
// Sanity check, not great, but it's probably helpful
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
|
|
|
if len(w.Peers()) == 0 {
|
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, options)
|
|
|
|
|
|
|
|
require.Error(t, err)
|
|
|
|
|
2024-06-28 18:12:12 +00:00
|
|
|
w.discV5BootstrapNodes = []string{testStoreENRBootstrap}
|
2022-12-09 16:16:21 +00:00
|
|
|
|
|
|
|
options = func(b *backoff.ExponentialBackOff) {
|
2023-11-01 15:47:41 +00:00
|
|
|
b.MaxElapsedTime = 90 * time.Second
|
2022-12-09 16:16:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
|
|
|
if len(w.Peers()) == 0 {
|
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, options)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
require.True(t, w.seededBootnodesForDiscV5)
|
|
|
|
require.NotEqual(t, 0, len(w.Peers()))
|
|
|
|
require.NoError(t, w.Stop())
|
|
|
|
}
|
2023-02-22 21:58:17 +00:00
|
|
|
|
2024-07-01 15:37:54 +00:00
|
|
|
func TestRelayPeers(t *testing.T) {
|
|
|
|
config := &Config{}
|
|
|
|
setDefaultConfig(config, false)
|
|
|
|
w, err := New(nil, "", config, nil, nil, nil, nil, nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w.Start())
|
|
|
|
_, err = w.RelayPeersByTopic(config.DefaultShardPubsubTopic)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Ensure function returns an error for lightclient
|
|
|
|
config = &Config{}
|
|
|
|
config.ClusterID = 16
|
|
|
|
config.LightClient = true
|
|
|
|
w, err = New(nil, "", config, nil, nil, nil, nil, nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w.Start())
|
|
|
|
_, err = w.RelayPeersByTopic(config.DefaultShardPubsubTopic)
|
|
|
|
require.Error(t, err)
|
|
|
|
}
|
|
|
|
|
2023-02-22 21:58:17 +00:00
|
|
|
func TestBasicWakuV2(t *testing.T) {
|
2024-07-05 10:00:33 +00:00
|
|
|
t.Skip("flaky test")
|
|
|
|
|
2024-07-11 14:37:16 +00:00
|
|
|
enrTreeAddress := testStoreENRBootstrap
|
2023-02-22 21:58:17 +00:00
|
|
|
envEnrTreeAddress := os.Getenv("ENRTREE_ADDRESS")
|
|
|
|
if envEnrTreeAddress != "" {
|
|
|
|
enrTreeAddress = envEnrTreeAddress
|
|
|
|
}
|
|
|
|
|
|
|
|
config := &Config{}
|
2024-06-14 12:41:45 +00:00
|
|
|
setDefaultConfig(config, false)
|
2023-02-22 21:58:17 +00:00
|
|
|
config.Port = 0
|
|
|
|
config.DiscV5BootstrapNodes = []string{enrTreeAddress}
|
|
|
|
config.DiscoveryLimit = 20
|
|
|
|
config.WakuNodes = []string{enrTreeAddress}
|
2024-05-15 23:15:22 +00:00
|
|
|
w, err := New(nil, "", config, nil, nil, nil, nil, nil)
|
2023-02-22 21:58:17 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w.Start())
|
|
|
|
|
|
|
|
// DNSDiscovery
|
|
|
|
ctx, cancel := context.WithTimeout(context.TODO(), 30*time.Second)
|
|
|
|
defer cancel()
|
|
|
|
|
|
|
|
discoveredNodes, err := dnsdisc.RetrieveNodes(ctx, enrTreeAddress)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Peer used for retrieving history
|
|
|
|
r, err := rand.Int(rand.Reader, big.NewInt(int64(len(discoveredNodes))))
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
storeNode := discoveredNodes[int(r.Int64())]
|
|
|
|
|
2023-11-01 15:47:41 +00:00
|
|
|
options := func(b *backoff.ExponentialBackOff) {
|
|
|
|
b.MaxElapsedTime = 30 * time.Second
|
|
|
|
}
|
2023-02-22 21:58:17 +00:00
|
|
|
|
2023-11-01 15:47:41 +00:00
|
|
|
// Sanity check, not great, but it's probably helpful
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
2024-06-14 12:41:45 +00:00
|
|
|
if len(w.Peers()) < 2 {
|
2023-11-01 15:47:41 +00:00
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, options)
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
2023-02-22 21:58:17 +00:00
|
|
|
|
|
|
|
filter := &common.Filter{
|
2024-06-06 13:52:51 +00:00
|
|
|
PubsubTopic: config.DefaultShardPubsubTopic,
|
2023-09-13 10:50:23 +00:00
|
|
|
Messages: common.NewMemoryMessageStore(),
|
2024-07-15 15:25:12 +00:00
|
|
|
ContentTopics: common.NewTopicSetFromBytes([][]byte{{1, 2, 3, 4}}),
|
2023-02-22 21:58:17 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
_, err = w.Subscribe(filter)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
msgTimestamp := w.timestamp()
|
2023-09-13 10:50:23 +00:00
|
|
|
contentTopic := maps.Keys(filter.ContentTopics)[0]
|
2023-02-22 21:58:17 +00:00
|
|
|
|
2024-06-06 13:52:51 +00:00
|
|
|
time.Sleep(2 * time.Second)
|
|
|
|
|
|
|
|
_, err = w.Send(config.DefaultShardPubsubTopic, &pb.WakuMessage{
|
2023-02-22 21:58:17 +00:00
|
|
|
Payload: []byte{1, 2, 3, 4, 5},
|
|
|
|
ContentTopic: contentTopic.ContentTopic(),
|
2023-12-05 04:29:27 +00:00
|
|
|
Version: proto.Uint32(0),
|
|
|
|
Timestamp: &msgTimestamp,
|
2023-02-22 21:58:17 +00:00
|
|
|
})
|
2024-06-14 12:41:45 +00:00
|
|
|
|
2023-02-22 21:58:17 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
|
|
|
|
messages := filter.Retrieve()
|
|
|
|
require.Len(t, messages, 1)
|
|
|
|
|
|
|
|
timestampInSeconds := msgTimestamp / int64(time.Second)
|
2023-10-19 20:32:55 +00:00
|
|
|
marginInSeconds := 20
|
|
|
|
|
2023-11-01 15:47:41 +00:00
|
|
|
options = func(b *backoff.ExponentialBackOff) {
|
2023-10-19 20:32:55 +00:00
|
|
|
b.MaxElapsedTime = 60 * time.Second
|
|
|
|
b.InitialInterval = 500 * time.Millisecond
|
|
|
|
}
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
2024-06-06 13:52:51 +00:00
|
|
|
_, envelopeCount, err := w.Query(
|
2024-05-23 12:26:21 +00:00
|
|
|
context.Background(),
|
|
|
|
storeNode.PeerID,
|
2024-06-06 13:52:51 +00:00
|
|
|
legacy_store.Query{
|
|
|
|
PubsubTopic: config.DefaultShardPubsubTopic,
|
|
|
|
ContentTopics: []string{contentTopic.ContentTopic()},
|
|
|
|
StartTime: proto.Int64((timestampInSeconds - int64(marginInSeconds)) * int64(time.Second)),
|
|
|
|
EndTime: proto.Int64((timestampInSeconds + int64(marginInSeconds)) * int64(time.Second)),
|
|
|
|
},
|
|
|
|
nil,
|
2024-05-23 12:26:21 +00:00
|
|
|
[]legacy_store.HistoryRequestOption{},
|
2024-06-06 13:52:51 +00:00
|
|
|
false,
|
2024-05-23 12:26:21 +00:00
|
|
|
)
|
2024-06-06 13:52:51 +00:00
|
|
|
if err != nil || envelopeCount == 0 {
|
2023-10-19 20:32:55 +00:00
|
|
|
// in case of failure extend timestamp margin up to 40secs
|
|
|
|
if marginInSeconds < 40 {
|
|
|
|
marginInSeconds += 5
|
|
|
|
}
|
|
|
|
return errors.New("no messages received from store node")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, options)
|
2023-02-22 21:58:17 +00:00
|
|
|
require.NoError(t, err)
|
2023-06-07 09:02:19 +00:00
|
|
|
|
|
|
|
require.NoError(t, w.Stop())
|
|
|
|
}
|
|
|
|
|
2023-12-07 00:28:08 +00:00
|
|
|
type mapResolver map[string]string
|
|
|
|
|
|
|
|
func (mr mapResolver) LookupTXT(ctx context.Context, name string) ([]string, error) {
|
|
|
|
if record, ok := mr[name]; ok {
|
|
|
|
return []string{record}, nil
|
|
|
|
}
|
|
|
|
return nil, errors.New("not found")
|
|
|
|
}
|
|
|
|
|
|
|
|
var signingKeyForTesting, _ = crypto.ToECDSA(hexutil.MustDecode("0xdc599867fc513f8f5e2c2c9c489cde5e71362d1d9ec6e693e0de063236ed1240"))
|
|
|
|
|
|
|
|
func makeTestTree(domain string, nodes []*enode.Node, links []string) (*ethdnsdisc.Tree, string) {
|
|
|
|
tree, err := ethdnsdisc.MakeTree(1, nodes, links)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
url, err := tree.Sign(signingKeyForTesting, domain)
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
return tree, url
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestPeerExchange(t *testing.T) {
|
2024-02-27 09:24:34 +00:00
|
|
|
logger, err := zap.NewDevelopment()
|
|
|
|
require.NoError(t, err)
|
2023-12-07 00:28:08 +00:00
|
|
|
// start node which serve as PeerExchange server
|
|
|
|
config := &Config{}
|
2024-06-06 13:52:51 +00:00
|
|
|
config.ClusterID = 16
|
2023-12-07 00:28:08 +00:00
|
|
|
config.EnableDiscV5 = true
|
2024-02-27 09:24:34 +00:00
|
|
|
config.EnablePeerExchangeServer = true
|
|
|
|
config.EnablePeerExchangeClient = false
|
2024-05-15 23:15:22 +00:00
|
|
|
pxServerNode, err := New(nil, "", config, logger.Named("pxServerNode"), nil, nil, nil, nil)
|
2023-12-07 00:28:08 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, pxServerNode.Start())
|
|
|
|
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
|
|
|
|
// start node that will be discovered by PeerExchange
|
|
|
|
config = &Config{}
|
2024-06-06 13:52:51 +00:00
|
|
|
config.ClusterID = 16
|
2023-12-07 00:28:08 +00:00
|
|
|
config.EnableDiscV5 = true
|
2024-02-27 09:24:34 +00:00
|
|
|
config.EnablePeerExchangeServer = false
|
|
|
|
config.EnablePeerExchangeClient = false
|
2023-12-07 00:28:08 +00:00
|
|
|
config.DiscV5BootstrapNodes = []string{pxServerNode.node.ENR().String()}
|
2024-05-15 23:15:22 +00:00
|
|
|
discV5Node, err := New(nil, "", config, logger.Named("discV5Node"), nil, nil, nil, nil)
|
2023-12-07 00:28:08 +00:00
|
|
|
require.NoError(t, err)
|
2024-02-27 09:24:34 +00:00
|
|
|
require.NoError(t, discV5Node.Start())
|
2023-12-07 00:28:08 +00:00
|
|
|
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
|
|
|
|
// start light node which use PeerExchange to discover peers
|
|
|
|
enrNodes := []*enode.Node{pxServerNode.node.ENR()}
|
|
|
|
tree, url := makeTestTree("n", enrNodes, nil)
|
|
|
|
resolver := mapResolver(tree.ToTXT("n"))
|
|
|
|
|
|
|
|
config = &Config{}
|
2024-06-06 13:52:51 +00:00
|
|
|
config.ClusterID = 16
|
2024-02-27 09:24:34 +00:00
|
|
|
config.EnablePeerExchangeServer = false
|
|
|
|
config.EnablePeerExchangeClient = true
|
2023-12-07 00:28:08 +00:00
|
|
|
config.LightClient = true
|
|
|
|
config.Resolver = resolver
|
|
|
|
|
|
|
|
config.WakuNodes = []string{url}
|
2024-05-15 23:15:22 +00:00
|
|
|
lightNode, err := New(nil, "", config, logger.Named("lightNode"), nil, nil, nil, nil)
|
2023-12-07 00:28:08 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, lightNode.Start())
|
|
|
|
|
|
|
|
// Sanity check, not great, but it's probably helpful
|
|
|
|
options := func(b *backoff.ExponentialBackOff) {
|
|
|
|
b.MaxElapsedTime = 30 * time.Second
|
|
|
|
}
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
2024-03-04 08:06:16 +00:00
|
|
|
// we should not use lightNode.Peers() here as it only indicates peers that are connected right now,
|
|
|
|
// in light client mode,the peer will be closed via `w.node.Host().Network().ClosePeer(peerInfo.ID)`
|
|
|
|
// after invoking identifyAndConnect, instead, we should check the peerStore, peers from peerStore
|
|
|
|
// won't get deleted especially if they are statically added.
|
|
|
|
if len(lightNode.node.Host().Peerstore().Peers()) == 2 {
|
2023-12-07 00:28:08 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}, options)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2024-07-12 13:21:21 +00:00
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
require.NoError(t, discV5Node.node.PeerExchange().Request(ctx, 1))
|
|
|
|
require.Error(t, discV5Node.node.PeerExchange().Request(ctx, 1)) //should fail due to rate limit
|
|
|
|
|
2023-12-07 00:28:08 +00:00
|
|
|
require.NoError(t, lightNode.Stop())
|
|
|
|
require.NoError(t, pxServerNode.Stop())
|
2024-02-27 09:24:34 +00:00
|
|
|
require.NoError(t, discV5Node.Stop())
|
2023-12-07 00:28:08 +00:00
|
|
|
}
|
|
|
|
|
2023-06-07 09:02:19 +00:00
|
|
|
func TestWakuV2Filter(t *testing.T) {
|
2024-07-05 10:00:33 +00:00
|
|
|
t.Skip("flaky test")
|
2024-06-28 08:02:41 +00:00
|
|
|
|
2024-06-28 18:12:12 +00:00
|
|
|
enrTreeAddress := testBootENRBootstrap
|
2023-06-07 09:02:19 +00:00
|
|
|
envEnrTreeAddress := os.Getenv("ENRTREE_ADDRESS")
|
|
|
|
if envEnrTreeAddress != "" {
|
|
|
|
enrTreeAddress = envEnrTreeAddress
|
|
|
|
}
|
|
|
|
config := &Config{}
|
2024-06-14 12:41:45 +00:00
|
|
|
setDefaultConfig(config, true)
|
2024-06-28 18:12:12 +00:00
|
|
|
config.EnablePeerExchangeClient = false
|
2023-06-07 09:02:19 +00:00
|
|
|
config.Port = 0
|
|
|
|
config.MinPeersForFilter = 2
|
2024-06-14 12:41:45 +00:00
|
|
|
|
2023-06-07 09:02:19 +00:00
|
|
|
config.DiscV5BootstrapNodes = []string{enrTreeAddress}
|
|
|
|
config.DiscoveryLimit = 20
|
|
|
|
config.WakuNodes = []string{enrTreeAddress}
|
2024-06-14 12:41:45 +00:00
|
|
|
w, err := New(nil, "", config, nil, nil, nil, nil, nil)
|
2023-06-07 09:02:19 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w.Start())
|
2024-07-04 05:04:53 +00:00
|
|
|
w.filterManager.filterSubBatchDuration = 1 * time.Second
|
2024-07-01 15:37:54 +00:00
|
|
|
|
2023-11-01 15:47:41 +00:00
|
|
|
options := func(b *backoff.ExponentialBackOff) {
|
|
|
|
b.MaxElapsedTime = 10 * time.Second
|
|
|
|
}
|
2024-06-14 12:41:45 +00:00
|
|
|
time.Sleep(10 * time.Second) //TODO: Check if we can remove this sleep.
|
2024-06-28 18:12:12 +00:00
|
|
|
|
2023-11-01 15:47:41 +00:00
|
|
|
// Sanity check, not great, but it's probably helpful
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
2024-06-14 12:41:45 +00:00
|
|
|
peers, err := w.node.PeerManager().FilterPeersByProto(nil, nil, filter.FilterSubscribeID_v20beta1)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if len(peers) < 2 {
|
2023-11-01 15:47:41 +00:00
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, options)
|
2023-11-01 15:47:50 +00:00
|
|
|
require.NoError(t, err)
|
2024-06-14 12:41:45 +00:00
|
|
|
testPubsubTopic := "/waku/2/rs/16/32"
|
2024-06-28 18:12:12 +00:00
|
|
|
contentTopicBytes := make([]byte, 4)
|
|
|
|
_, err = rand.Read(contentTopicBytes)
|
|
|
|
require.NoError(t, err)
|
2023-06-07 09:02:19 +00:00
|
|
|
filter := &common.Filter{
|
2023-09-13 10:50:23 +00:00
|
|
|
Messages: common.NewMemoryMessageStore(),
|
2024-06-14 12:41:45 +00:00
|
|
|
PubsubTopic: testPubsubTopic,
|
2024-06-28 18:12:12 +00:00
|
|
|
ContentTopics: common.NewTopicSetFromBytes([][]byte{contentTopicBytes}),
|
2023-06-07 09:02:19 +00:00
|
|
|
}
|
|
|
|
|
2024-07-04 05:04:53 +00:00
|
|
|
fID, err := w.Subscribe(filter)
|
2023-06-07 09:02:19 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
msgTimestamp := w.timestamp()
|
2023-09-13 10:50:23 +00:00
|
|
|
contentTopic := maps.Keys(filter.ContentTopics)[0]
|
2023-06-07 09:02:19 +00:00
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
_, err = w.Send(testPubsubTopic, &pb.WakuMessage{
|
2023-06-07 09:02:19 +00:00
|
|
|
Payload: []byte{1, 2, 3, 4, 5},
|
|
|
|
ContentTopic: contentTopic.ContentTopic(),
|
2023-12-05 04:29:27 +00:00
|
|
|
Version: proto.Uint32(0),
|
|
|
|
Timestamp: &msgTimestamp,
|
2023-06-07 09:02:19 +00:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
2024-06-14 12:41:45 +00:00
|
|
|
time.Sleep(5 * time.Second)
|
2023-06-07 09:02:19 +00:00
|
|
|
|
2023-09-13 10:50:23 +00:00
|
|
|
// Ensure there is at least 1 active filter subscription
|
|
|
|
subscriptions := w.node.FilterLightnode().Subscriptions()
|
|
|
|
require.Greater(t, len(subscriptions), 0)
|
|
|
|
|
2023-06-07 09:02:19 +00:00
|
|
|
messages := filter.Retrieve()
|
|
|
|
require.Len(t, messages, 1)
|
|
|
|
|
|
|
|
// Mock peers going down
|
2024-06-14 12:41:45 +00:00
|
|
|
_, err = w.node.FilterLightnode().UnsubscribeWithSubscription(w.ctx, subscriptions[0])
|
|
|
|
require.NoError(t, err)
|
2023-06-07 09:02:19 +00:00
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
time.Sleep(10 * time.Second)
|
2023-09-13 10:50:23 +00:00
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
// Ensure there is at least 1 active filter subscription
|
|
|
|
subscriptions = w.node.FilterLightnode().Subscriptions()
|
|
|
|
require.Greater(t, len(subscriptions), 0)
|
2023-06-07 09:02:19 +00:00
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
// Ensure that messages are retrieved with a fresh sub
|
|
|
|
_, err = w.Send(testPubsubTopic, &pb.WakuMessage{
|
|
|
|
Payload: []byte{1, 2, 3, 4, 5, 6},
|
|
|
|
ContentTopic: contentTopic.ContentTopic(),
|
|
|
|
Version: proto.Uint32(0),
|
|
|
|
Timestamp: &msgTimestamp,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
2023-06-07 09:02:19 +00:00
|
|
|
time.Sleep(10 * time.Second)
|
|
|
|
|
2024-06-14 12:41:45 +00:00
|
|
|
messages = filter.Retrieve()
|
|
|
|
require.Len(t, messages, 1)
|
2024-07-04 05:04:53 +00:00
|
|
|
err = w.Unsubscribe(context.Background(), fID)
|
|
|
|
require.NoError(t, err)
|
2023-06-07 09:02:19 +00:00
|
|
|
require.NoError(t, w.Stop())
|
2023-02-22 21:58:17 +00:00
|
|
|
}
|
2023-11-24 10:13:26 +00:00
|
|
|
|
|
|
|
func TestWakuV2Store(t *testing.T) {
|
|
|
|
// Configuration for the first Waku node
|
|
|
|
config1 := &Config{
|
2024-07-15 15:25:12 +00:00
|
|
|
Port: 0,
|
|
|
|
ClusterID: 16,
|
|
|
|
EnableDiscV5: false,
|
|
|
|
DiscoveryLimit: 20,
|
|
|
|
EnableStore: false,
|
|
|
|
StoreCapacity: 100,
|
|
|
|
StoreSeconds: 3600,
|
2023-11-24 10:13:26 +00:00
|
|
|
}
|
|
|
|
w1PeersCh := make(chan []string, 100) // buffered not to block on the send side
|
|
|
|
|
|
|
|
// Start the first Waku node
|
2024-05-15 23:15:22 +00:00
|
|
|
w1, err := New(nil, "", config1, nil, nil, nil, nil, func(cs types.ConnStatus) {
|
2023-11-24 10:13:26 +00:00
|
|
|
w1PeersCh <- maps.Keys(cs.Peers)
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w1.Start())
|
|
|
|
defer func() {
|
|
|
|
require.NoError(t, w1.Stop())
|
|
|
|
close(w1PeersCh)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Configuration for the second Waku node
|
|
|
|
sql2, err := helpers.SetupTestMemorySQLDB(appdatabase.DbInitializer{})
|
|
|
|
require.NoError(t, err)
|
|
|
|
config2 := &Config{
|
2024-07-15 15:25:12 +00:00
|
|
|
Port: 0,
|
|
|
|
ClusterID: 16,
|
|
|
|
EnableDiscV5: false,
|
|
|
|
DiscoveryLimit: 20,
|
|
|
|
EnableStore: true,
|
|
|
|
StoreCapacity: 100,
|
|
|
|
StoreSeconds: 3600,
|
2023-11-24 10:13:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Start the second Waku node
|
2024-05-15 23:15:22 +00:00
|
|
|
w2, err := New(nil, "", config2, nil, sql2, nil, nil, nil)
|
2023-11-24 10:13:26 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w2.Start())
|
|
|
|
w2EnvelopeCh := make(chan common.EnvelopeEvent, 100)
|
|
|
|
w2.SubscribeEnvelopeEvents(w2EnvelopeCh)
|
|
|
|
defer func() {
|
|
|
|
require.NoError(t, w2.Stop())
|
|
|
|
close(w2EnvelopeCh)
|
|
|
|
}()
|
|
|
|
|
|
|
|
// Connect the two nodes directly
|
|
|
|
peer2Addr := w2.node.ListenAddresses()[0].String()
|
|
|
|
err = w1.node.DialPeer(context.Background(), peer2Addr)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
waitForPeerConnection(t, w2.node.ID(), w1PeersCh)
|
|
|
|
|
|
|
|
// Create a filter for the second node to catch messages
|
|
|
|
filter := &common.Filter{
|
|
|
|
Messages: common.NewMemoryMessageStore(),
|
2024-06-06 13:52:51 +00:00
|
|
|
PubsubTopic: config2.DefaultShardPubsubTopic,
|
2023-11-24 10:13:26 +00:00
|
|
|
ContentTopics: common.NewTopicSetFromBytes([][]byte{{1, 2, 3, 4}}),
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = w2.Subscribe(filter)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2024-07-11 18:36:34 +00:00
|
|
|
time.Sleep(2 * time.Second)
|
|
|
|
|
2023-11-24 10:13:26 +00:00
|
|
|
// Send a message from the first node
|
|
|
|
msgTimestamp := w1.CurrentTime().UnixNano()
|
|
|
|
contentTopic := maps.Keys(filter.ContentTopics)[0]
|
2024-06-06 13:52:51 +00:00
|
|
|
_, err = w1.Send(config1.DefaultShardPubsubTopic, &pb.WakuMessage{
|
2023-11-24 10:13:26 +00:00
|
|
|
Payload: []byte{1, 2, 3, 4, 5},
|
|
|
|
ContentTopic: contentTopic.ContentTopic(),
|
2023-12-05 04:29:27 +00:00
|
|
|
Version: proto.Uint32(0),
|
|
|
|
Timestamp: &msgTimestamp,
|
2023-11-24 10:13:26 +00:00
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
waitForEnvelope(t, contentTopic.ContentTopic(), w2EnvelopeCh)
|
|
|
|
|
|
|
|
// Retrieve the message from the second node's filter
|
|
|
|
messages := filter.Retrieve()
|
|
|
|
require.Len(t, messages, 1)
|
|
|
|
|
|
|
|
timestampInSeconds := msgTimestamp / int64(time.Second)
|
|
|
|
marginInSeconds := 5
|
|
|
|
// Query the second node's store for the message
|
2024-06-06 13:52:51 +00:00
|
|
|
_, envelopeCount, err := w1.Query(
|
2024-05-23 12:26:21 +00:00
|
|
|
context.Background(),
|
|
|
|
w2.node.Host().ID(),
|
2024-06-06 13:52:51 +00:00
|
|
|
legacy_store.Query{
|
|
|
|
PubsubTopic: config1.DefaultShardPubsubTopic,
|
|
|
|
ContentTopics: []string{contentTopic.ContentTopic()},
|
|
|
|
StartTime: proto.Int64((timestampInSeconds - int64(marginInSeconds)) * int64(time.Second)),
|
|
|
|
EndTime: proto.Int64((timestampInSeconds + int64(marginInSeconds)) * int64(time.Second)),
|
|
|
|
},
|
|
|
|
nil,
|
2024-05-23 12:26:21 +00:00
|
|
|
[]legacy_store.HistoryRequestOption{},
|
2024-06-06 13:52:51 +00:00
|
|
|
false,
|
2024-05-23 12:26:21 +00:00
|
|
|
)
|
2023-11-24 10:13:26 +00:00
|
|
|
require.NoError(t, err)
|
2024-06-06 13:52:51 +00:00
|
|
|
require.True(t, envelopeCount > 0, "no messages received from store node")
|
2023-11-24 10:13:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func waitForPeerConnection(t *testing.T, peerID string, peerCh chan []string) {
|
2024-07-12 10:34:56 +00:00
|
|
|
waitForPeerConnectionWithTimeout(t, peerID, peerCh, 3*time.Second)
|
|
|
|
}
|
|
|
|
|
|
|
|
func waitForPeerConnectionWithTimeout(t *testing.T, peerID string, peerCh chan []string, timeout time.Duration) {
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), timeout)
|
2023-11-24 10:13:26 +00:00
|
|
|
defer cancel()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case peers := <-peerCh:
|
|
|
|
for _, p := range peers {
|
|
|
|
if p == peerID {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
case <-ctx.Done():
|
|
|
|
require.Fail(t, "timed out waiting for peer "+peerID)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func waitForEnvelope(t *testing.T, contentTopic string, envCh chan common.EnvelopeEvent) {
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 2*time.Second)
|
|
|
|
defer cancel()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case env := <-envCh:
|
|
|
|
if env.Topic.ContentTopic() == contentTopic {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
case <-ctx.Done():
|
|
|
|
require.Fail(t, "timed out waiting for envelope's topic "+contentTopic)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2024-06-11 07:45:01 +00:00
|
|
|
|
|
|
|
func TestConfirmMessageDelivered(t *testing.T) {
|
|
|
|
aliceConfig := &Config{}
|
|
|
|
aliceNode, err := New(nil, "", aliceConfig, nil, nil, nil, nil, nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, aliceNode.Start())
|
|
|
|
|
|
|
|
bobConfig := &Config{}
|
|
|
|
bobNode, err := New(nil, "", bobConfig, nil, nil, nil, nil, nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, bobNode.Start())
|
|
|
|
|
|
|
|
addrs := aliceNode.ListenAddresses()
|
|
|
|
require.Greater(t, len(addrs), 0)
|
|
|
|
_, err = bobNode.AddRelayPeer(addrs[0])
|
|
|
|
require.NoError(t, err)
|
|
|
|
err = bobNode.DialPeer(addrs[0])
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
filter := &common.Filter{
|
|
|
|
Messages: common.NewMemoryMessageStore(),
|
|
|
|
ContentTopics: common.NewTopicSetFromBytes([][]byte{[]byte{1, 2, 3, 4}}),
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = aliceNode.Subscribe(filter)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
msgTimestamp := aliceNode.timestamp()
|
|
|
|
contentTopic := maps.Keys(filter.ContentTopics)[0]
|
|
|
|
|
2024-07-15 15:25:12 +00:00
|
|
|
_, err = aliceNode.Send(shard.DefaultShardPubsubTopic(), &pb.WakuMessage{
|
2024-06-11 07:45:01 +00:00
|
|
|
Payload: []byte{1, 2, 3, 4, 5},
|
|
|
|
ContentTopic: contentTopic.ContentTopic(),
|
|
|
|
Version: proto.Uint32(0),
|
|
|
|
Timestamp: &msgTimestamp,
|
|
|
|
Ephemeral: proto.Bool(false),
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
|
|
|
|
messages := filter.Retrieve()
|
|
|
|
require.Len(t, messages, 1)
|
|
|
|
|
|
|
|
require.Len(t, aliceNode.sendMsgIDs, 1)
|
|
|
|
for _, msgs := range aliceNode.sendMsgIDs {
|
|
|
|
require.Len(t, msgs, 1)
|
|
|
|
for hash := range msgs {
|
|
|
|
require.Equal(t, hash, messages[0].Hash())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
aliceNode.ConfirmMessageDelivered([]ethcommon.Hash{messages[0].Hash()})
|
|
|
|
require.Len(t, aliceNode.sendMsgIDs, 0)
|
|
|
|
|
|
|
|
require.NoError(t, aliceNode.Stop())
|
|
|
|
require.NoError(t, bobNode.Stop())
|
|
|
|
}
|
2024-06-28 13:54:48 +00:00
|
|
|
|
|
|
|
func TestOnlineChecker(t *testing.T) {
|
|
|
|
w, err := New(nil, "shards.staging", nil, nil, nil, nil, nil, nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.False(t, w.onlineChecker.IsOnline())
|
|
|
|
|
|
|
|
w.ConnectionChanged(connection.State{Offline: false})
|
|
|
|
require.True(t, w.onlineChecker.IsOnline())
|
|
|
|
|
|
|
|
wg := sync.WaitGroup{}
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
|
|
|
defer wg.Done()
|
2024-07-11 18:36:34 +00:00
|
|
|
<-w.goingOnline
|
2024-06-28 13:54:48 +00:00
|
|
|
require.True(t, true)
|
|
|
|
}()
|
|
|
|
|
|
|
|
time.Sleep(100 * time.Millisecond)
|
|
|
|
|
|
|
|
w.ConnectionChanged(connection.State{Offline: true})
|
|
|
|
require.False(t, w.onlineChecker.IsOnline())
|
|
|
|
|
|
|
|
// Test lightnode online checker
|
|
|
|
config := &Config{}
|
|
|
|
config.ClusterID = 16
|
|
|
|
config.LightClient = true
|
|
|
|
lightNode, err := New(nil, "shards.staging", config, nil, nil, nil, nil, nil)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = lightNode.Start()
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
require.False(t, lightNode.onlineChecker.IsOnline())
|
|
|
|
|
|
|
|
lightNode.filterManager.addFilter("test", &common.Filter{})
|
|
|
|
|
|
|
|
}
|
2024-07-12 10:34:56 +00:00
|
|
|
|
|
|
|
func TestLightpushRateLimit(t *testing.T) {
|
|
|
|
logger, err := zap.NewDevelopment()
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
config0 := &Config{}
|
|
|
|
setDefaultConfig(config0, false)
|
|
|
|
w0PeersCh := make(chan []string, 5) // buffered not to block on the send side
|
|
|
|
|
|
|
|
// Start the relayu node
|
|
|
|
w0, err := New(nil, "", config0, logger.Named("relayNode"), nil, nil, nil, func(cs types.ConnStatus) {
|
|
|
|
w0PeersCh <- maps.Keys(cs.Peers)
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w0.Start())
|
|
|
|
defer func() {
|
|
|
|
require.NoError(t, w0.Stop())
|
|
|
|
close(w0PeersCh)
|
|
|
|
}()
|
|
|
|
|
|
|
|
contentTopics := common.NewTopicSetFromBytes([][]byte{{1, 2, 3, 4}})
|
|
|
|
filter := &common.Filter{
|
|
|
|
PubsubTopic: config0.DefaultShardPubsubTopic,
|
|
|
|
Messages: common.NewMemoryMessageStore(),
|
|
|
|
ContentTopics: contentTopics,
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err = w0.Subscribe(filter)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
config1 := &Config{}
|
|
|
|
setDefaultConfig(config1, false)
|
|
|
|
w1PeersCh := make(chan []string, 5) // buffered not to block on the send side
|
|
|
|
|
|
|
|
// Start the full node
|
|
|
|
w1, err := New(nil, "", config1, logger.Named("fullNode"), nil, nil, nil, func(cs types.ConnStatus) {
|
|
|
|
w1PeersCh <- maps.Keys(cs.Peers)
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w1.Start())
|
|
|
|
defer func() {
|
|
|
|
require.NoError(t, w1.Stop())
|
|
|
|
close(w1PeersCh)
|
|
|
|
}()
|
|
|
|
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
defer cancel()
|
|
|
|
//Connect the relay peer and full node
|
|
|
|
err = w1.node.DialPeer(ctx, w0.node.ListenAddresses()[0].String())
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
err = tt.RetryWithBackOff(func() error {
|
|
|
|
if len(w1.Peers()) == 0 {
|
|
|
|
return errors.New("no peers discovered")
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
config2 := &Config{}
|
|
|
|
setDefaultConfig(config2, true)
|
|
|
|
w2PeersCh := make(chan []string, 5) // buffered not to block on the send side
|
|
|
|
|
|
|
|
// Start the light node
|
|
|
|
w2, err := New(nil, "", config2, logger.Named("lightNode"), nil, nil, nil, func(cs types.ConnStatus) {
|
|
|
|
w2PeersCh <- maps.Keys(cs.Peers)
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, w2.Start())
|
|
|
|
defer func() {
|
|
|
|
require.NoError(t, w2.Stop())
|
|
|
|
close(w2PeersCh)
|
|
|
|
}()
|
|
|
|
|
|
|
|
//Use this instead of DialPeer to make sure the peer is added to PeerStore and can be selected for Lighpush
|
|
|
|
w2.node.AddDiscoveredPeer(w1.PeerID(), w1.node.ListenAddresses(), wps.Static, w1.cfg.DefaultShardedPubsubTopics, w1.node.ENR(), true)
|
|
|
|
|
|
|
|
waitForPeerConnectionWithTimeout(t, w2.node.ID(), w1PeersCh, 5*time.Second)
|
|
|
|
|
|
|
|
event := make(chan common.EnvelopeEvent, 10)
|
|
|
|
w2.SubscribeEnvelopeEvents(event)
|
|
|
|
|
|
|
|
for i := range [4]int{} {
|
|
|
|
msgTimestamp := w2.timestamp()
|
|
|
|
_, err := w2.Send(config2.DefaultShardPubsubTopic, &pb.WakuMessage{
|
|
|
|
Payload: []byte{1, 2, 3, 4, 5, 6, byte(i)},
|
|
|
|
ContentTopic: maps.Keys(contentTopics)[0].ContentTopic(),
|
|
|
|
Version: proto.Uint32(0),
|
|
|
|
Timestamp: &msgTimestamp,
|
|
|
|
})
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
time.Sleep(550 * time.Millisecond)
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
messages := filter.Retrieve()
|
|
|
|
require.Len(t, messages, 2)
|
|
|
|
|
|
|
|
}
|