mirror of
https://github.com/logos-messaging/logos-messaging-nim.git
synced 2026-01-06 16:03:08 +00:00
feat(filter): split waku filter protocol into server and client
This commit is contained in:
parent
1c46b61402
commit
8fee1b9bed
@ -326,9 +326,7 @@ proc writeAndPrint(c: Chat) {.async.} =
|
|||||||
if not c.node.wakuFilter.isNil():
|
if not c.node.wakuFilter.isNil():
|
||||||
echo "unsubscribing from content filters..."
|
echo "unsubscribing from content filters..."
|
||||||
|
|
||||||
await c.node.unsubscribe(
|
await c.node.unsubscribe(pubsubTopic=DefaultTopic, contentTopics=c.contentTopic)
|
||||||
FilterRequest(contentFilters: @[ContentFilter(contentTopic: c.contentTopic)], pubSubTopic: DefaultTopic, subscribe: false)
|
|
||||||
)
|
|
||||||
|
|
||||||
echo "quitting..."
|
echo "quitting..."
|
||||||
|
|
||||||
@ -499,18 +497,16 @@ proc processInput(rfd: AsyncFD) {.async.} =
|
|||||||
|
|
||||||
if conf.filternode != "":
|
if conf.filternode != "":
|
||||||
await node.mountFilter()
|
await node.mountFilter()
|
||||||
|
await node.mountFilterClient()
|
||||||
|
|
||||||
node.wakuFilter.setPeer(parseRemotePeerInfo(conf.filternode))
|
node.setFilterPeer(parseRemotePeerInfo(conf.filternode))
|
||||||
|
|
||||||
proc filterHandler(msg: WakuMessage) {.gcsafe.} =
|
proc filterHandler(pubsubTopic: string, msg: WakuMessage) {.gcsafe.} =
|
||||||
trace "Hit filter handler", contentTopic=msg.contentTopic
|
trace "Hit filter handler", contentTopic=msg.contentTopic
|
||||||
|
|
||||||
chat.printReceivedMessage(msg)
|
chat.printReceivedMessage(msg)
|
||||||
|
|
||||||
await node.subscribe(
|
await node.subscribe(pubsubTopic=DefaultTopic, contentTopics=chat.contentTopic, filterHandler)
|
||||||
FilterRequest(contentFilters: @[ContentFilter(contentTopic: chat.contentTopic)], pubSubTopic: DefaultTopic, subscribe: true),
|
|
||||||
filterHandler
|
|
||||||
)
|
|
||||||
|
|
||||||
# Subscribe to a topic, if relay is mounted
|
# Subscribe to a topic, if relay is mounted
|
||||||
if conf.relay:
|
if conf.relay:
|
||||||
|
|||||||
@ -418,17 +418,18 @@ proc setupProtocols(node: WakuNode, conf: WakuNodeConf,
|
|||||||
return err("failed to set node waku lightpush peer: " & getCurrentExceptionMsg())
|
return err("failed to set node waku lightpush peer: " & getCurrentExceptionMsg())
|
||||||
|
|
||||||
# Filter setup. NOTE Must be mounted after relay
|
# Filter setup. NOTE Must be mounted after relay
|
||||||
if (conf.filternode != "") or (conf.filter):
|
if conf.filter:
|
||||||
try:
|
try:
|
||||||
await mountFilter(node, filterTimeout = chronos.seconds(conf.filterTimeout))
|
await mountFilter(node, filterTimeout = chronos.seconds(conf.filterTimeout))
|
||||||
except:
|
except:
|
||||||
return err("failed to mount waku filter protocol: " & getCurrentExceptionMsg())
|
return err("failed to mount waku filter protocol: " & getCurrentExceptionMsg())
|
||||||
|
|
||||||
if conf.filternode != "":
|
if conf.filternode != "":
|
||||||
try:
|
try:
|
||||||
setFilterPeer(node, conf.filternode)
|
await mountFilterClient(node)
|
||||||
except:
|
setFilterPeer(node, conf.filternode)
|
||||||
return err("failed to set node waku filter peer: " & getCurrentExceptionMsg())
|
except:
|
||||||
|
return err("failed to set node waku filter peer: " & getCurrentExceptionMsg())
|
||||||
|
|
||||||
# waku peer exchange setup
|
# waku peer exchange setup
|
||||||
if (conf.peerExchangeNode != "") or (conf.peerExchange):
|
if (conf.peerExchangeNode != "") or (conf.peerExchange):
|
||||||
|
|||||||
@ -5,7 +5,7 @@ import
|
|||||||
chronicles,
|
chronicles,
|
||||||
testutils/unittests, stew/shims/net as stewNet,
|
testutils/unittests, stew/shims/net as stewNet,
|
||||||
json_rpc/[rpcserver, rpcclient],
|
json_rpc/[rpcserver, rpcclient],
|
||||||
eth/[keys, rlp], eth/common/eth_types,
|
eth/keys, eth/common/eth_types,
|
||||||
libp2p/[builders, switch, multiaddress],
|
libp2p/[builders, switch, multiaddress],
|
||||||
libp2p/protobuf/minprotobuf,
|
libp2p/protobuf/minprotobuf,
|
||||||
libp2p/stream/[bufferstream, connection],
|
libp2p/stream/[bufferstream, connection],
|
||||||
@ -27,6 +27,7 @@ import
|
|||||||
../../waku/v2/protocol/waku_store,
|
../../waku/v2/protocol/waku_store,
|
||||||
../../waku/v2/protocol/waku_swap/waku_swap,
|
../../waku/v2/protocol/waku_swap/waku_swap,
|
||||||
../../waku/v2/protocol/waku_filter,
|
../../waku/v2/protocol/waku_filter,
|
||||||
|
../../waku/v2/protocol/waku_filter/client,
|
||||||
../../waku/v2/utils/peers,
|
../../waku/v2/utils/peers,
|
||||||
../../waku/v2/utils/time,
|
../../waku/v2/utils/time,
|
||||||
./testlib/common
|
./testlib/common
|
||||||
@ -51,7 +52,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8546)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -78,7 +79,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8547)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -150,7 +151,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8548)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -164,7 +165,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
# First see if we can retrieve messages published on the default topic (node is already subscribed)
|
# First see if we can retrieve messages published on the default topic (node is already subscribed)
|
||||||
await node2.publish(DefaultPubsubTopic, message1)
|
await node2.publish(DefaultPubsubTopic, message1)
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
var messages = await client.get_waku_v2_relay_v1_messages(DefaultPubsubTopic)
|
var messages = await client.get_waku_v2_relay_v1_messages(DefaultPubsubTopic)
|
||||||
|
|
||||||
@ -182,7 +183,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
var response = await client.post_waku_v2_relay_v1_subscriptions(@[pubSubTopic])
|
var response = await client.post_waku_v2_relay_v1_subscriptions(@[pubSubTopic])
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
check:
|
check:
|
||||||
# Node is now subscribed to pubSubTopic
|
# Node is now subscribed to pubSubTopic
|
||||||
@ -191,7 +192,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
# Now publish a message on node1 and see if we receive it on node3
|
# Now publish a message on node1 and see if we receive it on node3
|
||||||
await node1.publish(pubSubTopic, message2)
|
await node1.publish(pubSubTopic, message2)
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
messages = await client.get_waku_v2_relay_v1_messages(pubSubTopic)
|
messages = await client.get_waku_v2_relay_v1_messages(pubSubTopic)
|
||||||
|
|
||||||
@ -219,7 +220,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8549)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -273,19 +274,26 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
await node.stop()
|
await node.stop()
|
||||||
|
|
||||||
asyncTest "Filter API: subscribe/unsubscribe":
|
asyncTest "Filter API: subscribe/unsubscribe":
|
||||||
await node.start()
|
let
|
||||||
|
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
|
node1 = WakuNode.new(nodeKey1, bindIp, Port(60390))
|
||||||
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
|
node2 = WakuNode.new(nodeKey2, bindIp, Port(60392))
|
||||||
|
|
||||||
await node.mountRelay()
|
await allFutures(node1.start(), node2.start())
|
||||||
|
|
||||||
await node.mountFilter()
|
await node1.mountFilter()
|
||||||
|
await node2.mountFilterClient()
|
||||||
|
|
||||||
|
node2.setFilterPeer(node1.peerInfo.toRemotePeerInfo())
|
||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8550)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
installFilterApiHandlers(node, server, newTable[ContentTopic, seq[WakuMessage]]())
|
installFilterApiHandlers(node2, server, newTable[ContentTopic, seq[WakuMessage]]())
|
||||||
server.start()
|
server.start()
|
||||||
|
|
||||||
let client = newRpcHttpClient()
|
let client = newRpcHttpClient()
|
||||||
@ -293,109 +301,31 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
check:
|
check:
|
||||||
# Light node has not yet subscribed to any filters
|
# Light node has not yet subscribed to any filters
|
||||||
node.filters.len() == 0
|
node2.wakuFilterClient.getSubscriptionsCount() == 0
|
||||||
|
|
||||||
let contentFilters = @[ContentFilter(contentTopic: DefaultContentTopic),
|
let contentFilters = @[
|
||||||
ContentFilter(contentTopic: ContentTopic("2")),
|
ContentFilter(contentTopic: DefaultContentTopic),
|
||||||
ContentFilter(contentTopic: ContentTopic("3")),
|
ContentFilter(contentTopic: ContentTopic("2")),
|
||||||
ContentFilter(contentTopic: ContentTopic("4")),
|
ContentFilter(contentTopic: ContentTopic("3")),
|
||||||
]
|
ContentFilter(contentTopic: ContentTopic("4")),
|
||||||
var response = await client.post_waku_v2_filter_v1_subscription(contentFilters = contentFilters, topic = some(DefaultPubsubTopic))
|
]
|
||||||
|
var response = await client.post_waku_v2_filter_v1_subscription(contentFilters=contentFilters, topic=some(DefaultPubsubTopic))
|
||||||
check:
|
check:
|
||||||
# Light node has successfully subscribed to a single filter
|
|
||||||
node.filters.len() == 1
|
|
||||||
response == true
|
response == true
|
||||||
|
# Light node has successfully subscribed to 4 content topics
|
||||||
|
node2.wakuFilterClient.getSubscriptionsCount() == 4
|
||||||
|
|
||||||
response = await client.delete_waku_v2_filter_v1_subscription(contentFilters = contentFilters, topic = some(DefaultPubsubTopic))
|
response = await client.delete_waku_v2_filter_v1_subscription(contentFilters=contentFilters, topic=some(DefaultPubsubTopic))
|
||||||
|
|
||||||
check:
|
check:
|
||||||
|
response == true
|
||||||
# Light node has successfully unsubscribed from all filters
|
# Light node has successfully unsubscribed from all filters
|
||||||
node.filters.len() == 0
|
node2.wakuFilterClient.getSubscriptionsCount() == 0
|
||||||
response == true
|
|
||||||
|
|
||||||
|
## Cleanup
|
||||||
await server.stop()
|
await server.stop()
|
||||||
await server.closeWait()
|
await server.closeWait()
|
||||||
|
|
||||||
await node.stop()
|
await allFutures(node1.stop(), node2.stop())
|
||||||
|
|
||||||
asyncTest "Filter API: get latest messages":
|
|
||||||
await node.start()
|
|
||||||
|
|
||||||
# RPC server setup
|
|
||||||
let
|
|
||||||
rpcPort = Port(8545)
|
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
|
||||||
server = newRpcHttpServer([ta])
|
|
||||||
|
|
||||||
installFilterApiHandlers(node, server, newTable[ContentTopic, seq[WakuMessage]]())
|
|
||||||
server.start()
|
|
||||||
|
|
||||||
await node.mountFilter()
|
|
||||||
|
|
||||||
let client = newRpcHttpClient()
|
|
||||||
await client.connect("127.0.0.1", rpcPort, false)
|
|
||||||
|
|
||||||
# First ensure subscription exists
|
|
||||||
|
|
||||||
let sub = await client.post_waku_v2_filter_v1_subscription(contentFilters = @[ContentFilter(contentTopic: DefaultContentTopic)], topic = some(DefaultPubsubTopic))
|
|
||||||
check:
|
|
||||||
sub
|
|
||||||
|
|
||||||
# Now prime the node with some messages before tests
|
|
||||||
var
|
|
||||||
msgList = @[WakuMessage(payload: @[byte 0], contentTopic: ContentTopic("2")),
|
|
||||||
WakuMessage(payload: @[byte 1], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 2], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 3], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 4], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 5], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 6], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 7], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 8], contentTopic: DefaultContentTopic),
|
|
||||||
WakuMessage(payload: @[byte 9], contentTopic: ContentTopic("2"))]
|
|
||||||
|
|
||||||
let
|
|
||||||
filters = node.filters
|
|
||||||
requestId = toSeq(Table(filters).keys)[0]
|
|
||||||
|
|
||||||
for wakuMsg in msgList:
|
|
||||||
filters.notify(wakuMsg, requestId)
|
|
||||||
|
|
||||||
var response = await client.get_waku_v2_filter_v1_messages(DefaultContentTopic)
|
|
||||||
check:
|
|
||||||
response.len() == 8
|
|
||||||
response.allIt(it.contentTopic == DefaultContentTopic)
|
|
||||||
|
|
||||||
# No new messages
|
|
||||||
response = await client.get_waku_v2_filter_v1_messages(DefaultContentTopic)
|
|
||||||
|
|
||||||
check:
|
|
||||||
response.len() == 0
|
|
||||||
|
|
||||||
# Now ensure that no more than the preset max messages can be cached
|
|
||||||
|
|
||||||
let maxSize = filter_api.maxCache
|
|
||||||
|
|
||||||
for x in 1..(maxSize + 1):
|
|
||||||
# Try to cache 1 more than maximum allowed
|
|
||||||
filters.notify(WakuMessage(payload: @[byte x], contentTopic: DefaultContentTopic), requestId)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
response = await client.get_waku_v2_filter_v1_messages(DefaultContentTopic)
|
|
||||||
check:
|
|
||||||
# Max messages has not been exceeded
|
|
||||||
response.len == maxSize
|
|
||||||
response.allIt(it.contentTopic == DefaultContentTopic)
|
|
||||||
# Check that oldest item has been removed
|
|
||||||
response[0].payload == @[byte 2]
|
|
||||||
response[maxSize - 1].payload == @[byte (maxSize + 1)]
|
|
||||||
|
|
||||||
await server.stop()
|
|
||||||
await server.closeWait()
|
|
||||||
|
|
||||||
await node.stop()
|
|
||||||
|
|
||||||
asyncTest "Admin API: connect to ad-hoc peers":
|
asyncTest "Admin API: connect to ad-hoc peers":
|
||||||
# Create a couple of nodes
|
# Create a couple of nodes
|
||||||
@ -417,7 +347,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8551)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -458,10 +388,10 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60220))
|
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60220))
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60222))
|
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60222))
|
||||||
peerInfo2 = node2.switch.peerInfo
|
peerInfo2 = node2.peerInfo
|
||||||
nodeKey3 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey3 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node3 = WakuNode.new(nodeKey3, ValidIpAddress.init("0.0.0.0"), Port(60224))
|
node3 = WakuNode.new(nodeKey3, ValidIpAddress.init("0.0.0.0"), Port(60224))
|
||||||
peerInfo3 = node3.switch.peerInfo
|
peerInfo3 = node3.peerInfo
|
||||||
|
|
||||||
await allFutures([node1.start(), node2.start(), node3.start()])
|
await allFutures([node1.start(), node2.start(), node3.start()])
|
||||||
|
|
||||||
@ -475,7 +405,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8552)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -510,7 +440,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# RPC server setup
|
# RPC server setup
|
||||||
let
|
let
|
||||||
rpcPort = Port(8545)
|
rpcPort = Port(8553)
|
||||||
ta = initTAddress(bindIp, rpcPort)
|
ta = initTAddress(bindIp, rpcPort)
|
||||||
server = newRpcHttpServer([ta])
|
server = newRpcHttpServer([ta])
|
||||||
|
|
||||||
@ -521,6 +451,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
await client.connect("127.0.0.1", rpcPort, false)
|
await client.connect("127.0.0.1", rpcPort, false)
|
||||||
|
|
||||||
await node.mountFilter()
|
await node.mountFilter()
|
||||||
|
await node.mountFilterClient()
|
||||||
await node.mountSwap()
|
await node.mountSwap()
|
||||||
let store = StoreQueueRef.new()
|
let store = StoreQueueRef.new()
|
||||||
await node.mountStore(store=store)
|
await node.mountStore(store=store)
|
||||||
@ -539,13 +470,13 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
storeKey = crypto.PrivateKey.random(ECDSA, rng[]).get()
|
storeKey = crypto.PrivateKey.random(ECDSA, rng[]).get()
|
||||||
storePeer = PeerInfo.new(storeKey, @[locationAddr])
|
storePeer = PeerInfo.new(storeKey, @[locationAddr])
|
||||||
|
|
||||||
node.wakuFilter.setPeer(filterPeer.toRemotePeerInfo())
|
|
||||||
node.wakuSwap.setPeer(swapPeer.toRemotePeerInfo())
|
node.wakuSwap.setPeer(swapPeer.toRemotePeerInfo())
|
||||||
|
|
||||||
node.setStorePeer(storePeer.toRemotePeerInfo())
|
node.setStorePeer(storePeer.toRemotePeerInfo())
|
||||||
|
node.setFilterPeer(filterPeer.toRemotePeerInfo())
|
||||||
|
|
||||||
let response = await client.get_waku_v2_admin_v1_peers()
|
let response = await client.get_waku_v2_admin_v1_peers()
|
||||||
|
|
||||||
|
## Then
|
||||||
check:
|
check:
|
||||||
response.len == 3
|
response.len == 3
|
||||||
# Check filter peer
|
# Check filter peer
|
||||||
@ -555,6 +486,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
# Check store peer
|
# Check store peer
|
||||||
(response.filterIt(it.protocol == WakuStoreCodec)[0]).multiaddr == constructMultiaddrStr(storePeer)
|
(response.filterIt(it.protocol == WakuStoreCodec)[0]).multiaddr == constructMultiaddrStr(storePeer)
|
||||||
|
|
||||||
|
## Cleanup
|
||||||
await server.stop()
|
await server.stop()
|
||||||
await server.closeWait()
|
await server.closeWait()
|
||||||
|
|
||||||
@ -588,10 +520,10 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# Setup two servers so we can see both sides of encrypted communication
|
# Setup two servers so we can see both sides of encrypted communication
|
||||||
let
|
let
|
||||||
rpcPort1 = Port(8545)
|
rpcPort1 = Port(8554)
|
||||||
ta1 = initTAddress(bindIp, rpcPort1)
|
ta1 = initTAddress(bindIp, rpcPort1)
|
||||||
server1 = newRpcHttpServer([ta1])
|
server1 = newRpcHttpServer([ta1])
|
||||||
rpcPort3 = Port(8546)
|
rpcPort3 = Port(8555)
|
||||||
ta3 = initTAddress(bindIp, rpcPort3)
|
ta3 = initTAddress(bindIp, rpcPort3)
|
||||||
server3 = newRpcHttpServer([ta3])
|
server3 = newRpcHttpServer([ta3])
|
||||||
|
|
||||||
@ -616,7 +548,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
let sub = await client3.post_waku_v2_relay_v1_subscriptions(@[pubSubTopic])
|
let sub = await client3.post_waku_v2_relay_v1_subscriptions(@[pubSubTopic])
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
check:
|
check:
|
||||||
# node3 is now subscribed to pubSubTopic
|
# node3 is now subscribed to pubSubTopic
|
||||||
@ -627,7 +559,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
check:
|
check:
|
||||||
posted
|
posted
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
# Let's see if we can receive, and decrypt, this message on node3
|
# Let's see if we can receive, and decrypt, this message on node3
|
||||||
var messages = await client3.get_waku_v2_private_v1_asymmetric_messages(pubSubTopic, privateKey = (%keypair.seckey).getStr())
|
var messages = await client3.get_waku_v2_private_v1_asymmetric_messages(pubSubTopic, privateKey = (%keypair.seckey).getStr())
|
||||||
@ -679,10 +611,10 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
# Setup two servers so we can see both sides of encrypted communication
|
# Setup two servers so we can see both sides of encrypted communication
|
||||||
let
|
let
|
||||||
rpcPort1 = Port(8545)
|
rpcPort1 = Port(8556)
|
||||||
ta1 = initTAddress(bindIp, rpcPort1)
|
ta1 = initTAddress(bindIp, rpcPort1)
|
||||||
server1 = newRpcHttpServer([ta1])
|
server1 = newRpcHttpServer([ta1])
|
||||||
rpcPort3 = Port(8546)
|
rpcPort3 = Port(8557)
|
||||||
ta3 = initTAddress(bindIp, rpcPort3)
|
ta3 = initTAddress(bindIp, rpcPort3)
|
||||||
server3 = newRpcHttpServer([ta3])
|
server3 = newRpcHttpServer([ta3])
|
||||||
|
|
||||||
@ -707,7 +639,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
|
|
||||||
let sub = await client3.post_waku_v2_relay_v1_subscriptions(@[pubSubTopic])
|
let sub = await client3.post_waku_v2_relay_v1_subscriptions(@[pubSubTopic])
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
check:
|
check:
|
||||||
# node3 is now subscribed to pubSubTopic
|
# node3 is now subscribed to pubSubTopic
|
||||||
@ -718,7 +650,7 @@ procSuite "Waku v2 JSON-RPC API":
|
|||||||
check:
|
check:
|
||||||
posted
|
posted
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
# Let's see if we can receive, and decrypt, this message on node3
|
# Let's see if we can receive, and decrypt, this message on node3
|
||||||
var messages = await client3.get_waku_v2_private_v1_symmetric_messages(pubSubTopic, symkey = (%symkey).getStr())
|
var messages = await client3.get_waku_v2_private_v1_symmetric_messages(pubSubTopic, symkey = (%symkey).getStr())
|
||||||
|
|||||||
@ -19,7 +19,6 @@ import
|
|||||||
../../waku/v2/node/peer_manager/peer_manager,
|
../../waku/v2/node/peer_manager/peer_manager,
|
||||||
../../waku/v2/node/storage/peer/waku_peer_storage,
|
../../waku/v2/node/storage/peer/waku_peer_storage,
|
||||||
../../waku/v2/node/waku_node,
|
../../waku/v2/node/waku_node,
|
||||||
../../waku/v2/protocol/waku_message,
|
|
||||||
../../waku/v2/protocol/waku_relay,
|
../../waku/v2/protocol/waku_relay,
|
||||||
../../waku/v2/protocol/waku_store,
|
../../waku/v2/protocol/waku_store,
|
||||||
../../waku/v2/protocol/waku_filter,
|
../../waku/v2/protocol/waku_filter,
|
||||||
@ -30,11 +29,9 @@ procSuite "Peer Manager":
|
|||||||
asyncTest "Peer dialing works":
|
asyncTest "Peer dialing works":
|
||||||
let
|
let
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"),
|
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60800))
|
||||||
Port(60000))
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"),
|
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60802))
|
||||||
Port(60002))
|
|
||||||
peerInfo2 = node2.switch.peerInfo
|
peerInfo2 = node2.switch.peerInfo
|
||||||
|
|
||||||
await allFutures([node1.start(), node2.start()])
|
await allFutures([node1.start(), node2.start()])
|
||||||
@ -63,11 +60,9 @@ procSuite "Peer Manager":
|
|||||||
asyncTest "Dialing fails gracefully":
|
asyncTest "Dialing fails gracefully":
|
||||||
let
|
let
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"),
|
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60810))
|
||||||
Port(60000))
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"),
|
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60812))
|
||||||
Port(60002))
|
|
||||||
peerInfo2 = node2.switch.peerInfo
|
peerInfo2 = node2.switch.peerInfo
|
||||||
|
|
||||||
await node1.start()
|
await node1.start()
|
||||||
@ -88,8 +83,7 @@ procSuite "Peer Manager":
|
|||||||
asyncTest "Adding, selecting and filtering peers work":
|
asyncTest "Adding, selecting and filtering peers work":
|
||||||
let
|
let
|
||||||
nodeKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node = WakuNode.new(nodeKey, ValidIpAddress.init("0.0.0.0"),
|
node = WakuNode.new(nodeKey, ValidIpAddress.init("0.0.0.0"), Port(60820))
|
||||||
Port(60000))
|
|
||||||
# Create filter peer
|
# Create filter peer
|
||||||
filterLoc = MultiAddress.init("/ip4/127.0.0.1/tcp/0").tryGet()
|
filterLoc = MultiAddress.init("/ip4/127.0.0.1/tcp/0").tryGet()
|
||||||
filterKey = crypto.PrivateKey.random(ECDSA, rng[]).get()
|
filterKey = crypto.PrivateKey.random(ECDSA, rng[]).get()
|
||||||
@ -105,14 +99,14 @@ procSuite "Peer Manager":
|
|||||||
|
|
||||||
await node.start()
|
await node.start()
|
||||||
|
|
||||||
await node.mountFilter()
|
await node.mountFilterClient()
|
||||||
await node.mountSwap()
|
await node.mountSwap()
|
||||||
node.mountStoreClient()
|
node.mountStoreClient()
|
||||||
|
|
||||||
node.wakuFilter.setPeer(filterPeer.toRemotePeerInfo())
|
|
||||||
node.wakuSwap.setPeer(swapPeer.toRemotePeerInfo())
|
node.wakuSwap.setPeer(swapPeer.toRemotePeerInfo())
|
||||||
|
|
||||||
node.setStorePeer(storePeer.toRemotePeerInfo())
|
node.setStorePeer(storePeer.toRemotePeerInfo())
|
||||||
|
node.setFilterPeer(filterPeer.toRemotePeerInfo())
|
||||||
|
|
||||||
# Check peers were successfully added to peer manager
|
# Check peers were successfully added to peer manager
|
||||||
check:
|
check:
|
||||||
@ -133,11 +127,9 @@ procSuite "Peer Manager":
|
|||||||
asyncTest "Peer manager keeps track of connections":
|
asyncTest "Peer manager keeps track of connections":
|
||||||
let
|
let
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"),
|
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60830))
|
||||||
Port(60000))
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"),
|
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60832))
|
||||||
Port(60002))
|
|
||||||
peerInfo2 = node2.switch.peerInfo
|
peerInfo2 = node2.switch.peerInfo
|
||||||
|
|
||||||
await node1.start()
|
await node1.start()
|
||||||
@ -178,11 +170,9 @@ procSuite "Peer Manager":
|
|||||||
database = SqliteDatabase.new(":memory:")[]
|
database = SqliteDatabase.new(":memory:")[]
|
||||||
storage = WakuPeerStorage.new(database)[]
|
storage = WakuPeerStorage.new(database)[]
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"),
|
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60840), peerStorage = storage)
|
||||||
Port(60000), peerStorage = storage)
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"),
|
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60842))
|
||||||
Port(60002))
|
|
||||||
peerInfo2 = node2.switch.peerInfo
|
peerInfo2 = node2.switch.peerInfo
|
||||||
|
|
||||||
await node1.start()
|
await node1.start()
|
||||||
@ -201,8 +191,7 @@ procSuite "Peer Manager":
|
|||||||
# Simulate restart by initialising a new node using the same storage
|
# Simulate restart by initialising a new node using the same storage
|
||||||
let
|
let
|
||||||
nodeKey3 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey3 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node3 = WakuNode.new(nodeKey3, ValidIpAddress.init("0.0.0.0"),
|
node3 = WakuNode.new(nodeKey3, ValidIpAddress.init("0.0.0.0"), Port(60844), peerStorage = storage)
|
||||||
Port(60004), peerStorage = storage)
|
|
||||||
|
|
||||||
await node3.start()
|
await node3.start()
|
||||||
check:
|
check:
|
||||||
@ -226,11 +215,9 @@ procSuite "Peer Manager":
|
|||||||
database = SqliteDatabase.new(":memory:")[]
|
database = SqliteDatabase.new(":memory:")[]
|
||||||
storage = WakuPeerStorage.new(database)[]
|
storage = WakuPeerStorage.new(database)[]
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"),
|
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60850), peerStorage = storage)
|
||||||
Port(60000), peerStorage = storage)
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"),
|
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60852))
|
||||||
Port(60002))
|
|
||||||
peerInfo2 = node2.switch.peerInfo
|
peerInfo2 = node2.switch.peerInfo
|
||||||
betaCodec = "/vac/waku/relay/2.0.0-beta2"
|
betaCodec = "/vac/waku/relay/2.0.0-beta2"
|
||||||
stableCodec = "/vac/waku/relay/2.0.0"
|
stableCodec = "/vac/waku/relay/2.0.0"
|
||||||
@ -254,8 +241,7 @@ procSuite "Peer Manager":
|
|||||||
# Simulate restart by initialising a new node using the same storage
|
# Simulate restart by initialising a new node using the same storage
|
||||||
let
|
let
|
||||||
nodeKey3 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
nodeKey3 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node3 = WakuNode.new(nodeKey3, ValidIpAddress.init("0.0.0.0"),
|
node3 = WakuNode.new(nodeKey3, ValidIpAddress.init("0.0.0.0"), Port(60854), peerStorage = storage)
|
||||||
Port(60004), peerStorage = storage)
|
|
||||||
|
|
||||||
await node3.mountRelay()
|
await node3.mountRelay()
|
||||||
node3.wakuRelay.codec = stableCodec
|
node3.wakuRelay.codec = stableCodec
|
||||||
|
|||||||
@ -5,193 +5,171 @@ import
|
|||||||
testutils/unittests,
|
testutils/unittests,
|
||||||
chronos,
|
chronos,
|
||||||
chronicles,
|
chronicles,
|
||||||
libp2p/crypto/crypto,
|
libp2p/crypto/crypto
|
||||||
libp2p/multistream
|
|
||||||
import
|
import
|
||||||
../../waku/v2/node/peer_manager/peer_manager,
|
../../waku/v2/node/peer_manager/peer_manager,
|
||||||
../../waku/v2/protocol/waku_message,
|
../../waku/v2/protocol/waku_message,
|
||||||
../../waku/v2/protocol/waku_filter,
|
../../waku/v2/protocol/waku_filter,
|
||||||
../test_helpers,
|
../../waku/v2/protocol/waku_filter/client,
|
||||||
./utils,
|
./utils,
|
||||||
./testlib/common,
|
./testlib/common,
|
||||||
./testlib/switch
|
./testlib/switch
|
||||||
|
|
||||||
|
|
||||||
const dummyHandler = proc(requestId: string, msg: MessagePush) {.async, gcsafe, closure.} = discard
|
proc newTestWakuFilterNode(switch: Switch, timeout: Duration = 2.hours): Future[WakuFilter] {.async.} =
|
||||||
|
let
|
||||||
|
peerManager = PeerManager.new(switch)
|
||||||
|
rng = crypto.newRng()
|
||||||
|
proto = WakuFilter.new(peerManager, rng, timeout)
|
||||||
|
|
||||||
|
await proto.start()
|
||||||
|
switch.mount(proto)
|
||||||
|
|
||||||
|
return proto
|
||||||
|
|
||||||
|
proc newTestWakuFilterClient(switch: Switch): Future[WakuFilterClient] {.async.} =
|
||||||
|
let
|
||||||
|
peerManager = PeerManager.new(switch)
|
||||||
|
rng = crypto.newRng()
|
||||||
|
proto = WakuFilterClient.new(peerManager, rng)
|
||||||
|
|
||||||
|
await proto.start()
|
||||||
|
switch.mount(proto)
|
||||||
|
|
||||||
|
return proto
|
||||||
|
|
||||||
|
|
||||||
# TODO: Extend test coverage
|
# TODO: Extend test coverage
|
||||||
procSuite "Waku Filter":
|
suite "Waku Filter":
|
||||||
|
|
||||||
asyncTest "should forward messages to client after subscribed":
|
asyncTest "should forward messages to client after subscribed":
|
||||||
## Setup
|
## Setup
|
||||||
let rng = crypto.newRng()
|
|
||||||
let
|
let
|
||||||
clientSwitch = newTestSwitch()
|
|
||||||
serverSwitch = newTestSwitch()
|
serverSwitch = newTestSwitch()
|
||||||
|
clientSwitch = newTestSwitch()
|
||||||
|
|
||||||
await allFutures(serverSwitch.start(), clientSwitch.start())
|
await allFutures(serverSwitch.start(), clientSwitch.start())
|
||||||
|
|
||||||
|
let
|
||||||
|
server = await newTestWakuFilterNode(serverSwitch)
|
||||||
|
client = await newTestWakuFilterClient(clientSwitch)
|
||||||
|
|
||||||
## Given
|
## Given
|
||||||
# Server
|
let serverAddr = serverSwitch.peerInfo.toRemotePeerInfo()
|
||||||
let
|
|
||||||
serverPeerManager = PeerManager.new(serverSwitch)
|
let pushHandlerFuture = newFuture[(string, WakuMessage)]()
|
||||||
serverProto = WakuFilter.init(serverPeerManager, rng, dummyHandler)
|
proc pushHandler(pubsubTopic: string, message: WakuMessage) {.gcsafe, closure.} =
|
||||||
await serverProto.start()
|
pushHandlerFuture.complete((pubsubTopic, message))
|
||||||
serverSwitch.mount(serverProto)
|
|
||||||
|
|
||||||
# Client
|
let
|
||||||
let handlerFuture = newFuture[(string, MessagePush)]()
|
pubsubTopic = DefaultPubsubTopic
|
||||||
proc handler(requestId: string, push: MessagePush) {.async, gcsafe, closure.} =
|
contentTopic = "test-content-topic"
|
||||||
handlerFuture.complete((requestId, push))
|
msg = fakeWakuMessage(contentTopic=contentTopic)
|
||||||
|
|
||||||
let
|
|
||||||
clientPeerManager = PeerManager.new(clientSwitch)
|
|
||||||
clientProto = WakuFilter.init(clientPeerManager, rng, handler)
|
|
||||||
await clientProto.start()
|
|
||||||
clientSwitch.mount(clientProto)
|
|
||||||
|
|
||||||
clientProto.setPeer(serverSwitch.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
## When
|
## When
|
||||||
let resSubscription = await clientProto.subscribe(DefaultPubsubTopic, @[DefaultContentTopic])
|
require (await client.subscribe(pubsubTopic, contentTopic, pushHandler, peer=serverAddr)).isOk()
|
||||||
require resSubscription.isOk()
|
|
||||||
|
|
||||||
|
# WARN: Sleep necessary to avoid a race condition between the subscription and the handle message proc
|
||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
let message = WakuMessage(payload: @[byte 1, 2, 3], contentTopic: DefaultContentTopic)
|
await server.handleMessage(pubsubTopic, msg)
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
|
||||||
|
require await pushHandlerFuture.withTimeout(5.seconds)
|
||||||
|
|
||||||
## Then
|
## Then
|
||||||
let subscriptionRequestId = resSubscription.get()
|
let (pushedMsgPubsubTopic, pushedMsg) = pushHandlerFuture.read()
|
||||||
let (requestId, push) = await handlerFuture
|
|
||||||
|
|
||||||
check:
|
check:
|
||||||
requestId == subscriptionRequestId
|
pushedMsgPubsubTopic == pubsubTopic
|
||||||
push.messages == @[message]
|
pushedMsg == msg
|
||||||
|
|
||||||
## Cleanup
|
## Cleanup
|
||||||
await allFutures(clientSwitch.stop(), serverSwitch.stop())
|
await allFutures(clientSwitch.stop(), serverSwitch.stop())
|
||||||
|
|
||||||
asyncTest "should not forward messages to client after unsuscribed":
|
asyncTest "should not forward messages to client after unsuscribed":
|
||||||
## Setup
|
## Setup
|
||||||
let rng = crypto.newRng()
|
|
||||||
let
|
let
|
||||||
clientSwitch = newTestSwitch()
|
|
||||||
serverSwitch = newTestSwitch()
|
serverSwitch = newTestSwitch()
|
||||||
|
clientSwitch = newTestSwitch()
|
||||||
|
|
||||||
await allFutures(serverSwitch.start(), clientSwitch.start())
|
await allFutures(serverSwitch.start(), clientSwitch.start())
|
||||||
|
|
||||||
|
let
|
||||||
|
server = await newTestWakuFilterNode(serverSwitch)
|
||||||
|
client = await newTestWakuFilterClient(clientSwitch)
|
||||||
|
|
||||||
## Given
|
## Given
|
||||||
# Server
|
let serverAddr = serverSwitch.peerInfo.toRemotePeerInfo()
|
||||||
let
|
|
||||||
serverPeerManager = PeerManager.new(serverSwitch)
|
|
||||||
serverProto = WakuFilter.init(serverPeerManager, rng, dummyHandler)
|
|
||||||
await serverProto.start()
|
|
||||||
serverSwitch.mount(serverProto)
|
|
||||||
|
|
||||||
# Client
|
var pushHandlerFuture = newFuture[void]()
|
||||||
var handlerFuture = newFuture[void]()
|
proc pushHandler(pubsubTopic: string, message: WakuMessage) {.gcsafe, closure.} =
|
||||||
proc handler(requestId: string, push: MessagePush) {.async, gcsafe, closure.} =
|
pushHandlerFuture.complete()
|
||||||
handlerFuture.complete()
|
|
||||||
|
|
||||||
let
|
let
|
||||||
clientPeerManager = PeerManager.new(clientSwitch)
|
pubsubTopic = DefaultPubsubTopic
|
||||||
clientProto = WakuFilter.init(clientPeerManager, rng, handler)
|
contentTopic = "test-content-topic"
|
||||||
await clientProto.start()
|
msg = fakeWakuMessage(contentTopic=contentTopic)
|
||||||
clientSwitch.mount(clientProto)
|
|
||||||
|
|
||||||
clientProto.setPeer(serverSwitch.peerInfo.toRemotePeerInfo())
|
## When
|
||||||
|
require (await client.subscribe(pubsubTopic, contentTopic, pushHandler, peer=serverAddr)).isOk()
|
||||||
## Given
|
|
||||||
let message = WakuMessage(payload: @[byte 1, 2, 3], contentTopic: DefaultContentTopic)
|
|
||||||
|
|
||||||
let resSubscription = await clientProto.subscribe(DefaultPubsubTopic, @[DefaultContentTopic])
|
|
||||||
require resSubscription.isOk()
|
|
||||||
|
|
||||||
|
# WARN: Sleep necessary to avoid a race condition between the subscription and the handle message proc
|
||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(pubsubTopic, msg)
|
||||||
let handlerWasCalledAfterSubscription = await handlerFuture.withTimeout(1.seconds)
|
|
||||||
require handlerWasCalledAfterSubscription
|
require await pushHandlerFuture.withTimeout(1.seconds)
|
||||||
|
|
||||||
# Reset to test unsubscribe
|
# Reset to test unsubscribe
|
||||||
handlerFuture = newFuture[void]()
|
pushHandlerFuture = newFuture[void]()
|
||||||
|
|
||||||
let resUnsubscription = await clientProto.unsubscribe(DefaultPubsubTopic, @[DefaultContentTopic])
|
require (await client.unsubscribe(pubsubTopic, contentTopic, peer=serverAddr)).isOk()
|
||||||
require resUnsubscription.isOk()
|
|
||||||
|
|
||||||
|
# WARN: Sleep necessary to avoid a race condition between the unsubscription and the handle message proc
|
||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(pubsubTopic, msg)
|
||||||
|
|
||||||
## Then
|
## Then
|
||||||
let handlerWasCalledAfterUnsubscription = await handlerFuture.withTimeout(1.seconds)
|
let handlerWasCalledAfterUnsubscription = await pushHandlerFuture.withTimeout(1.seconds)
|
||||||
check:
|
check:
|
||||||
not handlerWasCalledAfterUnsubscription
|
not handlerWasCalledAfterUnsubscription
|
||||||
|
|
||||||
## Cleanup
|
## Cleanup
|
||||||
await allFutures(clientSwitch.stop(), serverSwitch.stop())
|
await allFutures(clientSwitch.stop(), serverSwitch.stop())
|
||||||
|
|
||||||
asyncTest "subscription should fail if no filter peer is provided":
|
|
||||||
## Setup
|
|
||||||
let clientSwitch = newTestSwitch()
|
|
||||||
await clientSwitch.start()
|
|
||||||
|
|
||||||
## Given
|
|
||||||
let clientProto = WakuFilter.init(PeerManager.new(clientSwitch), crypto.newRng(), dummyHandler)
|
|
||||||
await clientProto.start()
|
|
||||||
clientSwitch.mount(clientProto)
|
|
||||||
|
|
||||||
## When
|
|
||||||
let resSubscription = await clientProto.subscribe(DefaultPubsubTopic, @[DefaultContentTopic])
|
|
||||||
|
|
||||||
## Then
|
|
||||||
check:
|
|
||||||
resSubscription.isErr()
|
|
||||||
resSubscription.error() == "peer_not_found_failure"
|
|
||||||
|
|
||||||
asyncTest "peer subscription should be dropped if connection fails for second time after the timeout has elapsed":
|
asyncTest "peer subscription should be dropped if connection fails for second time after the timeout has elapsed":
|
||||||
## Setup
|
## Setup
|
||||||
let rng = crypto.newRng()
|
|
||||||
let
|
let
|
||||||
clientSwitch = newTestSwitch()
|
|
||||||
serverSwitch = newTestSwitch()
|
serverSwitch = newTestSwitch()
|
||||||
|
clientSwitch = newTestSwitch()
|
||||||
|
|
||||||
await allFutures(serverSwitch.start(), clientSwitch.start())
|
await allFutures(serverSwitch.start(), clientSwitch.start())
|
||||||
|
|
||||||
|
let
|
||||||
|
server = await newTestWakuFilterNode(serverSwitch, timeout=200.milliseconds)
|
||||||
|
client = await newTestWakuFilterClient(clientSwitch)
|
||||||
|
|
||||||
## Given
|
## Given
|
||||||
# Server
|
let serverAddr = serverSwitch.peerInfo.toRemotePeerInfo()
|
||||||
let
|
|
||||||
serverPeerManager = PeerManager.new(serverSwitch)
|
|
||||||
serverProto = WakuFilter.init(serverPeerManager, rng, dummyHandler, timeout=1.seconds)
|
|
||||||
await serverProto.start()
|
|
||||||
serverSwitch.mount(serverProto)
|
|
||||||
|
|
||||||
# Client
|
var pushHandlerFuture = newFuture[void]()
|
||||||
var handlerFuture = newFuture[void]()
|
proc pushHandler(pubsubTopic: string, message: WakuMessage) {.gcsafe, closure.} =
|
||||||
proc handler(requestId: string, push: MessagePush) {.async, gcsafe, closure.} =
|
pushHandlerFuture.complete()
|
||||||
handlerFuture.complete()
|
|
||||||
|
|
||||||
let
|
let
|
||||||
clientPeerManager = PeerManager.new(clientSwitch)
|
pubsubTopic = DefaultPubsubTopic
|
||||||
clientProto = WakuFilter.init(clientPeerManager, rng, handler)
|
contentTopic = "test-content-topic"
|
||||||
await clientProto.start()
|
msg = fakeWakuMessage(contentTopic=contentTopic)
|
||||||
clientSwitch.mount(clientProto)
|
|
||||||
|
|
||||||
clientProto.setPeer(serverSwitch.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
## When
|
## When
|
||||||
let resSubscription = await clientProto.subscribe(DefaultPubsubTopic, @[DefaultContentTopic])
|
require (await client.subscribe(pubsubTopic, contentTopic, pushHandler, peer=serverAddr)).isOk()
|
||||||
check resSubscription.isOk()
|
|
||||||
|
|
||||||
|
# WARN: Sleep necessary to avoid a race condition between the unsubscription and the handle message proc
|
||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
let message = WakuMessage(payload: @[byte 1, 2, 3], contentTopic: DefaultContentTopic)
|
await server.handleMessage(DefaultPubsubTopic, msg)
|
||||||
|
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
# Push handler should be called
|
||||||
let handlerShouldHaveBeenCalled = await handlerFuture.withTimeout(1.seconds)
|
require await pushHandlerFuture.withTimeout(1.seconds)
|
||||||
require handlerShouldHaveBeenCalled
|
|
||||||
|
|
||||||
# Stop client node to test timeout unsubscription
|
# Stop client node to test timeout unsubscription
|
||||||
await clientSwitch.stop()
|
await clientSwitch.stop()
|
||||||
@ -199,19 +177,19 @@ procSuite "Waku Filter":
|
|||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
# First failure should not remove the subscription
|
# First failure should not remove the subscription
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(DefaultPubsubTopic, msg)
|
||||||
let
|
let
|
||||||
subscriptionsBeforeTimeout = serverProto.subscriptions.len()
|
subscriptionsBeforeTimeout = server.subscriptions.len()
|
||||||
failedPeersBeforeTimeout = serverProto.failedPeers.len()
|
failedPeersBeforeTimeout = server.failedPeers.len()
|
||||||
|
|
||||||
# Wait for peer connection failure timeout to elapse
|
# Wait for the configured peer connection timeout to elapse (200ms)
|
||||||
await sleepAsync(1.seconds)
|
await sleepAsync(200.milliseconds)
|
||||||
|
|
||||||
#Second failure should remove the subscription
|
# Second failure should remove the subscription
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(DefaultPubsubTopic, msg)
|
||||||
let
|
let
|
||||||
subscriptionsAfterTimeout = serverProto.subscriptions.len()
|
subscriptionsAfterTimeout = server.subscriptions.len()
|
||||||
failedPeersAfterTimeout = serverProto.failedPeers.len()
|
failedPeersAfterTimeout = server.failedPeers.len()
|
||||||
|
|
||||||
## Then
|
## Then
|
||||||
check:
|
check:
|
||||||
@ -222,55 +200,45 @@ procSuite "Waku Filter":
|
|||||||
|
|
||||||
## Cleanup
|
## Cleanup
|
||||||
await serverSwitch.stop()
|
await serverSwitch.stop()
|
||||||
|
|
||||||
asyncTest "peer subscription should not be dropped if connection recovers before timeout elapses":
|
asyncTest "peer subscription should not be dropped if connection recovers before timeout elapses":
|
||||||
## Setup
|
## Setup
|
||||||
let
|
let
|
||||||
clientKey = PrivateKey.random(ECDSA, rng[]).get()
|
|
||||||
clientAddress = MultiAddress.init("/ip4/127.0.0.1/tcp/65000").get()
|
|
||||||
|
|
||||||
let rng = crypto.newRng()
|
|
||||||
let
|
|
||||||
clientSwitch = newTestSwitch(some(clientKey), some(clientAddress))
|
|
||||||
serverSwitch = newTestSwitch()
|
serverSwitch = newTestSwitch()
|
||||||
|
clientSwitch = newTestSwitch()
|
||||||
|
|
||||||
await allFutures(serverSwitch.start(), clientSwitch.start())
|
await allFutures(serverSwitch.start(), clientSwitch.start())
|
||||||
|
|
||||||
|
let
|
||||||
|
server = await newTestWakuFilterNode(serverSwitch, timeout=200.milliseconds)
|
||||||
|
client = await newTestWakuFilterClient(clientSwitch)
|
||||||
|
|
||||||
## Given
|
## Given
|
||||||
# Server
|
let serverAddr = serverSwitch.peerInfo.toRemotePeerInfo()
|
||||||
let
|
|
||||||
serverPeerManager = PeerManager.new(serverSwitch)
|
|
||||||
serverProto = WakuFilter.init(serverPeerManager, rng, dummyHandler, timeout=2.seconds)
|
|
||||||
await serverProto.start()
|
|
||||||
serverSwitch.mount(serverProto)
|
|
||||||
|
|
||||||
# Client
|
var pushHandlerFuture = newFuture[void]()
|
||||||
var handlerFuture = newFuture[void]()
|
proc pushHandler(pubsubTopic: string, message: WakuMessage) {.gcsafe, closure.} =
|
||||||
proc handler(requestId: string, push: MessagePush) {.async, gcsafe, closure.} =
|
pushHandlerFuture.complete()
|
||||||
handlerFuture.complete()
|
|
||||||
|
|
||||||
let
|
let
|
||||||
clientPeerManager = PeerManager.new(clientSwitch)
|
pubsubTopic = DefaultPubsubTopic
|
||||||
clientProto = WakuFilter.init(clientPeerManager, rng, handler)
|
contentTopic = "test-content-topic"
|
||||||
await clientProto.start()
|
msg = fakeWakuMessage(contentTopic=contentTopic)
|
||||||
clientSwitch.mount(clientProto)
|
|
||||||
|
|
||||||
clientProto.setPeer(serverSwitch.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
## When
|
## When
|
||||||
let message = WakuMessage(payload: @[byte 1, 2, 3], contentTopic: DefaultContentTopic)
|
require (await client.subscribe(pubsubTopic, contentTopic, pushHandler, peer=serverAddr)).isOk()
|
||||||
|
|
||||||
let resSubscription = await clientProto.subscribe(DefaultPubsubTopic, @[DefaultContentTopic])
|
|
||||||
check resSubscription.isOk()
|
|
||||||
|
|
||||||
|
# WARN: Sleep necessary to avoid a race condition between the unsubscription and the handle message proc
|
||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(DefaultPubsubTopic, msg)
|
||||||
handlerFuture = newFuture[void]()
|
|
||||||
|
# Push handler should be called
|
||||||
|
require await pushHandlerFuture.withTimeout(1.seconds)
|
||||||
|
|
||||||
let
|
let
|
||||||
subscriptionsBeforeFailure = serverProto.subscriptions.len()
|
subscriptionsBeforeFailure = server.subscriptions.len()
|
||||||
failedPeersBeforeFailure = serverProto.failedPeers.len()
|
failedPeersBeforeFailure = server.failedPeers.len()
|
||||||
|
|
||||||
# Stop switch to test unsubscribe
|
# Stop switch to test unsubscribe
|
||||||
await clientSwitch.stop()
|
await clientSwitch.stop()
|
||||||
@ -278,29 +246,33 @@ procSuite "Waku Filter":
|
|||||||
await sleepAsync(5.milliseconds)
|
await sleepAsync(5.milliseconds)
|
||||||
|
|
||||||
# First failure should add to failure list
|
# First failure should add to failure list
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(DefaultPubsubTopic, msg)
|
||||||
handlerFuture = newFuture[void]()
|
|
||||||
|
pushHandlerFuture = newFuture[void]()
|
||||||
|
|
||||||
let
|
let
|
||||||
subscriptionsAfterFailure = serverProto.subscriptions.len()
|
subscriptionsAfterFailure = server.subscriptions.len()
|
||||||
failedPeersAfterFailure = serverProto.failedPeers.len()
|
failedPeersAfterFailure = server.failedPeers.len()
|
||||||
|
|
||||||
|
await sleepAsync(100.milliseconds)
|
||||||
|
|
||||||
await sleepAsync(250.milliseconds)
|
|
||||||
|
|
||||||
# Start switch with same key as before
|
# Start switch with same key as before
|
||||||
var clientSwitch2 = newTestSwitch(some(clientKey), some(clientAddress))
|
let clientSwitch2 = newTestSwitch(
|
||||||
|
some(clientSwitch.peerInfo.privateKey),
|
||||||
|
some(clientSwitch.peerInfo.addrs[0])
|
||||||
|
)
|
||||||
await clientSwitch2.start()
|
await clientSwitch2.start()
|
||||||
await clientProto.start()
|
await client.start()
|
||||||
clientSwitch2.mount(clientProto)
|
clientSwitch2.mount(client)
|
||||||
|
|
||||||
# If push succeeds after failure, the peer should removed from failed peers list
|
# If push succeeds after failure, the peer should removed from failed peers list
|
||||||
await serverProto.handleMessage(DefaultPubsubTopic, message)
|
await server.handleMessage(DefaultPubsubTopic, msg)
|
||||||
let handlerShouldHaveBeenCalled = await handlerFuture.withTimeout(1.seconds)
|
let handlerShouldHaveBeenCalled = await pushHandlerFuture.withTimeout(1.seconds)
|
||||||
|
|
||||||
let
|
|
||||||
subscriptionsAfterSuccessfulConnection = serverProto.subscriptions.len()
|
|
||||||
failedPeersAfterSuccessfulConnection = serverProto.failedPeers.len()
|
|
||||||
|
|
||||||
|
let
|
||||||
|
subscriptionsAfterSuccessfulConnection = server.subscriptions.len()
|
||||||
|
failedPeersAfterSuccessfulConnection = server.failedPeers.len()
|
||||||
|
|
||||||
## Then
|
## Then
|
||||||
check:
|
check:
|
||||||
handlerShouldHaveBeenCalled
|
handlerShouldHaveBeenCalled
|
||||||
@ -314,6 +286,6 @@ procSuite "Waku Filter":
|
|||||||
failedPeersBeforeFailure == 0
|
failedPeersBeforeFailure == 0
|
||||||
failedPeersAfterFailure == 1
|
failedPeersAfterFailure == 1
|
||||||
failedPeersAfterSuccessfulConnection == 0
|
failedPeersAfterSuccessfulConnection == 0
|
||||||
|
|
||||||
## Cleanup
|
## Cleanup
|
||||||
await allFutures(clientSwitch2.stop(), serverSwitch.stop())
|
await allFutures(clientSwitch2.stop(), serverSwitch.stop())
|
||||||
|
|||||||
@ -6,292 +6,60 @@ import
|
|||||||
testutils/unittests,
|
testutils/unittests,
|
||||||
chronicles,
|
chronicles,
|
||||||
chronos,
|
chronos,
|
||||||
libp2p/crypto/crypto,
|
libp2p/crypto/crypto
|
||||||
libp2p/peerid,
|
|
||||||
libp2p/multiaddress,
|
|
||||||
libp2p/switch,
|
|
||||||
libp2p/protocols/pubsub/rpc/messages,
|
|
||||||
libp2p/protocols/pubsub/pubsub,
|
|
||||||
libp2p/protocols/pubsub/gossipsub,
|
|
||||||
eth/keys
|
|
||||||
import
|
import
|
||||||
../../waku/v2/protocol/[waku_relay, waku_message],
|
|
||||||
../../waku/v2/protocol/waku_filter,
|
|
||||||
../../waku/v2/node/peer_manager/peer_manager,
|
../../waku/v2/node/peer_manager/peer_manager,
|
||||||
|
../../waku/v2/node/waku_node,
|
||||||
|
../../waku/v2/protocol/waku_message,
|
||||||
|
../../waku/v2/protocol/waku_filter,
|
||||||
../../waku/v2/utils/peers,
|
../../waku/v2/utils/peers,
|
||||||
../../waku/v2/node/waku_node
|
./testlib/common
|
||||||
|
|
||||||
|
|
||||||
procSuite "WakuNode - Filter":
|
suite "WakuNode - Filter":
|
||||||
let rng = keys.newRng()
|
|
||||||
|
|
||||||
asyncTest "Message published with content filter is retrievable":
|
asyncTest "subscriber should receive the message handled by the publisher":
|
||||||
|
## Setup
|
||||||
|
let rng = crypto.newRng()
|
||||||
let
|
let
|
||||||
nodeKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
node = WakuNode.new(nodeKey, ValidIpAddress.init("0.0.0.0"),
|
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60110))
|
||||||
Port(60000))
|
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
pubSubTopic = "chat"
|
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60111))
|
||||||
contentTopic = ContentTopic("/waku/2/default-content/proto")
|
|
||||||
filterRequest = FilterRequest(pubSubTopic: pubSubTopic, contentFilters: @[ContentFilter(contentTopic: contentTopic)], subscribe: true)
|
|
||||||
message = WakuMessage(payload: "hello world".toBytes(),
|
|
||||||
contentTopic: contentTopic)
|
|
||||||
|
|
||||||
# This could/should become a more fixed handler (at least default) that
|
await allFutures(server.start(), client.start())
|
||||||
# would be enforced on WakuNode level.
|
|
||||||
proc relayHandler(topic: string, data: seq[byte]) {.async, gcsafe.} =
|
|
||||||
let msg = WakuMessage.init(data)
|
|
||||||
if msg.isOk():
|
|
||||||
check:
|
|
||||||
topic == "chat"
|
|
||||||
node.filters.notify(msg.value(), topic)
|
|
||||||
|
|
||||||
var completionFut = newFuture[bool]()
|
await server.mountFilter()
|
||||||
|
await client.mountFilterClient()
|
||||||
|
|
||||||
# This would be the actual application handler
|
## Given
|
||||||
proc contentHandler(msg: WakuMessage) {.gcsafe, closure.} =
|
let serverPeerInfo = server.peerInfo.toRemotePeerInfo()
|
||||||
let message = string.fromBytes(msg.payload)
|
|
||||||
check:
|
|
||||||
message == "hello world"
|
|
||||||
completionFut.complete(true)
|
|
||||||
|
|
||||||
await node.start()
|
|
||||||
await node.mountRelay()
|
|
||||||
|
|
||||||
# Subscribe our node to the pubSubTopic where all chat data go onto.
|
|
||||||
node.subscribe(pubSubTopic, relayHandler)
|
|
||||||
|
|
||||||
# Subscribe a contentFilter to trigger a specific application handler when
|
|
||||||
# WakuMessages with that content are received
|
|
||||||
await node.subscribe(filterRequest, contentHandler)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
await node.publish(pubSubTopic, message)
|
|
||||||
|
|
||||||
check:
|
|
||||||
(await completionFut.withTimeout(5.seconds)) == true
|
|
||||||
|
|
||||||
await node.stop()
|
|
||||||
|
|
||||||
asyncTest "Content filtered publishing over network":
|
|
||||||
let
|
let
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
pubSubTopic = DefaultPubsubTopic
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
contentTopic = DefaultContentTopic
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
message = fakeWakuMessage(contentTopic=contentTopic)
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
|
||||||
let
|
|
||||||
pubSubTopic = "chat"
|
|
||||||
contentTopic = ContentTopic("/waku/2/default-content/proto")
|
|
||||||
filterRequest = FilterRequest(pubSubTopic: pubSubTopic, contentFilters: @[ContentFilter(contentTopic: contentTopic)], subscribe: true)
|
|
||||||
message = WakuMessage(payload: "hello world".toBytes(),
|
|
||||||
contentTopic: contentTopic)
|
|
||||||
|
|
||||||
var completionFut = newFuture[bool]()
|
var filterPushHandlerFut = newFuture[(PubsubTopic, WakuMessage)]()
|
||||||
|
proc filterPushHandler(pubsubTopic: PubsubTopic, msg: WakuMessage) {.gcsafe, closure.} =
|
||||||
|
filterPushHandlerFut.complete((pubsubTopic, msg))
|
||||||
|
|
||||||
# This could/should become a more fixed handler (at least default) that
|
## When
|
||||||
# would be enforced on WakuNode level.
|
await client.filterSubscribe(pubsubTopic, contentTopic, filterPushHandler, peer=serverPeerInfo)
|
||||||
proc relayHandler(topic: string, data: seq[byte]) {.async, gcsafe.} =
|
|
||||||
let msg = WakuMessage.init(data)
|
|
||||||
if msg.isOk():
|
|
||||||
check:
|
|
||||||
topic == "chat"
|
|
||||||
node1.filters.notify(msg.value(), topic)
|
|
||||||
|
|
||||||
# This would be the actual application handler
|
# Wait for subscription to take effect
|
||||||
proc contentHandler(msg: WakuMessage) {.gcsafe, closure.} =
|
await sleepAsync(100.millis)
|
||||||
let message = string.fromBytes(msg.payload)
|
|
||||||
check:
|
|
||||||
message == "hello world"
|
|
||||||
completionFut.complete(true)
|
|
||||||
|
|
||||||
await allFutures([node1.start(), node2.start()])
|
await server.filterHandleMessage(pubSubTopic, message)
|
||||||
|
|
||||||
await node1.mountRelay()
|
require await filterPushHandlerFut.withTimeout(5.seconds)
|
||||||
await node2.mountRelay()
|
|
||||||
|
|
||||||
await node1.mountFilter()
|
|
||||||
await node2.mountFilter()
|
|
||||||
|
|
||||||
# Subscribe our node to the pubSubTopic where all chat data go onto.
|
|
||||||
node1.subscribe(pubSubTopic, relayHandler)
|
|
||||||
# Subscribe a contentFilter to trigger a specific application handler when
|
|
||||||
# WakuMessages with that content are received
|
|
||||||
node1.wakuFilter.setPeer(node2.switch.peerInfo.toRemotePeerInfo())
|
|
||||||
await node1.subscribe(filterRequest, contentHandler)
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
# Connect peers by dialing from node2 to node1
|
|
||||||
discard await node2.switch.dial(node1.switch.peerInfo.peerId, node1.switch.peerInfo.addrs, WakuRelayCodec)
|
|
||||||
|
|
||||||
# We need to sleep to allow the subscription to go through
|
|
||||||
info "Going to sleep to allow subscribe to go through"
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
info "Waking up and publishing"
|
|
||||||
await node2.publish(pubSubTopic, message)
|
|
||||||
|
|
||||||
|
## Then
|
||||||
|
check filterPushHandlerFut.completed()
|
||||||
|
let (filterPubsubTopic, filterMessage) = filterPushHandlerFut.read()
|
||||||
check:
|
check:
|
||||||
(await completionFut.withTimeout(5.seconds)) == true
|
filterPubsubTopic == pubsubTopic
|
||||||
await node1.stop()
|
filterMessage == message
|
||||||
await node2.stop()
|
|
||||||
|
|
||||||
asyncTest "Can receive filtered messages published on both default and other topics":
|
|
||||||
let
|
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
|
||||||
defaultTopic = "/waku/2/default-waku/proto"
|
|
||||||
otherTopic = "/non/waku/formatted"
|
|
||||||
defaultContentTopic = "defaultCT"
|
|
||||||
otherContentTopic = "otherCT"
|
|
||||||
defaultPayload = @[byte 1]
|
|
||||||
otherPayload = @[byte 9]
|
|
||||||
defaultMessage = WakuMessage(payload: defaultPayload, contentTopic: defaultContentTopic)
|
|
||||||
otherMessage = WakuMessage(payload: otherPayload, contentTopic: otherContentTopic)
|
|
||||||
defaultFR = FilterRequest(contentFilters: @[ContentFilter(contentTopic: defaultContentTopic)], subscribe: true)
|
|
||||||
otherFR = FilterRequest(contentFilters: @[ContentFilter(contentTopic: otherContentTopic)], subscribe: true)
|
|
||||||
|
|
||||||
await node1.start()
|
## Cleanup
|
||||||
await node1.mountRelay()
|
await allFutures(client.stop(), server.stop())
|
||||||
await node1.mountFilter()
|
|
||||||
|
|
||||||
await node2.start()
|
|
||||||
await node2.mountRelay()
|
|
||||||
await node2.mountFilter()
|
|
||||||
node2.wakuFilter.setPeer(node1.switch.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
var defaultComplete = newFuture[bool]()
|
|
||||||
var otherComplete = newFuture[bool]()
|
|
||||||
|
|
||||||
# Subscribe nodes 1 and 2 to otherTopic
|
|
||||||
proc emptyHandler(topic: string, data: seq[byte]) {.async, gcsafe.} =
|
|
||||||
# Do not notify filters or subscriptions here. This should be default behaviour for all topics
|
|
||||||
discard
|
|
||||||
|
|
||||||
node1.subscribe(otherTopic, emptyHandler)
|
|
||||||
node2.subscribe(otherTopic, emptyHandler)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
proc defaultHandler(msg: WakuMessage) {.gcsafe, closure.} =
|
|
||||||
check:
|
|
||||||
msg.payload == defaultPayload
|
|
||||||
msg.contentTopic == defaultContentTopic
|
|
||||||
defaultComplete.complete(true)
|
|
||||||
|
|
||||||
proc otherHandler(msg: WakuMessage) {.gcsafe, closure.} =
|
|
||||||
check:
|
|
||||||
msg.payload == otherPayload
|
|
||||||
msg.contentTopic == otherContentTopic
|
|
||||||
otherComplete.complete(true)
|
|
||||||
|
|
||||||
# Subscribe a contentFilter to trigger a specific application handler when
|
|
||||||
# WakuMessages with that content are received
|
|
||||||
await node2.subscribe(defaultFR, defaultHandler)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
# Let's check that content filtering works on the default topic
|
|
||||||
await node1.publish(defaultTopic, defaultMessage)
|
|
||||||
|
|
||||||
check:
|
|
||||||
(await defaultComplete.withTimeout(5.seconds)) == true
|
|
||||||
|
|
||||||
# Now check that content filtering works on other topics
|
|
||||||
await node2.subscribe(otherFR, otherHandler)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
await node1.publish(otherTopic,otherMessage)
|
|
||||||
|
|
||||||
check:
|
|
||||||
(await otherComplete.withTimeout(5.seconds)) == true
|
|
||||||
|
|
||||||
await node1.stop()
|
|
||||||
await node2.stop()
|
|
||||||
|
|
||||||
asyncTest "Filter protocol works on node without relay capability":
|
|
||||||
let
|
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
|
||||||
defaultTopic = "/waku/2/default-waku/proto"
|
|
||||||
contentTopic = "defaultCT"
|
|
||||||
payload = @[byte 1]
|
|
||||||
message = WakuMessage(payload: payload, contentTopic: contentTopic)
|
|
||||||
filterRequest = FilterRequest(contentFilters: @[ContentFilter(contentTopic: contentTopic)], subscribe: true)
|
|
||||||
|
|
||||||
await node1.start()
|
|
||||||
await node1.mountRelay()
|
|
||||||
await node1.mountFilter()
|
|
||||||
|
|
||||||
await node2.start()
|
|
||||||
await node2.mountFilter()
|
|
||||||
node2.wakuFilter.setPeer(node1.switch.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
check:
|
|
||||||
node1.wakuRelay.isNil == false # Node1 is a full node
|
|
||||||
node2.wakuRelay.isNil == true # Node 2 is a light node
|
|
||||||
|
|
||||||
var completeFut = newFuture[bool]()
|
|
||||||
|
|
||||||
proc filterHandler(msg: WakuMessage) {.gcsafe, closure.} =
|
|
||||||
check:
|
|
||||||
msg.payload == payload
|
|
||||||
msg.contentTopic == contentTopic
|
|
||||||
completeFut.complete(true)
|
|
||||||
|
|
||||||
# Subscribe a contentFilter to trigger a specific application handler when
|
|
||||||
# WakuMessages with that content are received
|
|
||||||
await node2.subscribe(filterRequest, filterHandler)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
# Let's check that content filtering works on the default topic
|
|
||||||
await node1.publish(defaultTopic, message)
|
|
||||||
|
|
||||||
check:
|
|
||||||
(await completeFut.withTimeout(5.seconds)) == true
|
|
||||||
|
|
||||||
await node1.stop()
|
|
||||||
await node2.stop()
|
|
||||||
|
|
||||||
asyncTest "Filter protocol returns expected message":
|
|
||||||
let
|
|
||||||
nodeKey1 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
|
||||||
node1 = WakuNode.new(nodeKey1, ValidIpAddress.init("0.0.0.0"),
|
|
||||||
Port(60000))
|
|
||||||
nodeKey2 = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
|
||||||
node2 = WakuNode.new(nodeKey2, ValidIpAddress.init("0.0.0.0"),
|
|
||||||
Port(60002))
|
|
||||||
contentTopic = ContentTopic("/waku/2/default-content/proto")
|
|
||||||
message = WakuMessage(payload: "hello world".toBytes(), contentTopic: contentTopic)
|
|
||||||
|
|
||||||
var completionFut = newFuture[bool]()
|
|
||||||
|
|
||||||
await node1.start()
|
|
||||||
await node1.mountFilter()
|
|
||||||
await node2.start()
|
|
||||||
await node2.mountFilter()
|
|
||||||
|
|
||||||
node1.wakuFilter.setPeer(node2.switch.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
proc handler(msg: WakuMessage) {.gcsafe, closure.} =
|
|
||||||
check:
|
|
||||||
msg == message
|
|
||||||
completionFut.complete(true)
|
|
||||||
|
|
||||||
await node1.subscribe(FilterRequest(pubSubTopic: "/waku/2/default-waku/proto", contentFilters: @[ContentFilter(contentTopic: contentTopic)], subscribe: true), handler)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
await node2.wakuFilter.handleMessage("/waku/2/default-waku/proto", message)
|
|
||||||
|
|
||||||
await sleepAsync(2000.millis)
|
|
||||||
|
|
||||||
check:
|
|
||||||
(await completionFut.withTimeout(5.seconds)) == true
|
|
||||||
await node1.stop()
|
|
||||||
await node2.stop()
|
|
||||||
|
|||||||
@ -39,9 +39,9 @@ procSuite "WakuNode - Store":
|
|||||||
## Setup
|
## Setup
|
||||||
let
|
let
|
||||||
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60432))
|
||||||
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60430))
|
||||||
|
|
||||||
await allFutures(client.start(), server.start())
|
await allFutures(client.start(), server.start())
|
||||||
await server.mountStore(store=newTestMessageStore())
|
await server.mountStore(store=newTestMessageStore())
|
||||||
@ -73,34 +73,31 @@ procSuite "WakuNode - Store":
|
|||||||
## Setup
|
## Setup
|
||||||
let
|
let
|
||||||
filterSourceKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
filterSourceKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
filterSource = WakuNode.new(filterSourceKey, ValidIpAddress.init("0.0.0.0"), Port(60004))
|
filterSource = WakuNode.new(filterSourceKey, ValidIpAddress.init("0.0.0.0"), Port(60404))
|
||||||
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60402))
|
||||||
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60400))
|
||||||
|
|
||||||
await allFutures(client.start(), server.start(), filterSource.start())
|
await allFutures(client.start(), server.start(), filterSource.start())
|
||||||
|
|
||||||
await filterSource.mountFilter()
|
await filterSource.mountFilter()
|
||||||
await server.mountStore(store=newTestMessageStore())
|
await server.mountStore(store=newTestMessageStore())
|
||||||
await server.mountFilter()
|
await server.mountFilterClient()
|
||||||
await client.mountStore()
|
|
||||||
client.mountStoreClient()
|
client.mountStoreClient()
|
||||||
|
|
||||||
server.wakuFilter.setPeer(filterSource.peerInfo.toRemotePeerInfo())
|
|
||||||
|
|
||||||
## Given
|
## Given
|
||||||
let message = fakeWakuMessage()
|
let message = fakeWakuMessage()
|
||||||
let serverPeer = server.peerInfo.toRemotePeerInfo()
|
let
|
||||||
## Then
|
serverPeer = server.peerInfo.toRemotePeerInfo()
|
||||||
let filterFut = newFuture[bool]()
|
filterSourcePeer = filterSource.peerInfo.toRemotePeerInfo()
|
||||||
proc filterReqHandler(msg: WakuMessage) {.gcsafe, closure.} =
|
|
||||||
check:
|
|
||||||
msg == message
|
|
||||||
filterFut.complete(true)
|
|
||||||
|
|
||||||
let filterReq = FilterRequest(pubSubTopic: DefaultPubsubTopic, contentFilters: @[ContentFilter(contentTopic: DefaultContentTopic)], subscribe: true)
|
## Then
|
||||||
await server.subscribe(filterReq, filterReqHandler)
|
let filterFut = newFuture[(PubsubTopic, WakuMessage)]()
|
||||||
|
proc filterHandler(pubsubTopic: PubsubTopic, msg: WakuMessage) {.gcsafe, closure.} =
|
||||||
|
filterFut.complete((pubsubTopic, msg))
|
||||||
|
|
||||||
|
await server.filterSubscribe(DefaultPubsubTopic, DefaultContentTopic, filterHandler, peer=filterSourcePeer)
|
||||||
|
|
||||||
await sleepAsync(100.millis)
|
await sleepAsync(100.millis)
|
||||||
|
|
||||||
@ -108,7 +105,7 @@ procSuite "WakuNode - Store":
|
|||||||
await filterSource.wakuFilter.handleMessage(DefaultPubsubTopic, message)
|
await filterSource.wakuFilter.handleMessage(DefaultPubsubTopic, message)
|
||||||
|
|
||||||
# Wait for the server filter to receive the push message
|
# Wait for the server filter to receive the push message
|
||||||
require (await filterFut.withTimeout(5.seconds))
|
require await filterFut.withTimeout(5.seconds)
|
||||||
|
|
||||||
let res = await client.query(HistoryQuery(contentFilters: @[HistoryContentFilter(contentTopic: DefaultContentTopic)]), peer=serverPeer)
|
let res = await client.query(HistoryQuery(contentFilters: @[HistoryContentFilter(contentTopic: DefaultContentTopic)]), peer=serverPeer)
|
||||||
|
|
||||||
@ -120,6 +117,11 @@ procSuite "WakuNode - Store":
|
|||||||
response.messages.len == 1
|
response.messages.len == 1
|
||||||
response.messages[0] == message
|
response.messages[0] == message
|
||||||
|
|
||||||
|
let (handledPubsubTopic, handledMsg) = filterFut.read()
|
||||||
|
check:
|
||||||
|
handledPubsubTopic == DefaultPubsubTopic
|
||||||
|
handledMsg == message
|
||||||
|
|
||||||
## Cleanup
|
## Cleanup
|
||||||
await allFutures(client.stop(), server.stop(), filterSource.stop())
|
await allFutures(client.stop(), server.stop(), filterSource.stop())
|
||||||
|
|
||||||
@ -128,9 +130,9 @@ procSuite "WakuNode - Store":
|
|||||||
## Setup
|
## Setup
|
||||||
let
|
let
|
||||||
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60412))
|
||||||
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60410))
|
||||||
|
|
||||||
await allFutures(client.start(), server.start())
|
await allFutures(client.start(), server.start())
|
||||||
|
|
||||||
@ -161,9 +163,9 @@ procSuite "WakuNode - Store":
|
|||||||
## Setup
|
## Setup
|
||||||
let
|
let
|
||||||
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
serverKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60002))
|
server = WakuNode.new(serverKey, ValidIpAddress.init("0.0.0.0"), Port(60422))
|
||||||
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
clientKey = crypto.PrivateKey.random(Secp256k1, rng[])[]
|
||||||
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60000))
|
client = WakuNode.new(clientKey, ValidIpAddress.init("0.0.0.0"), Port(60420))
|
||||||
|
|
||||||
await allFutures(server.start(), client.start())
|
await allFutures(server.start(), client.start())
|
||||||
await server.mountStore(store=StoreQueueRef.new())
|
await server.mountStore(store=StoreQueueRef.new())
|
||||||
|
|||||||
@ -1,95 +1,100 @@
|
|||||||
{.push raises: [Defect].}
|
{.push raises: [Defect].}
|
||||||
|
|
||||||
import
|
import
|
||||||
std/[tables,sequtils],
|
std/[tables, sequtils],
|
||||||
chronicles,
|
chronicles,
|
||||||
json_rpc/rpcserver
|
json_rpc/rpcserver
|
||||||
import
|
import
|
||||||
../../protocol/waku_message,
|
../../protocol/waku_message,
|
||||||
../../protocol/waku_filter,
|
../../protocol/waku_filter,
|
||||||
|
../../protocol/waku_filter/client,
|
||||||
../waku_node,
|
../waku_node,
|
||||||
./jsonrpc_types
|
./jsonrpc_types
|
||||||
|
|
||||||
export jsonrpc_types
|
export jsonrpc_types
|
||||||
|
|
||||||
logScope:
|
logScope:
|
||||||
topics = "filter api"
|
topics = "wakunode.rpc.filter"
|
||||||
|
|
||||||
|
|
||||||
|
const DefaultPubsubTopic: PubsubTopic = "/waku/2/default-waku/proto"
|
||||||
const futTimeout* = 5.seconds # Max time to wait for futures
|
const futTimeout* = 5.seconds # Max time to wait for futures
|
||||||
const maxCache* = 30 # Max number of messages cached per topic @TODO make this configurable
|
const maxCache* = 30 # Max number of messages cached per topic TODO: make this configurable
|
||||||
|
|
||||||
|
|
||||||
proc installFilterApiHandlers*(node: WakuNode, rpcsrv: RpcServer, messageCache: MessageCache) =
|
proc installFilterApiHandlers*(node: WakuNode, rpcsrv: RpcServer, messageCache: MessageCache) =
|
||||||
|
|
||||||
proc filterHandler(msg: WakuMessage) {.gcsafe, closure, raises: [Defect].} =
|
|
||||||
# Add message to current cache
|
|
||||||
trace "WakuMessage received", msg=msg
|
|
||||||
|
|
||||||
# Make a copy of msgs for this topic to modify
|
|
||||||
var msgs = messageCache.getOrDefault(msg.contentTopic, @[])
|
|
||||||
|
|
||||||
if msgs.len >= maxCache:
|
|
||||||
# Message cache on this topic exceeds maximum. Delete oldest.
|
|
||||||
# @TODO this may become a bottle neck if called as the norm rather than exception when adding messages. Performance profile needed.
|
|
||||||
msgs.delete(0,0)
|
|
||||||
msgs.add(msg)
|
|
||||||
|
|
||||||
# Replace indexed entry with copy
|
|
||||||
# @TODO max number of content topics could be limited in node
|
|
||||||
messageCache[msg.contentTopic] = msgs
|
|
||||||
|
|
||||||
## Filter API version 1 definitions
|
## Filter API version 1 definitions
|
||||||
|
|
||||||
rpcsrv.rpc("get_waku_v2_filter_v1_messages") do(contentTopic: ContentTopic) -> seq[WakuMessage]:
|
rpcsrv.rpc("get_waku_v2_filter_v1_messages") do (contentTopic: ContentTopic) -> seq[WakuMessage]:
|
||||||
## Returns all WakuMessages received on a content topic since the
|
## Returns all WakuMessages received on a content topic since the
|
||||||
## last time this method was called
|
## last time this method was called
|
||||||
## @TODO ability to specify a return message limit
|
## TODO: ability to specify a return message limit
|
||||||
debug "get_waku_v2_filter_v1_messages", contentTopic=contentTopic
|
debug "get_waku_v2_filter_v1_messages", contentTopic=contentTopic
|
||||||
|
|
||||||
if messageCache.hasKey(contentTopic):
|
if not messageCache.hasKey(contentTopic):
|
||||||
let msgs = messageCache[contentTopic]
|
|
||||||
# Clear cache before next call
|
|
||||||
messageCache[contentTopic] = @[]
|
|
||||||
return msgs
|
|
||||||
else:
|
|
||||||
# Not subscribed to this content topic
|
|
||||||
raise newException(ValueError, "Not subscribed to content topic: " & $contentTopic)
|
raise newException(ValueError, "Not subscribed to content topic: " & $contentTopic)
|
||||||
|
|
||||||
|
let msgs = messageCache[contentTopic]
|
||||||
|
# Clear cache before next call
|
||||||
|
messageCache[contentTopic] = @[]
|
||||||
|
return msgs
|
||||||
|
|
||||||
rpcsrv.rpc("post_waku_v2_filter_v1_subscription") do(contentFilters: seq[ContentFilter], topic: Option[string]) -> bool:
|
|
||||||
|
rpcsrv.rpc("post_waku_v2_filter_v1_subscription") do (contentFilters: seq[ContentFilter], topic: Option[string]) -> bool:
|
||||||
## Subscribes a node to a list of content filters
|
## Subscribes a node to a list of content filters
|
||||||
debug "post_waku_v2_filter_v1_subscription"
|
debug "post_waku_v2_filter_v1_subscription"
|
||||||
|
|
||||||
# Construct a filter request
|
let
|
||||||
# @TODO use default PubSub topic if undefined
|
pubsubTopic: string = topic.get(DefaultPubsubTopic)
|
||||||
let fReq = if topic.isSome: FilterRequest(pubSubTopic: topic.get, contentFilters: contentFilters, subscribe: true) else: FilterRequest(contentFilters: contentFilters, subscribe: true)
|
contentTopics: seq[ContentTopic] = contentFilters.mapIt(it.contentTopic)
|
||||||
|
|
||||||
|
let pushHandler:FilterPushHandler = proc(pubsubTopic: string, msg: WakuMessage) {.gcsafe, closure.} =
|
||||||
|
# Add message to current cache
|
||||||
|
trace "WakuMessage received", msg=msg
|
||||||
|
|
||||||
|
# Make a copy of msgs for this topic to modify
|
||||||
|
var msgs = messageCache.getOrDefault(msg.contentTopic, @[])
|
||||||
|
|
||||||
|
if msgs.len >= maxCache:
|
||||||
|
# Message cache on this topic exceeds maximum. Delete oldest.
|
||||||
|
# TODO: this may become a bottle neck if called as the norm rather than exception when adding messages. Performance profile needed.
|
||||||
|
msgs.delete(0,0)
|
||||||
|
|
||||||
|
msgs.add(msg)
|
||||||
|
|
||||||
|
# Replace indexed entry with copy
|
||||||
|
# TODO: max number of content topics could be limited in node
|
||||||
|
messageCache[msg.contentTopic] = msgs
|
||||||
|
|
||||||
|
let subFut = node.subscribe(pubsubTopic, contentTopics, pushHandler)
|
||||||
|
|
||||||
|
if not await subFut.withTimeout(futTimeout):
|
||||||
|
raise newException(ValueError, "Failed to subscribe to contentFilters")
|
||||||
|
|
||||||
|
# Successfully subscribed to all content filters
|
||||||
|
for cTopic in contentTopics:
|
||||||
|
# Create message cache for each subscribed content topic
|
||||||
|
messageCache[cTopic] = @[]
|
||||||
|
|
||||||
if (await node.subscribe(fReq, filterHandler).withTimeout(futTimeout)):
|
return true
|
||||||
# Successfully subscribed to all content filters
|
|
||||||
|
|
||||||
for cTopic in contentFilters.mapIt(it.contentTopic):
|
|
||||||
# Create message cache for each subscribed content topic
|
|
||||||
messageCache[cTopic] = @[]
|
|
||||||
|
|
||||||
return true
|
|
||||||
else:
|
|
||||||
# Failed to subscribe to one or more content filters
|
|
||||||
raise newException(ValueError, "Failed to subscribe to contentFilters " & repr(fReq))
|
|
||||||
|
|
||||||
rpcsrv.rpc("delete_waku_v2_filter_v1_subscription") do(contentFilters: seq[ContentFilter], topic: Option[string]) -> bool:
|
rpcsrv.rpc("delete_waku_v2_filter_v1_subscription") do(contentFilters: seq[ContentFilter], topic: Option[string]) -> bool:
|
||||||
## Unsubscribes a node from a list of content filters
|
## Unsubscribes a node from a list of content filters
|
||||||
debug "delete_waku_v2_filter_v1_subscription"
|
debug "delete_waku_v2_filter_v1_subscription"
|
||||||
|
|
||||||
# Construct a filter request
|
let
|
||||||
# @TODO consider using default PubSub topic if undefined
|
pubsubTopic: string = topic.get(DefaultPubsubTopic)
|
||||||
let fReq = if topic.isSome: FilterRequest(pubSubTopic: topic.get, contentFilters: contentFilters, subscribe: false) else: FilterRequest(contentFilters: contentFilters, subscribe: false)
|
contentTopics: seq[ContentTopic] = contentFilters.mapIt(it.contentTopic)
|
||||||
|
|
||||||
if (await node.unsubscribe(fReq).withTimeout(futTimeout)):
|
let unsubFut = node.unsubscribe(pubsubTopic, contentTopics)
|
||||||
# Successfully unsubscribed from all content filters
|
|
||||||
|
|
||||||
for cTopic in contentFilters.mapIt(it.contentTopic):
|
if not await unsubFut.withTimeout(futTimeout):
|
||||||
# Remove message cache for each unsubscribed content topic
|
raise newException(ValueError, "Failed to unsubscribe from contentFilters")
|
||||||
messageCache.del(cTopic)
|
|
||||||
|
# Successfully unsubscribed from all content filters
|
||||||
|
for cTopic in contentTopics:
|
||||||
|
# Remove message cache for each unsubscribed content topic
|
||||||
|
messageCache.del(cTopic)
|
||||||
|
|
||||||
return true
|
return true
|
||||||
else:
|
|
||||||
# Failed to unsubscribe from one or more content filters
|
|
||||||
raise newException(ValueError, "Failed to unsubscribe from contentFilters " & repr(fReq))
|
|
||||||
|
|||||||
@ -1,14 +1,13 @@
|
|||||||
{.push raises: [Defect].}
|
{.push raises: [Defect].}
|
||||||
|
|
||||||
import
|
import
|
||||||
stew/results,
|
|
||||||
stew/shims/net,
|
stew/shims/net,
|
||||||
chronicles,
|
chronicles,
|
||||||
chronos,
|
chronos,
|
||||||
metrics,
|
metrics,
|
||||||
metrics/chronos_httpserver
|
metrics/chronos_httpserver
|
||||||
import
|
import
|
||||||
../protocol/waku_filter,
|
../protocol/waku_filter/protocol_metrics as filter_metrics,
|
||||||
../protocol/waku_store/protocol_metrics as store_metrics,
|
../protocol/waku_store/protocol_metrics as store_metrics,
|
||||||
../protocol/waku_lightpush/protocol_metrics as lightpush_metrics,
|
../protocol/waku_lightpush/protocol_metrics as lightpush_metrics,
|
||||||
../protocol/waku_swap/waku_swap,
|
../protocol/waku_swap/waku_swap,
|
||||||
|
|||||||
@ -21,11 +21,13 @@ import
|
|||||||
../protocol/waku_store/client,
|
../protocol/waku_store/client,
|
||||||
../protocol/waku_swap/waku_swap,
|
../protocol/waku_swap/waku_swap,
|
||||||
../protocol/waku_filter,
|
../protocol/waku_filter,
|
||||||
|
../protocol/waku_filter/client,
|
||||||
../protocol/waku_lightpush,
|
../protocol/waku_lightpush,
|
||||||
../protocol/waku_lightpush/client,
|
../protocol/waku_lightpush/client,
|
||||||
../protocol/waku_rln_relay/waku_rln_relay_types,
|
../protocol/waku_rln_relay/waku_rln_relay_types,
|
||||||
../protocol/waku_peer_exchange,
|
../protocol/waku_peer_exchange,
|
||||||
../utils/[peers, requests, wakuenr],
|
../utils/peers,
|
||||||
|
../utils/wakuenr,
|
||||||
./peer_manager/peer_manager,
|
./peer_manager/peer_manager,
|
||||||
./storage/message/waku_store_queue,
|
./storage/message/waku_store_queue,
|
||||||
./storage/message/message_retention_policy,
|
./storage/message/message_retention_policy,
|
||||||
@ -37,24 +39,26 @@ import
|
|||||||
|
|
||||||
declarePublicGauge waku_version, "Waku version info (in git describe format)", ["version"]
|
declarePublicGauge waku_version, "Waku version info (in git describe format)", ["version"]
|
||||||
declarePublicCounter waku_node_messages, "number of messages received", ["type"]
|
declarePublicCounter waku_node_messages, "number of messages received", ["type"]
|
||||||
declarePublicGauge waku_node_filters, "number of content filter subscriptions"
|
|
||||||
declarePublicGauge waku_node_errors, "number of wakunode errors", ["type"]
|
declarePublicGauge waku_node_errors, "number of wakunode errors", ["type"]
|
||||||
declarePublicGauge waku_lightpush_peers, "number of lightpush peers"
|
declarePublicGauge waku_lightpush_peers, "number of lightpush peers"
|
||||||
|
declarePublicGauge waku_filter_peers, "number of filter peers"
|
||||||
declarePublicGauge waku_store_peers, "number of store peers"
|
declarePublicGauge waku_store_peers, "number of store peers"
|
||||||
declarePublicGauge waku_px_peers, "number of peers (in the node's peerManager) supporting the peer exchange protocol"
|
declarePublicGauge waku_px_peers, "number of peers (in the node's peerManager) supporting the peer exchange protocol"
|
||||||
|
|
||||||
|
|
||||||
logScope:
|
logScope:
|
||||||
topics = "wakunode"
|
topics = "wakunode"
|
||||||
|
|
||||||
|
|
||||||
# Git version in git describe format (defined compile time)
|
# Git version in git describe format (defined compile time)
|
||||||
const git_version* {.strdefine.} = "n/a"
|
const git_version* {.strdefine.} = "n/a"
|
||||||
|
|
||||||
# Default clientId
|
# Default clientId
|
||||||
const clientId* = "Nimbus Waku v2 node"
|
const clientId* = "Nimbus Waku v2 node"
|
||||||
|
|
||||||
# Default topic
|
# TODO: Unify pubusub topic type and default value
|
||||||
const defaultTopic* = "/waku/2/default-waku/proto"
|
type PubsubTopic* = string
|
||||||
|
|
||||||
|
const defaultTopic*: PubsubTopic = "/waku/2/default-waku/proto"
|
||||||
|
|
||||||
# Default Waku Filter Timeout
|
# Default Waku Filter Timeout
|
||||||
const WakuFilterTimeout: Duration = 1.days
|
const WakuFilterTimeout: Duration = 1.days
|
||||||
@ -63,7 +67,6 @@ const WakuFilterTimeout: Duration = 1.days
|
|||||||
# key and crypto modules different
|
# key and crypto modules different
|
||||||
type
|
type
|
||||||
# XXX: Weird type, should probably be using pubsub PubsubTopic object name?
|
# XXX: Weird type, should probably be using pubsub PubsubTopic object name?
|
||||||
PubsubTopic* = string
|
|
||||||
Message* = seq[byte]
|
Message* = seq[byte]
|
||||||
|
|
||||||
WakuInfo* = object
|
WakuInfo* = object
|
||||||
@ -80,6 +83,7 @@ type
|
|||||||
wakuStore*: WakuStore
|
wakuStore*: WakuStore
|
||||||
wakuStoreClient*: WakuStoreClient
|
wakuStoreClient*: WakuStoreClient
|
||||||
wakuFilter*: WakuFilter
|
wakuFilter*: WakuFilter
|
||||||
|
wakuFilterClient*: WakuFilterClient
|
||||||
wakuSwap*: WakuSwap
|
wakuSwap*: WakuSwap
|
||||||
wakuRlnRelay*: WakuRLNRelay
|
wakuRlnRelay*: WakuRLNRelay
|
||||||
wakuLightPush*: WakuLightPush
|
wakuLightPush*: WakuLightPush
|
||||||
@ -87,7 +91,6 @@ type
|
|||||||
wakuPeerExchange*: WakuPeerExchange
|
wakuPeerExchange*: WakuPeerExchange
|
||||||
enr*: enr.Record
|
enr*: enr.Record
|
||||||
libp2pPing*: Ping
|
libp2pPing*: Ping
|
||||||
filters*: Filters
|
|
||||||
rng*: ref rand.HmacDrbgContext
|
rng*: ref rand.HmacDrbgContext
|
||||||
wakuDiscv5*: WakuDiscoveryV5
|
wakuDiscv5*: WakuDiscoveryV5
|
||||||
announcedAddresses* : seq[MultiAddress]
|
announcedAddresses* : seq[MultiAddress]
|
||||||
@ -217,7 +220,6 @@ proc new*(T: type WakuNode,
|
|||||||
switch: switch,
|
switch: switch,
|
||||||
rng: rng,
|
rng: rng,
|
||||||
enr: enr,
|
enr: enr,
|
||||||
filters: Filters.init(),
|
|
||||||
announcedAddresses: announcedAddresses
|
announcedAddresses: announcedAddresses
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -410,80 +412,123 @@ proc mountRelay*(node: WakuNode,
|
|||||||
## Waku filter
|
## Waku filter
|
||||||
|
|
||||||
proc mountFilter*(node: WakuNode, filterTimeout: Duration = WakuFilterTimeout) {.async, raises: [Defect, LPError]} =
|
proc mountFilter*(node: WakuNode, filterTimeout: Duration = WakuFilterTimeout) {.async, raises: [Defect, LPError]} =
|
||||||
info "mounting filter"
|
info "mounting filter protocol"
|
||||||
proc filterHandler(requestId: string, msg: MessagePush) {.async, gcsafe.} =
|
node.wakuFilter = WakuFilter.new(node.peerManager, node.rng, filterTimeout)
|
||||||
|
|
||||||
info "push received"
|
|
||||||
for message in msg.messages:
|
|
||||||
node.filters.notify(message, requestId) # Trigger filter handlers on a light node
|
|
||||||
|
|
||||||
if not node.wakuStore.isNil and (requestId in node.filters):
|
|
||||||
let pubSubTopic = node.filters[requestId].pubSubTopic
|
|
||||||
node.wakuStore.handleMessage(pubSubTopic, message)
|
|
||||||
|
|
||||||
waku_node_messages.inc(labelValues = ["filter"])
|
|
||||||
|
|
||||||
node.wakuFilter = WakuFilter.init(node.peerManager, node.rng, filterHandler, filterTimeout)
|
|
||||||
if node.started:
|
if node.started:
|
||||||
# Node has started already. Let's start filter too.
|
|
||||||
await node.wakuFilter.start()
|
await node.wakuFilter.start()
|
||||||
|
|
||||||
node.switch.mount(node.wakuFilter, protocolMatcher(WakuFilterCodec))
|
node.switch.mount(node.wakuFilter, protocolMatcher(WakuFilterCodec))
|
||||||
|
|
||||||
proc setFilterPeer*(node: WakuNode, peer: RemotePeerInfo|string) {.raises: [Defect, ValueError, LPError].} =
|
proc filterHandleMessage*(node: WakuNode, pubsubTopic: PubsubTopic, message: WakuMessage) {.async.}=
|
||||||
if node.wakuFilter.isNil():
|
if node.wakuFilter.isNil():
|
||||||
error "could not set peer, waku filter is nil"
|
error "cannot handle filter message", error="waku filter is nil"
|
||||||
return
|
return
|
||||||
|
|
||||||
info "Set filter peer", peer=peer
|
await node.wakuFilter.handleMessage(pubsubTopic, message)
|
||||||
|
|
||||||
|
|
||||||
|
proc mountFilterClient*(node: WakuNode) {.async, raises: [Defect, LPError].} =
|
||||||
|
info "mounting filter client"
|
||||||
|
|
||||||
|
node.wakuFilterClient = WakuFilterClient.new(node.peerManager, node.rng)
|
||||||
|
if node.started:
|
||||||
|
# Node has started already. Let's start filter too.
|
||||||
|
await node.wakuFilterClient.start()
|
||||||
|
|
||||||
|
node.switch.mount(node.wakuFilterClient, protocolMatcher(WakuFilterCodec))
|
||||||
|
|
||||||
|
proc filterSubscribe*(node: WakuNode, pubsubTopic: PubsubTopic, contentTopics: ContentTopic|seq[ContentTopic],
|
||||||
|
handler: FilterPushHandler, peer: RemotePeerInfo|string) {.async, gcsafe, raises: [Defect, ValueError].} =
|
||||||
|
## Registers for messages that match a specific filter. Triggers the handler whenever a message is received.
|
||||||
|
if node.wakuFilterClient.isNil():
|
||||||
|
error "cannot register filter subscription to topic", error="waku filter client is nil"
|
||||||
|
return
|
||||||
|
|
||||||
|
let remotePeer = when peer is string: parseRemotePeerInfo(peer)
|
||||||
|
else: peer
|
||||||
|
|
||||||
|
info "registering filter subscription to content", pubsubTopic=pubsubTopic, contentTopics=contentTopics, peer=remotePeer
|
||||||
|
|
||||||
|
# Add handler wrapper to store the message when pushed, when relay is disabled and filter enabled
|
||||||
|
# TODO: Move this logic to wakunode2 app
|
||||||
|
let handlerWrapper: FilterPushHandler = proc(pubsubTopic: string, message: WakuMessage) {.raises: [Exception].} =
|
||||||
|
if node.wakuRelay.isNil() and not node.wakuStore.isNil():
|
||||||
|
node.wakuStore.handleMessage(pubSubTopic, message)
|
||||||
|
|
||||||
|
handler(pubsubTopic, message)
|
||||||
|
|
||||||
|
let subRes = await node.wakuFilterClient.subscribe(pubsubTopic, contentTopics, handlerWrapper, peer=remotePeer)
|
||||||
|
if subRes.isOk():
|
||||||
|
info "subscribed to topic", pubsubTopic=pubsubTopic, contentTopics=contentTopics
|
||||||
|
else:
|
||||||
|
error "failed filter subscription", error=subRes.error
|
||||||
|
waku_node_errors.inc(labelValues = ["subscribe_filter_failure"])
|
||||||
|
|
||||||
|
proc filterUnsubscribe*(node: WakuNode, pubsubTopic: PubsubTopic, contentTopics: ContentTopic|seq[ContentTopic],
|
||||||
|
peer: RemotePeerInfo|string) {.async, gcsafe, raises: [Defect, ValueError].} =
|
||||||
|
## Unsubscribe from a content filter.
|
||||||
|
if node.wakuFilterClient.isNil():
|
||||||
|
error "cannot unregister filter subscription to content", error="waku filter client is nil"
|
||||||
|
return
|
||||||
|
|
||||||
|
let remotePeer = when peer is string: parseRemotePeerInfo(peer)
|
||||||
|
else: peer
|
||||||
|
|
||||||
|
info "deregistering filter subscription to content", pubsubTopic=pubsubTopic, contentTopics=contentTopics, peer=remotePeer
|
||||||
|
|
||||||
|
let unsubRes = await node.wakuFilterClient.unsubscribe(pubsubTopic, contentTopics, peer=remotePeer)
|
||||||
|
if unsubRes.isOk():
|
||||||
|
info "unsubscribed from topic", pubsubTopic=pubsubTopic, contentTopics=contentTopics
|
||||||
|
else:
|
||||||
|
error "failed filter unsubscription", error=unsubRes.error
|
||||||
|
waku_node_errors.inc(labelValues = ["unsubscribe_filter_failure"])
|
||||||
|
|
||||||
|
|
||||||
|
# TODO: Move to application module (e.g., wakunode2.nim)
|
||||||
|
proc setFilterPeer*(node: WakuNode, peer: RemotePeerInfo|string) {.raises: [Defect, ValueError, LPError],
|
||||||
|
deprecated: "Use the explicit destination peer procedures".} =
|
||||||
|
if node.wakuFilterClient.isNil():
|
||||||
|
error "could not set peer, waku filter client is nil"
|
||||||
|
return
|
||||||
|
|
||||||
|
info "seting filter client peer", peer=peer
|
||||||
|
|
||||||
let remotePeer = when peer is string: parseRemotePeerInfo(peer)
|
let remotePeer = when peer is string: parseRemotePeerInfo(peer)
|
||||||
else: peer
|
else: peer
|
||||||
node.wakuFilter.setPeer(remotePeer)
|
node.peerManager.addPeer(remotePeer, WakuFilterCodec)
|
||||||
|
|
||||||
proc subscribe*(node: WakuNode, request: FilterRequest, handler: ContentFilterHandler) {.async, gcsafe.} =
|
waku_filter_peers.inc()
|
||||||
|
|
||||||
|
# TODO: Move to application module (e.g., wakunode2.nim)
|
||||||
|
proc subscribe*(node: WakuNode, pubsubTopic: PubsubTopic, contentTopics: ContentTopic|seq[ContentTopic], handler: FilterPushHandler) {.async, gcsafe,
|
||||||
|
deprecated: "Use the explicit destination peer procedure. Use 'node.filterSubscribe()' instead.".} =
|
||||||
## Registers for messages that match a specific filter. Triggers the handler whenever a message is received.
|
## Registers for messages that match a specific filter. Triggers the handler whenever a message is received.
|
||||||
## FilterHandler is a method that takes a MessagePush.
|
if node.wakuFilterClient.isNil():
|
||||||
|
error "cannot register filter subscription to topic", error="waku filter client is nil"
|
||||||
|
return
|
||||||
|
|
||||||
# Sanity check for well-formed subscribe FilterRequest
|
let peerOpt = node.peerManager.selectPeer(WakuFilterCodec)
|
||||||
doAssert(request.subscribe, "invalid subscribe request")
|
if peerOpt.isNone():
|
||||||
|
error "cannot register filter subscription to topic", error="no suitable remote peers"
|
||||||
|
return
|
||||||
|
|
||||||
info "subscribe content", filter=request
|
await node.filterSubscribe(pubsubTopic, contentTopics, handler, peer=peerOpt.get())
|
||||||
|
|
||||||
var id = generateRequestId(node.rng)
|
# TODO: Move to application module (e.g., wakunode2.nim)
|
||||||
|
proc unsubscribe*(node: WakuNode, pubsubTopic: PubsubTopic, contentTopics: ContentTopic|seq[ContentTopic]) {.async, gcsafe,
|
||||||
if not node.wakuFilter.isNil():
|
deprecated: "Use the explicit destination peer procedure. Use 'node.filterUnsusbscribe()' instead.".} =
|
||||||
let
|
|
||||||
pubsubTopic = request.pubsubTopic
|
|
||||||
contentTopics = request.contentFilters.mapIt(it.contentTopic)
|
|
||||||
|
|
||||||
let resSubscription = await node.wakuFilter.subscribe(pubsubTopic, contentTopics)
|
|
||||||
if resSubscription.isOk():
|
|
||||||
id = resSubscription.get()
|
|
||||||
else:
|
|
||||||
# Failed to subscribe
|
|
||||||
error "remote subscription to filter failed", filter = request
|
|
||||||
waku_node_errors.inc(labelValues = ["subscribe_filter_failure"])
|
|
||||||
|
|
||||||
# Register handler for filter, whether remote subscription succeeded or not
|
|
||||||
node.filters.addContentFilters(id, request.pubSubTopic, request.contentFilters, handler)
|
|
||||||
waku_node_filters.set(node.filters.len.int64)
|
|
||||||
|
|
||||||
proc unsubscribe*(node: WakuNode, request: FilterRequest) {.async, gcsafe.} =
|
|
||||||
## Unsubscribe from a content filter.
|
## Unsubscribe from a content filter.
|
||||||
|
if node.wakuFilterClient.isNil():
|
||||||
|
error "cannot unregister filter subscription to content", error="waku filter client is nil"
|
||||||
|
return
|
||||||
|
|
||||||
# Sanity check for well-formed unsubscribe FilterRequest
|
let peerOpt = node.peerManager.selectPeer(WakuFilterCodec)
|
||||||
doAssert(request.subscribe == false, "invalid unsubscribe request")
|
if peerOpt.isNone():
|
||||||
|
error "cannot register filter subscription to topic", error="no suitable remote peers"
|
||||||
info "unsubscribe content", filter=request
|
return
|
||||||
|
|
||||||
let
|
|
||||||
pubsubTopic = request.pubsubTopic
|
|
||||||
contentTopics = request.contentFilters.mapIt(it.contentTopic)
|
|
||||||
discard await node.wakuFilter.unsubscribe(pubsubTopic, contentTopics)
|
|
||||||
node.filters.removeContentFilters(request.contentFilters)
|
|
||||||
|
|
||||||
waku_node_filters.set(node.filters.len.int64)
|
await node.filterUnsubscribe(pubsubTopic, contentTopics, peer=peerOpt.get())
|
||||||
|
|
||||||
|
|
||||||
## Waku swap
|
## Waku swap
|
||||||
@ -503,11 +548,6 @@ proc mountSwap*(node: WakuNode, swapConfig: SwapConfig = SwapConfig.init()) {.as
|
|||||||
|
|
||||||
## Waku store
|
## Waku store
|
||||||
|
|
||||||
proc mountStoreClient*(node: WakuNode, store: MessageStore = nil) =
|
|
||||||
info "mounting store client"
|
|
||||||
|
|
||||||
node.wakuStoreClient = WakuStoreClient.new(node.peerManager, node.rng, store)
|
|
||||||
|
|
||||||
const MessageStoreDefaultRetentionPolicyInterval* = 30.minutes
|
const MessageStoreDefaultRetentionPolicyInterval* = 30.minutes
|
||||||
|
|
||||||
proc executeMessageRetentionPolicy*(node: WakuNode) =
|
proc executeMessageRetentionPolicy*(node: WakuNode) =
|
||||||
@ -557,6 +597,12 @@ proc mountStore*(node: WakuNode, store: MessageStore = nil, retentionPolicy=none
|
|||||||
|
|
||||||
node.switch.mount(node.wakuStore, protocolMatcher(WakuStoreCodec))
|
node.switch.mount(node.wakuStore, protocolMatcher(WakuStoreCodec))
|
||||||
|
|
||||||
|
|
||||||
|
proc mountStoreClient*(node: WakuNode, store: MessageStore = nil) =
|
||||||
|
info "mounting store client"
|
||||||
|
|
||||||
|
node.wakuStoreClient = WakuStoreClient.new(node.peerManager, node.rng, store)
|
||||||
|
|
||||||
proc query*(node: WakuNode, query: HistoryQuery, peer: RemotePeerInfo): Future[WakuStoreResult[HistoryResponse]] {.async, gcsafe.} =
|
proc query*(node: WakuNode, query: HistoryQuery, peer: RemotePeerInfo): Future[WakuStoreResult[HistoryResponse]] {.async, gcsafe.} =
|
||||||
## Queries known nodes for historical messages
|
## Queries known nodes for historical messages
|
||||||
if node.wakuStoreClient.isNil():
|
if node.wakuStoreClient.isNil():
|
||||||
@ -626,12 +672,6 @@ proc resume*(node: WakuNode, peerList: Option[seq[RemotePeerInfo]] = none(seq[Re
|
|||||||
|
|
||||||
## Waku lightpush
|
## Waku lightpush
|
||||||
|
|
||||||
proc mountLightPushClient*(node: WakuNode) =
|
|
||||||
info "mounting light push client"
|
|
||||||
|
|
||||||
node.wakuLightpushClient = WakuLightPushClient.new(node.peerManager, node.rng)
|
|
||||||
|
|
||||||
|
|
||||||
proc mountLightPush*(node: WakuNode) {.async.} =
|
proc mountLightPush*(node: WakuNode) {.async.} =
|
||||||
info "mounting light push"
|
info "mounting light push"
|
||||||
|
|
||||||
@ -654,6 +694,12 @@ proc mountLightPush*(node: WakuNode) {.async.} =
|
|||||||
|
|
||||||
node.switch.mount(node.wakuLightPush, protocolMatcher(WakuLightPushCodec))
|
node.switch.mount(node.wakuLightPush, protocolMatcher(WakuLightPushCodec))
|
||||||
|
|
||||||
|
|
||||||
|
proc mountLightPushClient*(node: WakuNode) =
|
||||||
|
info "mounting light push client"
|
||||||
|
|
||||||
|
node.wakuLightpushClient = WakuLightPushClient.new(node.peerManager, node.rng)
|
||||||
|
|
||||||
proc lightpushPublish*(node: WakuNode, pubsubTopic: PubsubTopic, message: WakuMessage, peer: RemotePeerInfo): Future[WakuLightPushResult[void]] {.async, gcsafe.} =
|
proc lightpushPublish*(node: WakuNode, pubsubTopic: PubsubTopic, message: WakuMessage, peer: RemotePeerInfo): Future[WakuLightPushResult[void]] {.async, gcsafe.} =
|
||||||
## Pushes a `WakuMessage` to a node which relays it further on PubSub topic.
|
## Pushes a `WakuMessage` to a node which relays it further on PubSub topic.
|
||||||
## Returns whether relaying was successful or not.
|
## Returns whether relaying was successful or not.
|
||||||
|
|||||||
@ -3,11 +3,9 @@
|
|||||||
import
|
import
|
||||||
./waku_filter/rpc,
|
./waku_filter/rpc,
|
||||||
./waku_filter/rpc_codec,
|
./waku_filter/rpc_codec,
|
||||||
./waku_filter/protocol,
|
./waku_filter/protocol
|
||||||
./waku_filter/client
|
|
||||||
|
|
||||||
export
|
export
|
||||||
rpc,
|
rpc,
|
||||||
rpc_codec,
|
rpc_codec,
|
||||||
protocol,
|
protocol
|
||||||
client
|
|
||||||
|
|||||||
@ -1,69 +1,207 @@
|
|||||||
{.push raises: [Defect].}
|
{.push raises: [Defect].}
|
||||||
|
|
||||||
import
|
import
|
||||||
std/[tables, sequtils],
|
std/[options, tables, sequtils],
|
||||||
chronicles
|
stew/results,
|
||||||
|
chronicles,
|
||||||
|
chronos,
|
||||||
|
metrics,
|
||||||
|
bearssl/rand,
|
||||||
|
libp2p/protocols/protocol as libp2p_protocol
|
||||||
import
|
import
|
||||||
../waku_message,
|
../waku_message,
|
||||||
./rpc
|
../../node/peer_manager/peer_manager,
|
||||||
|
../../utils/requests,
|
||||||
type
|
./rpc,
|
||||||
ContentFilterHandler* = proc(msg: WakuMessage) {.gcsafe, closure, raises: [Defect].}
|
./rpc_codec,
|
||||||
|
./protocol,
|
||||||
Filter* = object
|
./protocol_metrics
|
||||||
pubSubTopic*: string
|
|
||||||
contentFilters*: seq[ContentFilter]
|
|
||||||
handler*: ContentFilterHandler
|
|
||||||
|
|
||||||
Filters* = Table[string, Filter]
|
|
||||||
|
|
||||||
|
|
||||||
proc init*(T: type Filters): T =
|
logScope:
|
||||||
initTable[string, Filter]()
|
topics = "wakufilter.client"
|
||||||
|
|
||||||
proc addContentFilters*(filters: var Filters, requestId: string, pubsubTopic: string, contentFilters: seq[ContentFilter], handler: ContentFilterHandler) {.gcsafe.}=
|
|
||||||
filters[requestId] = Filter(
|
const Defaultstring = "/waku/2/default-waku/proto"
|
||||||
pubSubTopic: pubsubTopic,
|
|
||||||
contentFilters: contentFilters,
|
|
||||||
handler: handler
|
### Client, filter subscripton manager
|
||||||
|
|
||||||
|
type FilterPushHandler* = proc(pubsubTopic: string, message: WakuMessage) {.gcsafe, closure.}
|
||||||
|
|
||||||
|
|
||||||
|
## Subscription manager
|
||||||
|
|
||||||
|
type SubscriptionManager = object
|
||||||
|
subscriptions: TableRef[(string, ContentTopic), FilterPushHandler]
|
||||||
|
|
||||||
|
proc init(T: type SubscriptionManager): T =
|
||||||
|
SubscriptionManager(subscriptions: newTable[(string, ContentTopic), FilterPushHandler]())
|
||||||
|
|
||||||
|
proc clear(m: var SubscriptionManager) =
|
||||||
|
m.subscriptions.clear()
|
||||||
|
|
||||||
|
proc registerSubscription(m: SubscriptionManager, pubsubTopic: string, contentTopic: ContentTopic, handler: FilterPushHandler) =
|
||||||
|
try:
|
||||||
|
m.subscriptions[(pubsubTopic, contentTopic)]= handler
|
||||||
|
except:
|
||||||
|
error "failed to register filter subscription", error=getCurrentExceptionMsg()
|
||||||
|
|
||||||
|
proc removeSubscription(m: SubscriptionManager, pubsubTopic: string, contentTopic: ContentTopic) =
|
||||||
|
m.subscriptions.del((pubsubTopic, contentTopic))
|
||||||
|
|
||||||
|
proc notifySubscriptionHandler(m: SubscriptionManager, pubsubTopic: string, contentTopic: ContentTopic, message: WakuMessage) =
|
||||||
|
if not m.subscriptions.hasKey((pubsubTopic, contentTopic)):
|
||||||
|
return
|
||||||
|
|
||||||
|
try:
|
||||||
|
let handler = m.subscriptions[(pubsubTopic, contentTopic)]
|
||||||
|
handler(pubsubTopic, message)
|
||||||
|
except:
|
||||||
|
discard
|
||||||
|
|
||||||
|
proc getSubscriptionsCount(m: SubscriptionManager): int =
|
||||||
|
m.subscriptions.len()
|
||||||
|
|
||||||
|
|
||||||
|
## Client
|
||||||
|
|
||||||
|
type MessagePushHandler* = proc(requestId: string, msg: MessagePush): Future[void] {.gcsafe, closure.}
|
||||||
|
|
||||||
|
type WakuFilterClient* = ref object of LPProtocol
|
||||||
|
rng: ref rand.HmacDrbgContext
|
||||||
|
peerManager: PeerManager
|
||||||
|
subManager: SubscriptionManager
|
||||||
|
|
||||||
|
|
||||||
|
proc handleMessagePush(wf: WakuFilterClient, peerId: PeerId, requestId: string, rpc: MessagePush) =
|
||||||
|
for msg in rpc.messages:
|
||||||
|
let
|
||||||
|
pubsubTopic = Defaultstring # TODO: Extend the filter push rpc to provide the pubsub topic. This is a limitation
|
||||||
|
contentTopic = msg.contentTopic
|
||||||
|
|
||||||
|
wf.subManager.notifySubscriptionHandler(pubsubTopic, contentTopic, msg)
|
||||||
|
|
||||||
|
|
||||||
|
proc initProtocolHandler(wf: WakuFilterClient) =
|
||||||
|
proc handle(conn: Connection, proto: string) {.async, gcsafe, closure.} =
|
||||||
|
let buffer = await conn.readLp(MaxRpcSize.int)
|
||||||
|
|
||||||
|
let decodeReqRes = FilterRPC.init(buffer)
|
||||||
|
if decodeReqRes.isErr():
|
||||||
|
waku_filter_errors.inc(labelValues = [decodeRpcFailure])
|
||||||
|
return
|
||||||
|
|
||||||
|
let rpc = decodeReqRes.get()
|
||||||
|
trace "filter message received"
|
||||||
|
|
||||||
|
if rpc.push == MessagePush():
|
||||||
|
waku_filter_errors.inc(labelValues = [emptyMessagePushFailure])
|
||||||
|
# TODO: Manage the empty push message error. Perform any action?
|
||||||
|
return
|
||||||
|
|
||||||
|
waku_filter_messages.inc(labelValues = ["MessagePush"])
|
||||||
|
|
||||||
|
let
|
||||||
|
peerId = conn.peerId
|
||||||
|
requestId = rpc.requestId
|
||||||
|
push = rpc.push
|
||||||
|
|
||||||
|
info "received filter message push", peerId=conn.peerId, requestId=requestId
|
||||||
|
wf.handleMessagePush(peerId, requestId, push)
|
||||||
|
|
||||||
|
wf.handler = handle
|
||||||
|
wf.codec = WakuFilterCodec
|
||||||
|
|
||||||
|
proc new*(T: type WakuFilterClient,
|
||||||
|
peerManager: PeerManager,
|
||||||
|
rng: ref rand.HmacDrbgContext): T =
|
||||||
|
|
||||||
|
let wf = WakuFilterClient(
|
||||||
|
peerManager: peerManager,
|
||||||
|
rng: rng,
|
||||||
|
subManager: SubscriptionManager.init()
|
||||||
|
)
|
||||||
|
wf.initProtocolHandler()
|
||||||
|
wf
|
||||||
|
|
||||||
|
|
||||||
|
proc sendFilterRpc(wf: WakuFilterClient, rpc: FilterRPC, peer: PeerId|RemotePeerInfo): Future[WakuFilterResult[void]] {.async, gcsafe.}=
|
||||||
|
let connOpt = await wf.peerManager.dialPeer(peer, WakuFilterCodec)
|
||||||
|
if connOpt.isNone():
|
||||||
|
return err(dialFailure)
|
||||||
|
let connection = connOpt.get()
|
||||||
|
|
||||||
|
await connection.writeLP(rpc.encode().buffer)
|
||||||
|
return ok()
|
||||||
|
|
||||||
|
proc sendFilterRequestRpc(wf: WakuFilterClient,
|
||||||
|
pubsubTopic: string,
|
||||||
|
contentTopics: seq[ContentTopic],
|
||||||
|
subscribe: bool,
|
||||||
|
peer: PeerId|RemotePeerInfo): Future[WakuFilterResult[void]] {.async.} =
|
||||||
|
|
||||||
|
let requestId = generateRequestId(wf.rng)
|
||||||
|
let contentFilters = contentTopics.mapIt(ContentFilter(contentTopic: it))
|
||||||
|
|
||||||
|
let rpc = FilterRpc(
|
||||||
|
requestId: requestId,
|
||||||
|
request: FilterRequest(
|
||||||
|
subscribe: subscribe,
|
||||||
|
pubSubTopic: pubsubTopic,
|
||||||
|
contentFilters: contentFilters
|
||||||
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
proc removeContentFilters*(filters: var Filters, contentFilters: seq[ContentFilter]) {.gcsafe.} =
|
let sendRes = await wf.sendFilterRpc(rpc, peer)
|
||||||
# Flatten all unsubscribe topics into single seq
|
if sendRes.isErr():
|
||||||
let unsubscribeTopics = contentFilters.mapIt(it.contentTopic)
|
waku_filter_errors.inc(labelValues = [sendRes.error])
|
||||||
|
return err(sendRes.error)
|
||||||
debug "unsubscribing", unsubscribeTopics=unsubscribeTopics
|
|
||||||
|
return ok()
|
||||||
|
|
||||||
var rIdToRemove: seq[string] = @[]
|
|
||||||
for rId, f in filters.mpairs:
|
|
||||||
# Iterate filter entries to remove matching content topics
|
|
||||||
|
|
||||||
# make sure we delete the content filter
|
|
||||||
# if no more topics are left
|
|
||||||
f.contentFilters.keepIf(proc (cf: auto): bool = cf.contentTopic notin unsubscribeTopics)
|
|
||||||
|
|
||||||
if f.contentFilters.len == 0:
|
proc subscribe*(wf: WakuFilterClient,
|
||||||
rIdToRemove.add(rId)
|
pubsubTopic: string,
|
||||||
|
contentTopic: ContentTopic|seq[ContentTopic],
|
||||||
|
handler: FilterPushHandler,
|
||||||
|
peer: PeerId|RemotePeerInfo): Future[WakuFilterResult[void]] {.async.} =
|
||||||
|
var topics: seq[ContentTopic]
|
||||||
|
when contentTopic is seq[ContentTopic]:
|
||||||
|
topics = contentTopic
|
||||||
|
else:
|
||||||
|
topics = @[contentTopic]
|
||||||
|
|
||||||
# make sure we delete the filter entry
|
let sendRes = await wf.sendFilterRequestRpc(pubsubTopic, topics, subscribe=true, peer=peer)
|
||||||
# if no more content filters left
|
if sendRes.isErr():
|
||||||
for rId in rIdToRemove:
|
return err(sendRes.error)
|
||||||
filters.del(rId)
|
|
||||||
|
|
||||||
debug "filters modified", filters=filters
|
|
||||||
|
|
||||||
proc notify*(filters: Filters, msg: WakuMessage, requestId: string) =
|
for topic in topics:
|
||||||
for key, filter in filters.pairs:
|
wf.subManager.registerSubscription(pubsubTopic, topic, handler)
|
||||||
# We do this because the key for the filter is set to the requestId received from the filter protocol.
|
|
||||||
# This means we do not need to check the content filter explicitly as all MessagePushs already contain
|
|
||||||
# the requestId of the coresponding filter.
|
|
||||||
if requestId != "" and requestId == key:
|
|
||||||
filter.handler(msg)
|
|
||||||
continue
|
|
||||||
|
|
||||||
# TODO: In case of no topics we should either trigger here for all messages,
|
return ok()
|
||||||
# or we should not allow such filter to exist in the first place.
|
|
||||||
for contentFilter in filter.contentFilters:
|
proc unsubscribe*(wf: WakuFilterClient,
|
||||||
if msg.contentTopic == contentFilter.contentTopic:
|
pubsubTopic: string,
|
||||||
filter.handler(msg)
|
contentTopic: ContentTopic|seq[ContentTopic],
|
||||||
break
|
peer: PeerId|RemotePeerInfo): Future[WakuFilterResult[void]] {.async.} =
|
||||||
|
var topics: seq[ContentTopic]
|
||||||
|
when contentTopic is seq[ContentTopic]:
|
||||||
|
topics = contentTopic
|
||||||
|
else:
|
||||||
|
topics = @[contentTopic]
|
||||||
|
|
||||||
|
let sendRes = await wf.sendFilterRequestRpc(pubsubTopic, topics, subscribe=false, peer=peer)
|
||||||
|
if sendRes.isErr():
|
||||||
|
return err(sendRes.error)
|
||||||
|
|
||||||
|
for topic in topics:
|
||||||
|
wf.subManager.removeSubscription(pubsubTopic, topic)
|
||||||
|
|
||||||
|
return ok()
|
||||||
|
|
||||||
|
proc clearSubscriptions*(wf: WakuFilterClient) =
|
||||||
|
wf.subManager.clear()
|
||||||
|
|
||||||
|
proc getSubscriptionsCount*(wf: WakuFilterClient): int =
|
||||||
|
wf.subManager.getSubscriptionsCount()
|
||||||
@ -10,37 +10,26 @@ import
|
|||||||
import
|
import
|
||||||
../waku_message,
|
../waku_message,
|
||||||
../../node/peer_manager/peer_manager,
|
../../node/peer_manager/peer_manager,
|
||||||
../../utils/requests,
|
|
||||||
./rpc,
|
./rpc,
|
||||||
./rpc_codec
|
./rpc_codec,
|
||||||
|
./protocol_metrics
|
||||||
|
|
||||||
|
|
||||||
declarePublicGauge waku_filter_peers, "number of filter peers"
|
|
||||||
declarePublicGauge waku_filter_subscribers, "number of light node filter subscribers"
|
|
||||||
declarePublicGauge waku_filter_errors, "number of filter protocol errors", ["type"]
|
|
||||||
declarePublicGauge waku_filter_messages, "number of filter messages received", ["type"]
|
|
||||||
|
|
||||||
logScope:
|
logScope:
|
||||||
topics = "wakufilter"
|
topics = "wakufilter"
|
||||||
|
|
||||||
|
|
||||||
const
|
const
|
||||||
# We add a 64kB safety buffer for protocol overhead.
|
|
||||||
# 10x-multiplier also for safety: currently we never
|
|
||||||
# push more than 1 message at a time.
|
|
||||||
MaxRpcSize* = 10 * MaxWakuMessageSize + 64 * 1024
|
|
||||||
|
|
||||||
WakuFilterCodec* = "/vac/waku/filter/2.0.0-beta1"
|
WakuFilterCodec* = "/vac/waku/filter/2.0.0-beta1"
|
||||||
|
|
||||||
WakuFilterTimeout: Duration = 2.hours
|
WakuFilterTimeout: Duration = 2.hours
|
||||||
|
|
||||||
|
|
||||||
# Error types (metric label values)
|
type WakuFilterResult*[T] = Result[T, string]
|
||||||
const
|
|
||||||
dialFailure = "dial_failure"
|
|
||||||
decodeRpcFailure = "decode_rpc_failure"
|
|
||||||
peerNotFoundFailure = "peer_not_found_failure"
|
|
||||||
|
|
||||||
|
|
||||||
|
## Subscription manager
|
||||||
|
|
||||||
type Subscription = object
|
type Subscription = object
|
||||||
requestId: string
|
requestId: string
|
||||||
peer: PeerID
|
peer: PeerID
|
||||||
@ -68,108 +57,86 @@ proc removeSubscription(subscriptions: var seq[Subscription], peer: PeerId, unsu
|
|||||||
subscriptions.keepItIf(it.contentTopics.len > 0)
|
subscriptions.keepItIf(it.contentTopics.len > 0)
|
||||||
|
|
||||||
|
|
||||||
|
## Protocol
|
||||||
|
|
||||||
type
|
type
|
||||||
MessagePushHandler* = proc(requestId: string, msg: MessagePush): Future[void] {.gcsafe, closure.}
|
MessagePushHandler* = proc(requestId: string, msg: MessagePush): Future[void] {.gcsafe, closure.}
|
||||||
|
|
||||||
WakuFilterResult*[T] = Result[T, string]
|
|
||||||
|
|
||||||
WakuFilter* = ref object of LPProtocol
|
WakuFilter* = ref object of LPProtocol
|
||||||
rng*: ref rand.HmacDrbgContext
|
rng*: ref rand.HmacDrbgContext
|
||||||
peerManager*: PeerManager
|
peerManager*: PeerManager
|
||||||
pushHandler*: MessagePushHandler
|
|
||||||
subscriptions*: seq[Subscription]
|
subscriptions*: seq[Subscription]
|
||||||
failedPeers*: Table[string, chronos.Moment]
|
failedPeers*: Table[string, chronos.Moment]
|
||||||
timeout*: chronos.Duration
|
timeout*: chronos.Duration
|
||||||
|
|
||||||
proc init(wf: WakuFilter) =
|
proc handleFilterRequest(wf: WakuFilter, peerId: PeerId, rpc: FilterRPC) =
|
||||||
|
let
|
||||||
|
requestId = rpc.requestId
|
||||||
|
subscribe = rpc.request.subscribe
|
||||||
|
pubsubTopic = rpc.request.pubsubTopic
|
||||||
|
contentTopics = rpc.request.contentFilters.mapIt(it.contentTopic)
|
||||||
|
|
||||||
proc handle(conn: Connection, proto: string) {.async, gcsafe, closure.} =
|
if subscribe:
|
||||||
let message = await conn.readLp(MaxRpcSize.int)
|
info "added filter subscritpiton", peerId=peerId, pubsubTopic=pubsubTopic, contentTopics=contentTopics
|
||||||
|
wf.subscriptions.addSubscription(peerId, requestId, pubsubTopic, contentTopics)
|
||||||
|
else:
|
||||||
|
info "removed filter subscritpiton", peerId=peerId, contentTopics=contentTopics
|
||||||
|
wf.subscriptions.removeSubscription(peerId, contentTopics)
|
||||||
|
|
||||||
let res = FilterRPC.init(message)
|
waku_filter_subscribers.set(wf.subscriptions.len.int64)
|
||||||
if res.isErr():
|
|
||||||
|
|
||||||
|
proc initProtocolHandler(wf: WakuFilter) =
|
||||||
|
proc handler(conn: Connection, proto: string) {.async, gcsafe, closure.} =
|
||||||
|
let buffer = await conn.readLp(MaxRpcSize.int)
|
||||||
|
|
||||||
|
let decodeRpcRes = FilterRPC.init(buffer)
|
||||||
|
if decodeRpcRes.isErr():
|
||||||
waku_filter_errors.inc(labelValues = [decodeRpcFailure])
|
waku_filter_errors.inc(labelValues = [decodeRpcFailure])
|
||||||
return
|
return
|
||||||
|
|
||||||
trace "filter message received"
|
trace "filter message received"
|
||||||
|
|
||||||
let rpc = res.get()
|
let rpc = decodeRpcRes.get()
|
||||||
|
|
||||||
## Filter request
|
## Filter request
|
||||||
# We are receiving a subscription/unsubscription request
|
# Subscription/unsubscription request
|
||||||
if rpc.request != FilterRequest():
|
if rpc.request == FilterRequest():
|
||||||
waku_filter_messages.inc(labelValues = ["FilterRequest"])
|
waku_filter_errors.inc(labelValues = [emptyFilterRequestFailure])
|
||||||
|
# TODO: Manage the empty filter request message error. Perform any action?
|
||||||
|
return
|
||||||
|
|
||||||
let
|
waku_filter_messages.inc(labelValues = ["FilterRequest"])
|
||||||
requestId = rpc.requestId
|
wf.handleFilterRequest(conn.peerId, rpc)
|
||||||
subscribe = rpc.request.subscribe
|
|
||||||
pubsubTopic = rpc.request.pubsubTopic
|
|
||||||
contentTopics = rpc.request.contentFilters.mapIt(it.contentTopic)
|
|
||||||
|
|
||||||
if subscribe:
|
|
||||||
info "added filter subscritpiton", peerId=conn.peerId, pubsubTopic=pubsubTopic, contentTopics=contentTopics
|
|
||||||
wf.subscriptions.addSubscription(conn.peerId, requestId, pubsubTopic, contentTopics)
|
|
||||||
else:
|
|
||||||
info "removed filter subscritpiton", peerId=conn.peerId, contentTopics=contentTopics
|
|
||||||
wf.subscriptions.removeSubscription(conn.peerId, contentTopics)
|
|
||||||
|
|
||||||
waku_filter_subscribers.set(wf.subscriptions.len.int64)
|
|
||||||
|
|
||||||
|
|
||||||
## Push message
|
wf.handler = handler
|
||||||
# We are receiving a messages from the peer that we subscribed to
|
|
||||||
if rpc.push != MessagePush():
|
|
||||||
waku_filter_messages.inc(labelValues = ["MessagePush"])
|
|
||||||
|
|
||||||
let
|
|
||||||
requestId = rpc.requestId
|
|
||||||
push = rpc.push
|
|
||||||
|
|
||||||
info "received filter message push", peerId=conn.peerId
|
|
||||||
await wf.pushHandler(requestId, push)
|
|
||||||
|
|
||||||
|
|
||||||
wf.handler = handle
|
|
||||||
wf.codec = WakuFilterCodec
|
wf.codec = WakuFilterCodec
|
||||||
|
|
||||||
|
proc new*(T: type WakuFilter,
|
||||||
|
peerManager: PeerManager,
|
||||||
|
rng: ref rand.HmacDrbgContext,
|
||||||
|
timeout: Duration = WakuFilterTimeout): T =
|
||||||
|
let wf = WakuFilter(rng: rng,
|
||||||
|
peerManager: peerManager,
|
||||||
|
timeout: timeout)
|
||||||
|
wf.initProtocolHandler()
|
||||||
|
return wf
|
||||||
|
|
||||||
proc init*(T: type WakuFilter,
|
proc init*(T: type WakuFilter,
|
||||||
peerManager: PeerManager,
|
peerManager: PeerManager,
|
||||||
rng: ref rand.HmacDrbgContext,
|
rng: ref rand.HmacDrbgContext,
|
||||||
handler: MessagePushHandler,
|
timeout: Duration = WakuFilterTimeout): T {.
|
||||||
timeout: Duration = WakuFilterTimeout): T =
|
deprecated: "WakuFilter.new()' instead".} =
|
||||||
let wf = WakuFilter(rng: rng,
|
WakuFilter.new(peerManager, rng, timeout)
|
||||||
peerManager: peerManager,
|
|
||||||
pushHandler: handler,
|
|
||||||
timeout: timeout)
|
|
||||||
wf.init()
|
|
||||||
return wf
|
|
||||||
|
|
||||||
|
|
||||||
proc setPeer*(wf: WakuFilter, peer: RemotePeerInfo) =
|
proc sendFilterRpc(wf: WakuFilter, rpc: FilterRPC, peer: PeerId|RemotePeerInfo): Future[WakuFilterResult[void]] {.async, gcsafe.}=
|
||||||
wf.peerManager.addPeer(peer, WakuFilterCodec)
|
|
||||||
waku_filter_peers.inc()
|
|
||||||
|
|
||||||
|
|
||||||
proc sendFilterRpcToPeer(wf: WakuFilter, rpc: FilterRPC, peer: PeerId): Future[WakuFilterResult[void]] {.async, gcsafe.}=
|
|
||||||
let connOpt = await wf.peerManager.dialPeer(peer, WakuFilterCodec)
|
let connOpt = await wf.peerManager.dialPeer(peer, WakuFilterCodec)
|
||||||
if connOpt.isNone():
|
if connOpt.isNone():
|
||||||
return err(dialFailure)
|
return err(dialFailure)
|
||||||
|
|
||||||
let connection = connOpt.get()
|
let connection = connOpt.get()
|
||||||
|
|
||||||
await connection.writeLP(rpc.encode().buffer)
|
await connection.writeLP(rpc.encode().buffer)
|
||||||
|
|
||||||
return ok()
|
|
||||||
|
|
||||||
proc sendFilterRpcToRemotePeer(wf: WakuFilter, rpc: FilterRPC, peer: RemotePeerInfo): Future[WakuFilterResult[void]] {.async, gcsafe.}=
|
|
||||||
let connOpt = await wf.peerManager.dialPeer(peer, WakuFilterCodec)
|
|
||||||
if connOpt.isNone():
|
|
||||||
return err(dialFailure)
|
|
||||||
|
|
||||||
let connection = connOpt.get()
|
|
||||||
|
|
||||||
await connection.writeLP(rpc.encode().buffer)
|
|
||||||
|
|
||||||
return ok()
|
return ok()
|
||||||
|
|
||||||
|
|
||||||
@ -199,6 +166,9 @@ proc handleClientError(wf: WakuFilter, subs: seq[Subscription]) {.raises: [Defec
|
|||||||
|
|
||||||
|
|
||||||
proc handleMessage*(wf: WakuFilter, pubsubTopic: string, msg: WakuMessage) {.async.} =
|
proc handleMessage*(wf: WakuFilter, pubsubTopic: string, msg: WakuMessage) {.async.} =
|
||||||
|
|
||||||
|
trace "handling message", pubsubTopic, contentTopic=msg.contentTopic, subscriptions=wf.subscriptions.len
|
||||||
|
|
||||||
if wf.subscriptions.len <= 0:
|
if wf.subscriptions.len <= 0:
|
||||||
return
|
return
|
||||||
|
|
||||||
@ -218,70 +188,14 @@ proc handleMessage*(wf: WakuFilter, pubsubTopic: string, msg: WakuMessage) {.asy
|
|||||||
push: MessagePush(messages: @[msg])
|
push: MessagePush(messages: @[msg])
|
||||||
)
|
)
|
||||||
|
|
||||||
let res = await wf.sendFilterRpcToPeer(rpc, sub.peer)
|
let res = await wf.sendFilterRpc(rpc, sub.peer)
|
||||||
if res.isErr():
|
if res.isErr():
|
||||||
waku_filter_errors.inc(labelValues = [res.error()])
|
waku_filter_errors.inc(labelValues = [res.error()])
|
||||||
failedSubscriptions.add(sub)
|
failedSubscriptions.add(sub)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
connectedSubscriptions.add(sub)
|
connectedSubscriptions.add(sub)
|
||||||
|
|
||||||
wf.removePeerFromFailedPeersTable(connectedSubscriptions)
|
wf.removePeerFromFailedPeersTable(connectedSubscriptions)
|
||||||
|
|
||||||
wf.handleClientError(failedSubscriptions)
|
wf.handleClientError(failedSubscriptions)
|
||||||
|
|
||||||
|
|
||||||
### Send subscription/unsubscription
|
|
||||||
|
|
||||||
proc subscribe(wf: WakuFilter, pubsubTopic: string, contentTopics: seq[ContentTopic], peer: RemotePeerInfo): Future[WakuFilterResult[string]] {.async, gcsafe.} =
|
|
||||||
let id = generateRequestId(wf.rng)
|
|
||||||
let rpc = FilterRPC(
|
|
||||||
requestId: id,
|
|
||||||
request: FilterRequest(
|
|
||||||
subscribe: true,
|
|
||||||
pubSubTopic: pubsubTopic,
|
|
||||||
contentFilters: contentTopics.mapIt(ContentFilter(contentTopic: it))
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
let res = await wf.sendFilterRpcToRemotePeer(rpc, peer)
|
|
||||||
if res.isErr():
|
|
||||||
waku_filter_errors.inc(labelValues = [res.error()])
|
|
||||||
return err(res.error())
|
|
||||||
|
|
||||||
return ok(id)
|
|
||||||
|
|
||||||
proc subscribe*(wf: WakuFilter, pubsubTopic: string, contentTopics: seq[ContentTopic]): Future[WakuFilterResult[string]] {.async, gcsafe.} =
|
|
||||||
let peerOpt = wf.peerManager.selectPeer(WakuFilterCodec)
|
|
||||||
if peerOpt.isNone():
|
|
||||||
waku_filter_errors.inc(labelValues = [peerNotFoundFailure])
|
|
||||||
return err(peerNotFoundFailure)
|
|
||||||
|
|
||||||
return await wf.subscribe(pubsubTopic, contentTopics, peerOpt.get())
|
|
||||||
|
|
||||||
|
|
||||||
proc unsubscribe(wf: WakuFilter, pubsubTopic: string, contentTopics: seq[ContentTopic], peer: RemotePeerInfo): Future[WakuFilterResult[void]] {.async, gcsafe.} =
|
|
||||||
let id = generateRequestId(wf.rng)
|
|
||||||
let rpc = FilterRPC(
|
|
||||||
requestId: id,
|
|
||||||
request: FilterRequest(
|
|
||||||
subscribe: false,
|
|
||||||
pubSubTopic: pubsubTopic,
|
|
||||||
contentFilters: contentTopics.mapIt(ContentFilter(contentTopic: it))
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
let res = await wf.sendFilterRpcToRemotePeer(rpc, peer)
|
|
||||||
if res.isErr():
|
|
||||||
waku_filter_errors.inc(labelValues = [res.error()])
|
|
||||||
return err(res.error())
|
|
||||||
|
|
||||||
return ok()
|
|
||||||
|
|
||||||
proc unsubscribe*(wf: WakuFilter, pubsubTopic: string, contentTopics: seq[ContentTopic]): Future[WakuFilterResult[void]] {.async, gcsafe.} =
|
|
||||||
let peerOpt = wf.peerManager.selectPeer(WakuFilterCodec)
|
|
||||||
if peerOpt.isNone():
|
|
||||||
waku_filter_errors.inc(labelValues = [peerNotFoundFailure])
|
|
||||||
return err(peerNotFoundFailure)
|
|
||||||
|
|
||||||
return await wf.unsubscribe(pubsubTopic, contentTopics, peerOpt.get())
|
|
||||||
19
waku/v2/protocol/waku_filter/protocol_metrics.nim
Normal file
19
waku/v2/protocol/waku_filter/protocol_metrics.nim
Normal file
@ -0,0 +1,19 @@
|
|||||||
|
{.push raises: [Defect].}
|
||||||
|
|
||||||
|
import metrics
|
||||||
|
|
||||||
|
|
||||||
|
declarePublicGauge waku_filter_subscribers, "number of light node filter subscribers"
|
||||||
|
declarePublicGauge waku_filter_errors, "number of filter protocol errors", ["type"]
|
||||||
|
declarePublicGauge waku_filter_messages, "number of filter messages received", ["type"]
|
||||||
|
declarePublicGauge waku_node_filters, "number of content filter subscriptions"
|
||||||
|
|
||||||
|
|
||||||
|
# Error types (metric label values)
|
||||||
|
const
|
||||||
|
dialFailure* = "dial_failure"
|
||||||
|
decodeRpcFailure* = "decode_rpc_failure"
|
||||||
|
peerNotFoundFailure* = "peer_not_found_failure"
|
||||||
|
emptyMessagePushFailure* = "empty_message_push_failure"
|
||||||
|
emptyFilterRequestFailure* = "empty_filter_request_failure"
|
||||||
|
|
||||||
@ -9,6 +9,11 @@ import
|
|||||||
./rpc
|
./rpc
|
||||||
|
|
||||||
|
|
||||||
|
# Multiply by 10 for safety. Currently we never push more than 1 message at a time
|
||||||
|
# We add a 64kB safety buffer for protocol overhead.
|
||||||
|
const MaxRpcSize* = 10 * MaxWakuMessageSize + 64 * 1024
|
||||||
|
|
||||||
|
|
||||||
proc encode*(filter: ContentFilter): ProtoBuffer =
|
proc encode*(filter: ContentFilter): ProtoBuffer =
|
||||||
var output = initProtoBuffer()
|
var output = initProtoBuffer()
|
||||||
output.write3(1, filter.contentTopic)
|
output.write3(1, filter.contentTopic)
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user