mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-04-18 17:53:09 +00:00
102 lines
20 KiB
Plaintext
102 lines
20 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_relay_receiving_node_not_connected_directly_to_relaying_node with id: 2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.filter:filter.py:28 Running fixture setup: filter_setup
|
||
[35mDEBUG [0m src.steps.light_push:light_push.py:28 Running fixture setup: light_push_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node3_2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:33 Network waku created
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.237.89
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['8530', '8531', '8532', '8533', '8534']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 8530:8530 -p 8531:8531 -p 8532:8532 -p 8533:8533 -p 8534:8534 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=8532 --rest-port=8530 --tcp-port=8531 --discv5-udp-port=8533 --rest-address=0.0.0.0 --nat=extip:172.18.237.89 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=d2ab98ff5efb620bc2bec51a6bad3c5b80ff9f1df3ca94d5dc191ee8aaf8b1dd --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=8534 --metrics-logging=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.237.89 waku 4768a7b2a6227c33427e4d9a4bec4e2fee687381dea70d3c3ecdeef6427a144b
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 4768a7b2a622. Setting up logs at ./log/docker/node1_2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 8530
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:8530/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:8530/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.237.89/tcp/8531/p2p/16Uiu2HAmJdaf31mVSvKZKGTtnECKWqkcFsNtziVFJoz8oX5uTjag","/ip4/172.18.237.89/tcp/8532/ws/p2p/16Uiu2HAmJdaf31mVSvKZKGTtnECKWqkcFsNtziVFJoz8oX5uTjag"],"enrUri":"enr:-L24QCjCVU5GMQmrrccJp9728UKCbw5NAllC7cuLscF2-KiVK3cvvdTS3oOP3kQ5joUkNXjKgZpDotVRpEObyFYCGtMCgmlkgnY0gmlwhKwS7VmKbXVsdGlhZGRyc5YACASsEu1ZBiFTAAoErBLtWQYhVN0DgnJzhQADAQAAiXNlY3AyNTZrMaEDWMm_866EOWOpVUbJRoUMuGzQ5FVUSdIFYIcCyC5xhNmDdGNwgiFTg3VkcIIhVYV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.96.186
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['23317', '23318', '23319', '23320', '23321']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 23317:23317 -p 23318:23318 -p 23319:23319 -p 23320:23320 -p 23321:23321 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=23319 --rest-port=23317 --tcp-port=23318 --discv5-udp-port=23320 --rest-address=0.0.0.0 --nat=extip:172.18.96.186 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=cc1ebe6ce25bc3f1fa69f36831d5f91deb4caed837aeeff5cd576da96ac83acb --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=23321 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QCjCVU5GMQmrrccJp9728UKCbw5NAllC7cuLscF2-KiVK3cvvdTS3oOP3kQ5joUkNXjKgZpDotVRpEObyFYCGtMCgmlkgnY0gmlwhKwS7VmKbXVsdGlhZGRyc5YACASsEu1ZBiFTAAoErBLtWQYhVN0DgnJzhQADAQAAiXNlY3AyNTZrMaEDWMm_866EOWOpVUbJRoUMuGzQ5FVUSdIFYIcCyC5xhNmDdGNwgiFTg3VkcIIhVYV3YWt1MgE
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.96.186 waku f25c88d502fda6af435852637a24a6d013ceafbf1bfb432ba0b2543f941f0f3d
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID f25c88d502fd. Setting up logs at ./log/docker/node2_2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 23317
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23317/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23317/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.96.186/tcp/23318/p2p/16Uiu2HAmAw2yFLNvbLMjPBmxMTXfEY4hwnVAC9L3jBgbyYUK1tYB","/ip4/172.18.96.186/tcp/23319/ws/p2p/16Uiu2HAmAw2yFLNvbLMjPBmxMTXfEY4hwnVAC9L3jBgbyYUK1tYB"],"enrUri":"enr:-L24QO-9jTKwd6O8uDPeAEFA7k5NyLFXQGVZJF5cAsWZbS4ARgGDLdjDnPANQq9SGDkMtgweFpKHepn3AcmgSVF0RjkCgmlkgnY0gmlwhKwSYLqKbXVsdGlhZGRyc5YACASsEmC6BlsWAAoErBJgugZbF90DgnJzhQADAQAAiXNlY3AyNTZrMaEC5mThqeLupgV5rEg6tw2oFTPzUUR_VSwvmIW9iUhl1fyDdGNwglsWg3VkcIJbGIV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.45.95
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['11690', '11691', '11692', '11693', '11694']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 11690:11690 -p 11691:11691 -p 11692:11692 -p 11693:11693 -p 11694:11694 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=11692 --rest-port=11690 --tcp-port=11691 --discv5-udp-port=11693 --rest-address=0.0.0.0 --nat=extip:172.18.45.95 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=db293aeb2f0ee6a0ff18d56ac1c5eb829ffac099c8ca8bd2a0a6baf8047b990d --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=11694 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QO-9jTKwd6O8uDPeAEFA7k5NyLFXQGVZJF5cAsWZbS4ARgGDLdjDnPANQq9SGDkMtgweFpKHepn3AcmgSVF0RjkCgmlkgnY0gmlwhKwSYLqKbXVsdGlhZGRyc5YACASsEmC6BlsWAAoErBJgugZbF90DgnJzhQADAQAAiXNlY3AyNTZrMaEC5mThqeLupgV5rEg6tw2oFTPzUUR_VSwvmIW9iUhl1fyDdGNwglsWg3VkcIJbGIV3YWt1MgE
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.45.95 waku a19dfe801f35727a8673cc40cc07fc70b221950e0a0a41ce38a1469c59434225
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID a19dfe801f35. Setting up logs at ./log/docker/node3_2026-03-11_04-32-35__bf2370af-230c-4480-a922-94342b2d93a3__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 11690
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:11690/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:11690/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.45.95/tcp/11691/p2p/16Uiu2HAmEkkWfHi8B1MaTFTfbDXTGwESJdWUHUSKukryhFcVysAu","/ip4/172.18.45.95/tcp/11692/ws/p2p/16Uiu2HAmEkkWfHi8B1MaTFTfbDXTGwESJdWUHUSKukryhFcVysAu"],"enrUri":"enr:-L24QBAvLjaR2dLcvVSdJQhHCfZf8oTzqcCF9xYDjts4eU6DJyrP-mj_qFhCnScmYwEuj6MAK70dWwBlrL8EJi3C3ygCgmlkgnY0gmlwhKwSLV-KbXVsdGlhZGRyc5YACASsEi1fBi2rAAoErBItXwYtrN0DgnJzhQADAQAAiXNlY3AyNTZrMaEDHzE2J8289punXqydtJTzmmvvHeR-W3iqHjurOKhqCCaDdGNwgi2rg3VkcIItrYV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8530/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/1"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:23317/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/1"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:11690/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/1"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:8530/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.96.186/tcp/48390/p2p/16Uiu2HAmAw2yFLNvbLMjPBmxMTXfEY4hwnVAC9L3jBgbyYUK1tYB","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"logos-delivery-v0.36.0-164-g0ad551","origin":"UnknownOrigin"}]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23317/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.45.95/tcp/44762/p2p/16Uiu2HAmEkkWfHi8B1MaTFTfbDXTGwESJdWUHUSKukryhFcVysAu","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"logos-delivery-v0.36.0-164-g0ad551","origin":"UnknownOrigin"},{"multiaddr":"/ip4/172.18.237.89/tcp/8531/p2p/16Uiu2HAmJdaf31mVSvKZKGTtnECKWqkcFsNtziVFJoz8oX5uTjag","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"logos-delivery-v0.36.0-164-g0ad551","origin":"Discv5"}]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:11690/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.96.186/tcp/23318/p2p/16Uiu2HAmAw2yFLNvbLMjPBmxMTXfEY4hwnVAC9L3jBgbyYUK1tYB","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"logos-delivery-v0.36.0-164-g0ad551","origin":"Discv5"}]'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 30 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8530/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F1" -H "Content-Type: application/json" -d '{"payload": "RmlsdGVyIHdvcmtzISE=", "contentTopic": "/test/1/waku-filter/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[35mDEBUG [0m src.steps.relay:relay.py:123 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:11690/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F1" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"RmlsdGVyIHdvcmtzISE=","contentTopic":"/test/1/waku-filter/proto","version":0,"timestamp":1773203592265130060,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 4768a7b2a622
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id f25c88d502fd
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id a19dfe801f35
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 4768a7b2a622. Exiting log stream.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |