mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-04 01:23:07 +00:00
102 lines
20 KiB
Plaintext
102 lines
20 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_relay_receiving_node_not_connected_directly_to_relaying_node with id: 2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.filter:filter.py:28 Running fixture setup: filter_setup
|
||
[35mDEBUG [0m src.steps.light_push:light_push.py:28 Running fixture setup: light_push_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node3_2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:33 Network waku created
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.206.24
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['19743', '19744', '19745', '19746', '19747']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 19743:19743 -p 19744:19744 -p 19745:19745 -p 19746:19746 -p 19747:19747 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=19745 --rest-port=19743 --tcp-port=19744 --discv5-udp-port=19746 --rest-address=0.0.0.0 --nat=extip:172.18.206.24 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=9e7deed83fad4adefe1eaedf89afcfcbe6b4ecd6cba18b0b1da625a8dcbbf7c9 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=19747 --metrics-logging=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.206.24 waku 85d26f108fb857d8e74f778778dceee111088f356aee2f51e2f1ce0a0f032d36
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 85d26f108fb8. Setting up logs at ./log/docker/node1_2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 19743
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:19743/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:19743/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.206.24/tcp/19744/p2p/16Uiu2HAmVYUHUBphsZyfL8KYszmxLQbWwR7ZQPSUeyaEhXKwxjX8","/ip4/172.18.206.24/tcp/19745/ws/p2p/16Uiu2HAmVYUHUBphsZyfL8KYszmxLQbWwR7ZQPSUeyaEhXKwxjX8"],"enrUri":"enr:-L24QL4dkrnnXQiXSBEw7TTJ2634rJQcCYkU8qE41DJRbQP_dD4YVBWmpqVQtzaLlsU1qR2RtnIFcyH-W0WkGYHe1AACgmlkgnY0gmlwhKwSzhiKbXVsdGlhZGRyc5YACASsEs4YBk0gAAoErBLOGAZNId0DgnJzhQADAQAAiXNlY3AyNTZrMaED-usygtCjYhOZNB1gk9PetwooRkJUqr2ce9T4xQfUMnODdGNwgk0gg3VkcIJNIoV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.128.108
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['27016', '27017', '27018', '27019', '27020']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 27016:27016 -p 27017:27017 -p 27018:27018 -p 27019:27019 -p 27020:27020 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=27018 --rest-port=27016 --tcp-port=27017 --discv5-udp-port=27019 --rest-address=0.0.0.0 --nat=extip:172.18.128.108 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=12acf39ba17dbd052a17dfaaf57e3e5aaae81e4d63cb537dedab3cda9c4b920a --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=27020 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QL4dkrnnXQiXSBEw7TTJ2634rJQcCYkU8qE41DJRbQP_dD4YVBWmpqVQtzaLlsU1qR2RtnIFcyH-W0WkGYHe1AACgmlkgnY0gmlwhKwSzhiKbXVsdGlhZGRyc5YACASsEs4YBk0gAAoErBLOGAZNId0DgnJzhQADAQAAiXNlY3AyNTZrMaED-usygtCjYhOZNB1gk9PetwooRkJUqr2ce9T4xQfUMnODdGNwgk0gg3VkcIJNIoV3YWt1MgE
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.128.108 waku bc093e72b248f81eedcb6db60f7b83f2e65b475e4f5d936ba14f001184ea085c
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID bc093e72b248. Setting up logs at ./log/docker/node2_2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 27016
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:27016/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:27016/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.128.108/tcp/27017/p2p/16Uiu2HAkuTBnKjp9zPVLGDzNc8bRBRX6H8C8dTNWf5jNHgozMxLy","/ip4/172.18.128.108/tcp/27018/ws/p2p/16Uiu2HAkuTBnKjp9zPVLGDzNc8bRBRX6H8C8dTNWf5jNHgozMxLy"],"enrUri":"enr:-L24QBupcLoFR9tm2KsFf_n2ry927bbrwBBmJ-q3UWw27uHXVwoKG0ryOsTVlw0_SwodzWcEQnpejPHErITNAJhkZwMCgmlkgnY0gmlwhKwSgGyKbXVsdGlhZGRyc5YACASsEoBsBmmJAAoErBKAbAZpit0DgnJzhQADAQAAiXNlY3AyNTZrMaECAGNHP9pK-5Gqyu3UJHwfjgkP23ricg2q_5RUdz9NMnKDdGNwgmmJg3VkcIJpi4V3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.62.51
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['12069', '12070', '12071', '12072', '12073']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 12069:12069 -p 12070:12070 -p 12071:12071 -p 12072:12072 -p 12073:12073 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=12071 --rest-port=12069 --tcp-port=12070 --discv5-udp-port=12072 --rest-address=0.0.0.0 --nat=extip:172.18.62.51 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=b2fff36961ea11ad98b8bcf1716aee9bb84e77cdd0a83bc1022b4beceeb8331e --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=12073 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QBupcLoFR9tm2KsFf_n2ry927bbrwBBmJ-q3UWw27uHXVwoKG0ryOsTVlw0_SwodzWcEQnpejPHErITNAJhkZwMCgmlkgnY0gmlwhKwSgGyKbXVsdGlhZGRyc5YACASsEoBsBmmJAAoErBKAbAZpit0DgnJzhQADAQAAiXNlY3AyNTZrMaECAGNHP9pK-5Gqyu3UJHwfjgkP23ricg2q_5RUdz9NMnKDdGNwgmmJg3VkcIJpi4V3YWt1MgE
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.62.51 waku c951a9ee207b7983f3f84b375bdfe20219628783cc06d258502965203c123f35
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID c951a9ee207b. Setting up logs at ./log/docker/node3_2026-02-26_04-34-36__fb585df4-c801-4284-8bfb-a6402d8dc20f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 12069
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:12069/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:12069/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.62.51/tcp/12070/p2p/16Uiu2HAmBxmpricrouZq5wQmS8Q3w4GLAj59Go4MF4w1CSb7eNeF","/ip4/172.18.62.51/tcp/12071/ws/p2p/16Uiu2HAmBxmpricrouZq5wQmS8Q3w4GLAj59Go4MF4w1CSb7eNeF"],"enrUri":"enr:-L24QAy86Kz75NVaSlCC0haOU9gCm7zFY87GqyZRMLnksIZePP9h_Ix-Nxe9QBWALZfTa4BCFzMAeXOF0dsIsB3bl1cCgmlkgnY0gmlwhKwSPjOKbXVsdGlhZGRyc5YACASsEj4zBi8mAAoErBI-MwYvJ90DgnJzhQADAQAAiXNlY3AyNTZrMaEC9bKbHhiTahfR8K070o6ja9Li0CBdbCkQQAnMWtJBeYyDdGNwgi8mg3VkcIIvKIV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:19743/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/1"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:27016/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/1"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:12069/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/1"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:19743/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.128.108/tcp/34140/p2p/16Uiu2HAkuTBnKjp9zPVLGDzNc8bRBRX6H8C8dTNWf5jNHgozMxLy","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.36.0-156-gc7e0cc","origin":"UnknownOrigin"}]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:27016/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.62.51/tcp/46834/p2p/16Uiu2HAmBxmpricrouZq5wQmS8Q3w4GLAj59Go4MF4w1CSb7eNeF","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.36.0-156-gc7e0cc","origin":"UnknownOrigin"},{"multiaddr":"/ip4/172.18.206.24/tcp/19744/p2p/16Uiu2HAmVYUHUBphsZyfL8KYszmxLQbWwR7ZQPSUeyaEhXKwxjX8","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.36.0-156-gc7e0cc","origin":"Discv5"}]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:12069/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.128.108/tcp/27017/p2p/16Uiu2HAkuTBnKjp9zPVLGDzNc8bRBRX6H8C8dTNWf5jNHgozMxLy","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.36.0-156-gc7e0cc","origin":"Discv5"}]'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 30 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:19743/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F1" -H "Content-Type: application/json" -d '{"payload": "RmlsdGVyIHdvcmtzISE=", "contentTopic": "/test/1/waku-filter/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[35mDEBUG [0m src.steps.relay:relay.py:123 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:12069/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F1" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"RmlsdGVyIHdvcmtzISE=","contentTopic":"/test/1/waku-filter/proto","version":0,"timestamp":1772080513433827117,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 85d26f108fb8
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id bc093e72b248
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id c951a9ee207b
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 85d26f108fb8. Exiting log stream.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |