mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-05 10:03:16 +00:00
84 lines
15 KiB
Plaintext
84 lines
15 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_unsubscribe_from_some_content_topics[content_topic_list1] with id: 2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.47.81
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['52133', '52134', '52135', '52136', '52137']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 52133:52133 -p 52134:52134 -p 52135:52135 -p 52136:52136 -p 52137:52137 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=52135 --rest-port=52133 --tcp-port=52134 --discv5-udp-port=52136 --rest-address=0.0.0.0 --nat=extip:172.18.47.81 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=b4eceef363f42acc4ab3daa5fae8ab3c5b0de1ca11641bc3fa8a32ba13739fb9 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=52137 --metrics-logging=true --relay=true --filter=true --content-topic=/myapp/1/latest/proto --num-shards-in-network=8
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.47.81 waku 21274de86d679291e7c57c2fffd50b470402f336f4cdaaef13e020ae8076c285
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 21274de86d67. Setting up logs at ./log/docker/node1_2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 52133
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container a12d19dd1699. Exiting log stream.
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 8095ddde0f72. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52133/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52133/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52133/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.47.81/tcp/52134/p2p/16Uiu2HAm4mZoaqmVEzqjD9m9ghFFrFyCfVFN5y8UHhHaamg9zYeS","/ip4/172.18.47.81/tcp/52135/ws/p2p/16Uiu2HAm4mZoaqmVEzqjD9m9ghFFrFyCfVFN5y8UHhHaamg9zYeS"],"enrUri":"enr:-L24QDI5CTRnxUtljgGPoX19NRgGUE5721Iud7BBRGQrzxGlAYbkW7bpqvK3B6or0gZ1s4OW7Nvmae83KNBbYWgq0tECgmlkgnY0gmlwhKwSL1GKbXVsdGlhZGRyc5YACASsEi9RBsumAAoErBIvUQbLp90DgnJzhQACAQAAiXNlY3AyNTZrMaECitHQtXk9k1QkD4A96ehB-dq-iuwarRB58FOdrC48BR-DdGNwgsumg3VkcILLqIV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.235.41
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['58819', '58820', '58821', '58822', '58823']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 58819:58819 -p 58820:58820 -p 58821:58821 -p 58822:58822 -p 58823:58823 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=58821 --rest-port=58819 --tcp-port=58820 --discv5-udp-port=58822 --rest-address=0.0.0.0 --nat=extip:172.18.235.41 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=1bdddcde2fbe2c46ce0dc4d7f0aeefcf3eaa24c4c8ee4f2392f7cdebe69f19e0 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=58823 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QDI5CTRnxUtljgGPoX19NRgGUE5721Iud7BBRGQrzxGlAYbkW7bpqvK3B6or0gZ1s4OW7Nvmae83KNBbYWgq0tECgmlkgnY0gmlwhKwSL1GKbXVsdGlhZGRyc5YACASsEi9RBsumAAoErBIvUQbLp90DgnJzhQACAQAAiXNlY3AyNTZrMaECitHQtXk9k1QkD4A96ehB-dq-iuwarRB58FOdrC48BR-DdGNwgsumg3VkcILLqIV3YWt1MgU --content-topic=/myapp/1/latest/proto --num-shards-in-network=8
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.235.41 waku 8cc96586b2f395e8d8613fbadb1a9e40528cf3b6c3ce27193c514c99ebcfcc30
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 8cc96586b2f3. Setting up logs at ./log/docker/node2_2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 58819
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:58819/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:58819/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:58819/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:58819/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:58819/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:58819/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.235.41/tcp/58820/p2p/16Uiu2HAmMgsUy6Jx8ZwSSBCgdwAxg5siDxfMratPxq5GWXyjwAqk","/ip4/172.18.235.41/tcp/58821/ws/p2p/16Uiu2HAmMgsUy6Jx8ZwSSBCgdwAxg5siDxfMratPxq5GWXyjwAqk"],"enrUri":"enr:-L24QFSmjHQpkiLs8SVYmS4zbA5hAeQQ8L4fSmVkLV3-g0h4H5fEBhhmOaiUP1yltSIrqGCJA3L4f0k8HMJGdPPgBCACgmlkgnY0gmlwhKwS6ymKbXVsdGlhZGRyc5YACASsEuspBuXEAAoErBLrKQblxd0DgnJzhQACAQAAiXNlY3AyNTZrMaEDhjSUJoF84zfsvQ7hI39xxHzO-AmVKHwqSWIYE7JN4F-DdGNwguXEg3VkcILlxoV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:58819/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.47.81/tcp/52134/p2p/16Uiu2HAm4mZoaqmVEzqjD9m9ghFFrFyCfVFN5y8UHhHaamg9zYeS"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:52133/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/myapp/1/latest/proto", "/waku/2/content/test.js", "/app/22/sometopic/someencoding", "/toychat/2/huilong/proto", "/statusim/1/community/cbor", "/app/27/sometopic/someencoding", "/app/29/sometopic/someencoding", "/app/20/sometopic/someencoding"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:58819/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/myapp/1/latest/proto", "/waku/2/content/test.js", "/app/22/sometopic/someencoding", "/toychat/2/huilong/proto", "/statusim/1/community/cbor", "/app/27/sometopic/someencoding", "/app/29/sometopic/someencoding", "/app/20/sometopic/someencoding"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:52133/relay/v1/auto/messages" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52133/relay/v1/auto/messages/%2Fmyapp%2F1%2Flatest%2Fproto" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1774500495000804122,"ephemeral":false,"proof":""},{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1774500495465882818,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 21274de86d67
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 8cc96586b2f3
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-26_04-48-12__71b11a3a-f658-474a-8919-4997fae04d11__wakuorg_nwaku:latest.log |