mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-05 18:09:25 +00:00
278 lines
49 KiB
Plaintext
278 lines
49 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_2_nodes_different_content_topic_same_shard with id: 2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.150.31
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['35986', '35987', '35988', '35989', '35990']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 35986:35986 -p 35987:35987 -p 35988:35988 -p 35989:35989 -p 35990:35990 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=35988 --rest-port=35986 --tcp-port=35987 --discv5-udp-port=35989 --rest-address=0.0.0.0 --nat=extip:172.18.150.31 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=c6eb6badaeda04baa76fc57a1f8c1d50beea0891abd9f72dfcd95f0eaebebffe --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=35990 --metrics-logging=true --relay=true --filter=true --content-topic=/newsService/1.0/weekly/protobuf
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container a1498f5c7791. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.150.31 waku 8fa3fe4666147ec3d071e1c107aab3872665cfe76e3cf00d8e5314eaad6f656c
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 8fa3fe466614. Setting up logs at ./log/docker/node1_2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 35986
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container a9a1729bf9a6. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:35986/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:35986/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:35986/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.150.31/tcp/35987/p2p/16Uiu2HAmTeNmf3ZcmSZdTj6L13x91uX2nib2NMJrnqURGwEd6rcV","/ip4/172.18.150.31/tcp/35988/ws/p2p/16Uiu2HAmTeNmf3ZcmSZdTj6L13x91uX2nib2NMJrnqURGwEd6rcV"],"enrUri":"enr:-L24QPprPDnk8L4YweVucF41VGXW1sZzeWRAOAkYWdo2ZNO8PDVfGr45c7YlPPLUBvLVQhveWb4RtMzSukmLEfjG5W8CgmlkgnY0gmlwhKwSlh-KbXVsdGlhZGRyc5YACASsEpYfBoyTAAoErBKWHwaMlN0DgnJzhQACAQAAiXNlY3AyNTZrMaED3rcS1x_7opkADH9Dw6Ftp-M3CPmzvaMP-6_SI8lCexaDdGNwgoyTg3VkcIKMlYV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.80.229
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['17251', '17252', '17253', '17254', '17255']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 17251:17251 -p 17252:17252 -p 17253:17253 -p 17254:17254 -p 17255:17255 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=17253 --rest-port=17251 --tcp-port=17252 --discv5-udp-port=17254 --rest-address=0.0.0.0 --nat=extip:172.18.80.229 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=7c55d9ccaee9cde94a52a3be31c9965af8a0dd87fdfa4b7ac1ea9e7b107de15b --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=17255 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QPprPDnk8L4YweVucF41VGXW1sZzeWRAOAkYWdo2ZNO8PDVfGr45c7YlPPLUBvLVQhveWb4RtMzSukmLEfjG5W8CgmlkgnY0gmlwhKwSlh-KbXVsdGlhZGRyc5YACASsEpYfBoyTAAoErBKWHwaMlN0DgnJzhQACAQAAiXNlY3AyNTZrMaED3rcS1x_7opkADH9Dw6Ftp-M3CPmzvaMP-6_SI8lCexaDdGNwgoyTg3VkcIKMlYV3YWt1MgU --content-topic=/newsService/1.0/alerts/xml
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.80.229 waku d636ff0b52b0a6fe05ddb0cf22efa9d97405200113cc7bb1cbf8873b2d2d30bf
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID d636ff0b52b0. Setting up logs at ./log/docker/node2_2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 17251
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.80.229/tcp/17252/p2p/16Uiu2HAmLxHvUZqWHvGebE8W49HmXBW5GP5uwWTujveWGCK3cm8L","/ip4/172.18.80.229/tcp/17253/ws/p2p/16Uiu2HAmLxHvUZqWHvGebE8W49HmXBW5GP5uwWTujveWGCK3cm8L"],"enrUri":"enr:-L24QHevUQeKUlzqGhbDsxGZrw6OFbSXEDJttmigI6XwJr2dQopl2CvjQywe9O9ff24LsDu0NQ5fF2NBm54KG0r3eecCgmlkgnY0gmlwhKwSUOWKbXVsdGlhZGRyc5YACASsElDlBkNkAAoErBJQ5QZDZd0DgnJzhQACAQAAiXNlY3AyNTZrMaEDe0w881ahsbJz4F3uSD1me2XG2UKIEP_insCRecBMW_GDdGNwgkNkg3VkcIJDZoV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:17251/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.150.31/tcp/35987/p2p/16Uiu2HAmTeNmf3ZcmSZdTj6L13x91uX2nib2NMJrnqURGwEd6rcV"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:35986/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/newsService/1.0/weekly/protobuf"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:17251/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/newsService/1.0/alerts/xml"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:35986/relay/v1/auto/messages" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/newsService/1.0/weekly/protobuf", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:35986/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/newsService/1.0/weekly/protobuf","version":0,"timestamp":1772685325371726422,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_2:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:17251/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf" -H "Content-Type: application/json" -d 'None'
|
||
[31m[1mERROR [0m src.node.api_clients.base_client:base_client.py:16 HTTP error occurred: 404 Client Error: Not Found for url: http://127.0.0.1:17251/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf. Response content: b'/newsService/1.0/weekly/protobuf'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 8fa3fe466614
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id d636ff0b52b0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-05_04-35-22__4032a241-b8ee-44b1-ae55-f72c04e22f33__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_2_nodes_different_content_topic_same_shard with id: 2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.68.229
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['43912', '43913', '43914', '43915', '43916']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 43912:43912 -p 43913:43913 -p 43914:43914 -p 43915:43915 -p 43916:43916 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=43914 --rest-port=43912 --tcp-port=43913 --discv5-udp-port=43915 --rest-address=0.0.0.0 --nat=extip:172.18.68.229 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=da2dbadf05c9cfbbed05fecb705eeb4cc20eb7c1b5ed333da64010448a667743 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=43916 --metrics-logging=true --relay=true --filter=true --content-topic=/newsService/1.0/weekly/protobuf
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 8fa3fe466614. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.68.229 waku 6958f5151d08397d35160243b7133095dfa5a6b2264e86ae550c0b0fa9942a63
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 6958f5151d08. Setting up logs at ./log/docker/node1_2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 43912
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container d636ff0b52b0. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:43912/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:43912/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:43912/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:43912/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.68.229/tcp/43913/p2p/16Uiu2HAm1b5no4xWinqqH9Zm5rkN2qN7BDjouYWY9oyp2xd2Fu49","/ip4/172.18.68.229/tcp/43914/ws/p2p/16Uiu2HAm1b5no4xWinqqH9Zm5rkN2qN7BDjouYWY9oyp2xd2Fu49"],"enrUri":"enr:-L24QEcwkBTPKZz96OOck8WSP4g4VqBGB9FKk_JPKIeOmt9MbfiewXXIYZwa7bbWUK6pV4AXh_A7lILLsPOJmqmAAkYCgmlkgnY0gmlwhKwSROWKbXVsdGlhZGRyc5YACASsEkTlBquJAAoErBJE5Qarit0DgnJzhQACAQAAiXNlY3AyNTZrMaECW49FdMncK-v56nY_Qdrb3U5BNiIEdJDvYr_wpC3qOwaDdGNwgquJg3VkcIKri4V3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.192.194
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['31028', '31029', '31030', '31031', '31032']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 31028:31028 -p 31029:31029 -p 31030:31030 -p 31031:31031 -p 31032:31032 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=31030 --rest-port=31028 --tcp-port=31029 --discv5-udp-port=31031 --rest-address=0.0.0.0 --nat=extip:172.18.192.194 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=e4efca0e7ffcedd5dcbcaff8caca8af58007ebfebc5fc5ecd06c6b1862eee7c8 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=31032 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QEcwkBTPKZz96OOck8WSP4g4VqBGB9FKk_JPKIeOmt9MbfiewXXIYZwa7bbWUK6pV4AXh_A7lILLsPOJmqmAAkYCgmlkgnY0gmlwhKwSROWKbXVsdGlhZGRyc5YACASsEkTlBquJAAoErBJE5Qarit0DgnJzhQACAQAAiXNlY3AyNTZrMaECW49FdMncK-v56nY_Qdrb3U5BNiIEdJDvYr_wpC3qOwaDdGNwgquJg3VkcIKri4V3YWt1MgU --content-topic=/newsService/1.0/alerts/xml
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.192.194 waku 7eedbde664fe799530828f494754087a95d6c6637af79016f7dfa83d6ef5a1d7
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 7eedbde664fe. Setting up logs at ./log/docker/node2_2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 31028
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.192.194/tcp/31029/p2p/16Uiu2HAmUWMRTpz4pzukkfjgcZUFzsvneCjje4RCHEai37vYUtKG","/ip4/172.18.192.194/tcp/31030/ws/p2p/16Uiu2HAmUWMRTpz4pzukkfjgcZUFzsvneCjje4RCHEai37vYUtKG"],"enrUri":"enr:-L24QMf0qGU8DpfkxhtTD_1qr4gE7hL4Nimkos6K9viNKjGdeal44N3BKZC3RUJ8xluK-eDtzT6znRUot4ilCIp5MVMCgmlkgnY0gmlwhKwSwMKKbXVsdGlhZGRyc5YACASsEsDCBnk1AAoErBLAwgZ5Nt0DgnJzhQACAQAAiXNlY3AyNTZrMaED64SW9RW3uIvwAYPn_CejQt-RnSwnuiNwmOVWGNAvlReDdGNwgnk1g3VkcIJ5N4V3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:31028/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.68.229/tcp/43913/p2p/16Uiu2HAm1b5no4xWinqqH9Zm5rkN2qN7BDjouYWY9oyp2xd2Fu49"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:43912/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/newsService/1.0/weekly/protobuf"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:31028/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/newsService/1.0/alerts/xml"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:43912/relay/v1/auto/messages" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/newsService/1.0/weekly/protobuf", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:43912/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/newsService/1.0/weekly/protobuf","version":0,"timestamp":1772685330228846530,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_2:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31028/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf" -H "Content-Type: application/json" -d 'None'
|
||
[31m[1mERROR [0m src.node.api_clients.base_client:base_client.py:16 HTTP error occurred: 404 Client Error: Not Found for url: http://127.0.0.1:31028/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf. Response content: b'/newsService/1.0/weekly/protobuf'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 6958f5151d08
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 7eedbde664fe
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-05_04-35-26__3e7e9343-d632-4b41-adb0-b22d82caaf77__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_2_nodes_different_content_topic_same_shard with id: 2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.18.16
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['56359', '56360', '56361', '56362', '56363']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 56359:56359 -p 56360:56360 -p 56361:56361 -p 56362:56362 -p 56363:56363 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=56361 --rest-port=56359 --tcp-port=56360 --discv5-udp-port=56362 --rest-address=0.0.0.0 --nat=extip:172.18.18.16 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=c1bee17fa41c89b8b0ceb04fa26e7b2666da68f1ede95a7dacf047a76dfab6f5 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=56363 --metrics-logging=true --relay=true --filter=true --content-topic=/newsService/1.0/weekly/protobuf
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 6958f5151d08. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.18.16 waku 17198e4e0f30bfb99be6decb7de70beef108ba4131fdc6df8d0182e16b1315fe
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 17198e4e0f30. Setting up logs at ./log/docker/node1_2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 56359
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 7eedbde664fe. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:56359/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:56359/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:56359/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.18.16/tcp/56360/p2p/16Uiu2HAm7K5fph8W1S3FE71fgPN9KUWVpTZMmJzYiELYBfrzTiGQ","/ip4/172.18.18.16/tcp/56361/ws/p2p/16Uiu2HAm7K5fph8W1S3FE71fgPN9KUWVpTZMmJzYiELYBfrzTiGQ"],"enrUri":"enr:-L24QOi-e06edCt2WaES5fQOiFc1UodNiyC3FOJBxhmfLSZ_Jk1HA4WAR4-4nDV3Vls0OLKYhT897ZHk2SNnX0IE44ACgmlkgnY0gmlwhKwSEhCKbXVsdGlhZGRyc5YACASsEhIQBtwoAAoErBISEAbcKd0DgnJzhQACAQAAiXNlY3AyNTZrMaECsJv2yTXk7KvHz7agv_m0y7kpNjArXLsMdw2Jjef2wgGDdGNwgtwog3VkcILcKoV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.126.111
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['9727', '9728', '9729', '9730', '9731']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 9727:9727 -p 9728:9728 -p 9729:9729 -p 9730:9730 -p 9731:9731 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=9729 --rest-port=9727 --tcp-port=9728 --discv5-udp-port=9730 --rest-address=0.0.0.0 --nat=extip:172.18.126.111 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=1f1fa31adfefdec44ade093ad7ea5886545fd287ab5b8c1facfe2abc214da3f5 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=9731 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QOi-e06edCt2WaES5fQOiFc1UodNiyC3FOJBxhmfLSZ_Jk1HA4WAR4-4nDV3Vls0OLKYhT897ZHk2SNnX0IE44ACgmlkgnY0gmlwhKwSEhCKbXVsdGlhZGRyc5YACASsEhIQBtwoAAoErBISEAbcKd0DgnJzhQACAQAAiXNlY3AyNTZrMaECsJv2yTXk7KvHz7agv_m0y7kpNjArXLsMdw2Jjef2wgGDdGNwgtwog3VkcILcKoV3YWt1MgU --content-topic=/newsService/1.0/alerts/xml
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.126.111 waku 54c655378f2636fa878eb1b67c7f8f5a84957cba24d02280ee9e55854867c861
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 54c655378f26. Setting up logs at ./log/docker/node2_2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 9727
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:9727/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.126.111/tcp/9728/p2p/16Uiu2HAmHPxiD6dwEJM5eSVBquU8fCbk11Q8UyhrFSxmBmAJVvqR","/ip4/172.18.126.111/tcp/9729/ws/p2p/16Uiu2HAmHPxiD6dwEJM5eSVBquU8fCbk11Q8UyhrFSxmBmAJVvqR"],"enrUri":"enr:-L24QCksxy6jv6NeRJLq9s-gRyT6HrI2hbpwgROGbzxEkowwMF1YCnZqSAg4xpltc7oG9Qr8-yziixmS7Q2-GjwHAvUCgmlkgnY0gmlwhKwSfm-KbXVsdGlhZGRyc5YACASsEn5vBiYAAAoErBJ-bwYmAd0DgnJzhQACAQAAiXNlY3AyNTZrMaEDRnDecQSYVacJoo4aywmkn-Vc5LhA75lM4TVV7NQSYnyDdGNwgiYAg3VkcIImAoV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:9727/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.18.16/tcp/56360/p2p/16Uiu2HAm7K5fph8W1S3FE71fgPN9KUWVpTZMmJzYiELYBfrzTiGQ"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:56359/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/newsService/1.0/weekly/protobuf"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:9727/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/newsService/1.0/alerts/xml"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:56359/relay/v1/auto/messages" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/newsService/1.0/weekly/protobuf", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:56359/relay/v1/auto/messages/%2FnewsService%2F1.0%2Fweekly%2Fprotobuf" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/newsService/1.0/weekly/protobuf","version":0,"timestamp":1772685330228846530,"ephemeral":false,"proof":""},{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/newsService/1.0/weekly/protobuf","version":0,"timestamp":1772685335372864381,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 17198e4e0f30
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 54c655378f26
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-05_04-35-31__2ba532f6-e204-438a-99fa-2693e0c39916__wakuorg_nwaku:latest.log |