mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-03 09:03:25 +00:00
264 lines
46 KiB
Plaintext
264 lines
46 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_subscribe_one_by_one_to_different_pubsub_topics_and_send_messages with id: 2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.213.154
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['41790', '41791', '41792', '41793', '41794']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 41790:41790 -p 41791:41791 -p 41792:41792 -p 41793:41793 -p 41794:41794 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=41792 --rest-port=41790 --tcp-port=41791 --discv5-udp-port=41793 --rest-address=0.0.0.0 --nat=extip:172.18.213.154 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=5c9601348d4f9efd5eb8c78dcaed19f5e8c9b02093b84aa09aad828b1f1ba9fb --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=41794 --metrics-logging=true --relay=true --filter=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.213.154 waku 670f0306d12062ccdb07b12b3a8cd2079cf340256cbb8e824aaa143c8f5e217c
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 670f0306d120. Setting up logs at ./log/docker/node1_2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 41790
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 289fbbc6a11c. Exiting log stream.
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container b6b50ca23770. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41790/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41790/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41790/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.213.154/tcp/41791/p2p/16Uiu2HAkuo2T9PWTTioYRcRvPFPJGGydF3pBEdv7QMwhUQj3YAa5","/ip4/172.18.213.154/tcp/41792/ws/p2p/16Uiu2HAkuo2T9PWTTioYRcRvPFPJGGydF3pBEdv7QMwhUQj3YAa5"],"enrUri":"enr:-L24QBJb8kChmrj3vELt32MivUgZ0nj9tyTXczU3vrTujRmRHSFSZBW7Z7PxmJxIMM-bZDCyRLIt-H16ksrUEW5kiWgCgmlkgnY0gmlwhKwS1ZqKbXVsdGlhZGRyc5YACASsEtWaBqM_AAoErBLVmgajQN0DgnJzhQACAQAAiXNlY3AyNTZrMaECBXhYIxXUkQb3slI23tjVTVfVLpzSBFW4OKFxpZaKpjqDdGNwgqM_g3VkcIKjQYV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.99.173
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['38097', '38098', '38099', '38100', '38101']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 38097:38097 -p 38098:38098 -p 38099:38099 -p 38100:38100 -p 38101:38101 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=38099 --rest-port=38097 --tcp-port=38098 --discv5-udp-port=38100 --rest-address=0.0.0.0 --nat=extip:172.18.99.173 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=dcff4dc8fe79cd01ab6da71d8dc4d406bf16b970df58789cc3dcebba06866ce2 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=38101 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QBJb8kChmrj3vELt32MivUgZ0nj9tyTXczU3vrTujRmRHSFSZBW7Z7PxmJxIMM-bZDCyRLIt-H16ksrUEW5kiWgCgmlkgnY0gmlwhKwS1ZqKbXVsdGlhZGRyc5YACASsEtWaBqM_AAoErBLVmgajQN0DgnJzhQACAQAAiXNlY3AyNTZrMaECBXhYIxXUkQb3slI23tjVTVfVLpzSBFW4OKFxpZaKpjqDdGNwgqM_g3VkcIKjQYV3YWt1MgU
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.99.173 waku 993f6813c392bbcbaa978523a471d9aa13201624d9b4d4d345ea9afab23f0995
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 993f6813c392. Setting up logs at ./log/docker/node2_2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 38097
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:38097/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:38097/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:38097/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:38097/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:38097/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:38097/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.99.173/tcp/38098/p2p/16Uiu2HAm17LM2DZeVcY2aS1SaxrS4ShnfE7E8DqZEWTwVGyh3fEu","/ip4/172.18.99.173/tcp/38099/ws/p2p/16Uiu2HAm17LM2DZeVcY2aS1SaxrS4ShnfE7E8DqZEWTwVGyh3fEu"],"enrUri":"enr:-L24QDpEsVpPb7wNtYCXHLqi2PmJRpyE8T-CRk40BdyWc0wIDkfwPcfG1s906k6ampWoBw4FUrX5jY2b_fZC9ZRTNgYCgmlkgnY0gmlwhKwSY62KbXVsdGlhZGRyc5YACASsEmOtBpTSAAoErBJjrQaU090DgnJzhQACAQAAiXNlY3AyNTZrMaECVHN3ffKFq8NYO1qmtppa6wXy63TghKEklEQGdvZICg6DdGNwgpTSg3VkcIKU1IV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:38097/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.213.154/tcp/41791/p2p/16Uiu2HAkuo2T9PWTTioYRcRvPFPJGGydF3pBEdv7QMwhUQj3YAa5"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:41790/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/2/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:38097/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/2/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:41790/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41790/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376509042404129,"ephemeral":false,"proof":""},{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376510450618049,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 670f0306d120
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 993f6813c392
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-13_04-35-07__8394f9e5-1339-48c7-b626-c98b8b583414__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_subscribe_one_by_one_to_different_pubsub_topics_and_send_messages with id: 2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.75.117
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['23791', '23792', '23793', '23794', '23795']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 23791:23791 -p 23792:23792 -p 23793:23793 -p 23794:23794 -p 23795:23795 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=23793 --rest-port=23791 --tcp-port=23792 --discv5-udp-port=23794 --rest-address=0.0.0.0 --nat=extip:172.18.75.117 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=a1ac07c0af97bcad4bef04f625d7195c37cdaddfcf8aa333d53e95e25ace9601 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=23795 --metrics-logging=true --relay=true --filter=true
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 670f0306d120. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.75.117 waku b13508e6ccc71b74e19c8d47738d64356b4915452a71f794963786774d1bcf18
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID b13508e6ccc7. Setting up logs at ./log/docker/node1_2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 23791
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 993f6813c392. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Connected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.75.117/tcp/23792/p2p/16Uiu2HAmFf9y8x8Xc9q6rAf28kTWJXb2keoD1qkV2xzRPTpBU9TP","/ip4/172.18.75.117/tcp/23793/ws/p2p/16Uiu2HAmFf9y8x8Xc9q6rAf28kTWJXb2keoD1qkV2xzRPTpBU9TP"],"enrUri":"enr:-L24QITQufnrjSVSkyDNpHnpkEcdywRrakKeA1mtwRbSAvkqUrvBBK2D11AmphumX_R8DOObFMYQgk_zdH4mffCtvTACgmlkgnY0gmlwhKwSS3WKbXVsdGlhZGRyc5YACASsEkt1BlzwAAoErBJLdQZc8d0DgnJzhQACAQAAiXNlY3AyNTZrMaEDLJ3vm2FLUloHsi5bidi-9SskEBsFAKyDINtk74EqO9KDdGNwglzwg3VkcIJc8oV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.149.158
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['2828', '2829', '2830', '2831', '2832']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 2828:2828 -p 2829:2829 -p 2830:2830 -p 2831:2831 -p 2832:2832 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=2830 --rest-port=2828 --tcp-port=2829 --discv5-udp-port=2831 --rest-address=0.0.0.0 --nat=extip:172.18.149.158 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=23eeee003eaab8f7c0f9dd0ebfdc328bba2c41dbd523dbb40cf15d2dac80b505 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=2832 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QITQufnrjSVSkyDNpHnpkEcdywRrakKeA1mtwRbSAvkqUrvBBK2D11AmphumX_R8DOObFMYQgk_zdH4mffCtvTACgmlkgnY0gmlwhKwSS3WKbXVsdGlhZGRyc5YACASsEkt1BlzwAAoErBJLdQZc8d0DgnJzhQACAQAAiXNlY3AyNTZrMaEDLJ3vm2FLUloHsi5bidi-9SskEBsFAKyDINtk74EqO9KDdGNwglzwg3VkcIJc8oV3YWt1MgU
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.149.158 waku 0d4f295181b89cddc53318d707eb9d93b679fd017f0103de21f33197b83b38b1
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 0d4f295181b8. Setting up logs at ./log/docker/node2_2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 2828
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:2828/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:2828/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:2828/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.149.158/tcp/2829/p2p/16Uiu2HAmBYELLiuTF2c4ZBni8zqJZK7rDdkUbvSWqnUtUzdW9EB1","/ip4/172.18.149.158/tcp/2830/ws/p2p/16Uiu2HAmBYELLiuTF2c4ZBni8zqJZK7rDdkUbvSWqnUtUzdW9EB1"],"enrUri":"enr:-L24QNdGxJu6fj8gdaqGwzy9e_R7XPWFIDzDpkpiLzIfOeVPAt79sK2-U68oB4Nszj43-cxOyNlTiJA2-QicVb1eIeMCgmlkgnY0gmlwhKwSlZ6KbXVsdGlhZGRyc5YACASsEpWeBgsNAAoErBKVngYLDt0DgnJzhQACAQAAiXNlY3AyNTZrMaEC72kPCt60wY-mkp5PnMzTK7xm4TYbvTLy6DeAMoz1GGiDdGNwggsNg3VkcIILD4V3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:2828/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.75.117/tcp/23792/p2p/16Uiu2HAmFf9y8x8Xc9q6rAf28kTWJXb2keoD1qkV2xzRPTpBU9TP"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:23791/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/2/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:2828/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/2/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:23791/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:23791/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376514966831403,"ephemeral":false,"proof":""},{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376515247226714,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id b13508e6ccc7
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 0d4f295181b8
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-13_04-35-11__d2664650-82c6-41ad-90c3-4d26771e7808__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_subscribe_one_by_one_to_different_pubsub_topics_and_send_messages with id: 2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.69.11
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['41787', '41788', '41789', '41790', '41791']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 41787:41787 -p 41788:41788 -p 41789:41789 -p 41790:41790 -p 41791:41791 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=41789 --rest-port=41787 --tcp-port=41788 --discv5-udp-port=41790 --rest-address=0.0.0.0 --nat=extip:172.18.69.11 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=46eb56c4294bc4e82d1b00df0bf1dd1a346fc39b88e73f8ca82cffff86df5d62 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=41791 --metrics-logging=true --relay=true --filter=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.69.11 waku 16bc1264a3f554f571742bc2309f2fae49dd9ada286a7439879bd6ec8287c839
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 16bc1264a3f5. Setting up logs at ./log/docker/node1_2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 41787
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container b13508e6ccc7. Exiting log stream.
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 0d4f295181b8. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41787/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41787/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41787/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41787/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.69.11/tcp/41788/p2p/16Uiu2HAmDgQsDnvV6saDdRGez6zx6EGaMFxm9PAbSDteJ1FYgzNg","/ip4/172.18.69.11/tcp/41789/ws/p2p/16Uiu2HAmDgQsDnvV6saDdRGez6zx6EGaMFxm9PAbSDteJ1FYgzNg"],"enrUri":"enr:-L24QBh10SZzf-LvmkD5TQ-npEzNdJANYv_SDVf3DqFT4gUZLMnVzvz9e6t80URjYmFktDf8619PP0RF6yDfsr3y78ECgmlkgnY0gmlwhKwSRQuKbXVsdGlhZGRyc5YACASsEkULBqM8AAoErBJFCwajPd0DgnJzhQACAQAAiXNlY3AyNTZrMaEDDzj89kjBw_67kzgBL0C97xy4WHw64eYuF1EDqqOarzWDdGNwgqM8g3VkcIKjPoV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.58.57
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['34207', '34208', '34209', '34210', '34211']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 34207:34207 -p 34208:34208 -p 34209:34209 -p 34210:34210 -p 34211:34211 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=34209 --rest-port=34207 --tcp-port=34208 --discv5-udp-port=34210 --rest-address=0.0.0.0 --nat=extip:172.18.58.57 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=da2ab485eb1bffbfedd40167df18adaa39ea221487ccc9304ca97c78d1ae932e --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=34211 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QBh10SZzf-LvmkD5TQ-npEzNdJANYv_SDVf3DqFT4gUZLMnVzvz9e6t80URjYmFktDf8619PP0RF6yDfsr3y78ECgmlkgnY0gmlwhKwSRQuKbXVsdGlhZGRyc5YACASsEkULBqM8AAoErBJFCwajPd0DgnJzhQACAQAAiXNlY3AyNTZrMaEDDzj89kjBw_67kzgBL0C97xy4WHw64eYuF1EDqqOarzWDdGNwgqM8g3VkcIKjPoV3YWt1MgU
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.58.57 waku fe2cd468907d8388894287352fa742bf8f94b91625e97ae4bc9a0144e881478e
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID fe2cd468907d. Setting up logs at ./log/docker/node2_2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 34207
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34207/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34207/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34207/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.58.57/tcp/34208/p2p/16Uiu2HAkzn2Gp4QfjXMprtRqDD3eVDqEM6wa6GZu3Ce5f4CkTbJm","/ip4/172.18.58.57/tcp/34209/ws/p2p/16Uiu2HAkzn2Gp4QfjXMprtRqDD3eVDqEM6wa6GZu3Ce5f4CkTbJm"],"enrUri":"enr:-L24QDNk6H-JqZtoIxU4qfheVNeq33K1DZfHbEtwilP1DWlUS_y8gCxHt8FE5b_aldkGmufFdQuDzbwhJlxt6okUl0kCgmlkgnY0gmlwhKwSOjmKbXVsdGlhZGRyc5YACASsEjo5BoWgAAoErBI6OQaFod0DgnJzhQACAQAAiXNlY3AyNTZrMaECT4D_JUEX4bVi2u2mX7U2Dl3IzfjTOxp-d7ZFdAOQTi6DdGNwgoWgg3VkcIKFooV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:34207/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.69.11/tcp/41788/p2p/16Uiu2HAmDgQsDnvV6saDdRGez6zx6EGaMFxm9PAbSDteJ1FYgzNg"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:41787/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/2/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:34207/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/2/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:41787/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:41787/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376519481489399,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_2:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34207/relay/v1/messages/%2Fwaku%2F2%2Frs%2F2%2F0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376519481489399,"ephemeral":false,"proof":""},{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773376519551847461,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 16bc1264a3f5
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id fe2cd468907d
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:141 Found matches for keyword 'fatal': ['/app/vendor/nimbus-build-system/vendor/Nim/lib/system/fatal.nim(53) _ZN10assertions8sysFatalE8typeDescIN10exceptions15AssertionDefectEE6string']
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:141 Found matches for keyword 'unhandled': ['Error: unhandled exception: /app/vendor/nim-json-serialization/json_serialization/writer.nim(191, 3) `w.inObject()` [AssertionDefect]']
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:141 Found matches for keyword 'stacktrace': ['/app/vendor/nimbus-build-system/vendor/Nim/lib/system/excpt.nim(349) _ZN6system18rawWriteStackTraceE3varI3seqIN6system15StackTraceEntryEEE', '/app/vendor/nimbus-build-system/vendor/Nim/lib/system/stacktraces.nim(62) _ZN11stacktraces30auxWriteStackTraceWithOverrideE3varI3seqIN6system15StackTraceEntryEEE']
|
||
[35mDEBUG [0m tests.conftest:conftest.py:73 Running fixture teardown: attach_logs_on_fail
|
||
[35mDEBUG [0m tests.conftest:conftest.py:74 Test failed, attempting to attach logs to the allure reports
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node2_2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.libs.common:common.py:42 Attaching file ./log/docker/node1_2026-03-13_04-35-16__2cf18cf2-e57a-4aef-96f5-4600948f8429__wakuorg_nwaku:latest.log |