mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-03 00:53:08 +00:00
75 lines
13 KiB
Plaintext
75 lines
13 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_content_topic_and_pubsub_topic_not_in_docker_flags with id: 2026-03-26_04-48-43__b3cac5fb-7194-47d4-a8af-7cc21c351ade
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-26_04-48-43__b3cac5fb-7194-47d4-a8af-7cc21c351ade__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.118.72
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['7891', '7892', '7893', '7894', '7895']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 7891:7891 -p 7892:7892 -p 7893:7893 -p 7894:7894 -p 7895:7895 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=7893 --rest-port=7891 --tcp-port=7892 --discv5-udp-port=7894 --rest-address=0.0.0.0 --nat=extip:172.18.118.72 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=cfe66b4d114defff5efb7dee0bd5d4f1ee2baca311c90abd9b078667eccdbdab --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=7895 --metrics-logging=true --relay=true --filter=true
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container f127a1c8f039. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.118.72 waku 59bea6cbc4f43cfaa93c3ef1a6cb780143b60d78037b781ed75c78df9f4595f8
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 59bea6cbc4f4. Setting up logs at ./log/docker/node1_2026-03-26_04-48-43__b3cac5fb-7194-47d4-a8af-7cc21c351ade__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 7891
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container dc6e3a910516. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7891/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"INITIALIZING","connectionStatus":"Disconnected","protocolsHealth":[]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7891/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"READY"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7891/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.118.72/tcp/7892/p2p/16Uiu2HAm7gmazDF1m1UEcTiMPeA1Q4M22kwJauctv7i8uqwBLCNg","/ip4/172.18.118.72/tcp/7893/ws/p2p/16Uiu2HAm7gmazDF1m1UEcTiMPeA1Q4M22kwJauctv7i8uqwBLCNg"],"enrUri":"enr:-L24QLiEIBqeh8FNmXp7HL6vAHW_47OdmQq4-XmnjwWXqfH6Il8UWYtp0tvOX8juByYjnRzxJgUjluuAWG5P9W4W9ysCgmlkgnY0gmlwhKwSdkiKbXVsdGlhZGRyc5YACASsEnZIBh7UAAoErBJ2SAYe1d0DgnJzhQACAQAAiXNlY3AyNTZrMaECtipLfHHBaeWn5SKtmWsFEPBCbhxqwWsGazrXMOMnp82DdGNwgh7Ug3VkcIIe1oV3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-26_04-48-43__b3cac5fb-7194-47d4-a8af-7cc21c351ade__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.56.144
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['4341', '4342', '4343', '4344', '4345']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 4341:4341 -p 4342:4342 -p 4343:4343 -p 4344:4344 -p 4345:4345 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=4343 --rest-port=4341 --tcp-port=4342 --discv5-udp-port=4344 --rest-address=0.0.0.0 --nat=extip:172.18.56.144 --peer-exchange=true --discv5-discovery=true --cluster-id=2 --nodekey=b31700138bbdb0be030befc3a7b958daec30aff74eac0ffbac330aaecaad1455 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=4345 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QLiEIBqeh8FNmXp7HL6vAHW_47OdmQq4-XmnjwWXqfH6Il8UWYtp0tvOX8juByYjnRzxJgUjluuAWG5P9W4W9ysCgmlkgnY0gmlwhKwSdkiKbXVsdGlhZGRyc5YACASsEnZIBh7UAAoErBJ2SAYe1d0DgnJzhQACAQAAiXNlY3AyNTZrMaECtipLfHHBaeWn5SKtmWsFEPBCbhxqwWsGazrXMOMnp82DdGNwgh7Ug3VkcIIe1oV3YWt1MgU
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.56.144 waku 998a75023039e6481d4f962109249a26405f4cf64de93c7c66db7b31bbea5cba
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 998a75023039. Setting up logs at ./log/docker/node2_2026-03-26_04-48-43__b3cac5fb-7194-47d4-a8af-7cc21c351ade__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 4341
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:4341/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"READY"},{"Legacy Lightpush Client":"READY"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"READY"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:4341/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.56.144/tcp/4342/p2p/16Uiu2HAm8SWeYq12TuDpT6rVf4dpiHyA3w5AbR51Fa4anfxftUqd","/ip4/172.18.56.144/tcp/4343/ws/p2p/16Uiu2HAm8SWeYq12TuDpT6rVf4dpiHyA3w5AbR51Fa4anfxftUqd"],"enrUri":"enr:-L24QG4nx-XkjstZ4qBv-oLefHONlErVcTOxcLYnQoxXdWUJZGSfVfvAxsEbuDCOVTdFaC601n4uZd3KjBfymdPkO7ACgmlkgnY0gmlwhKwSOJCKbXVsdGlhZGRyc5YACASsEjiQBhD2AAoErBI4kAYQ990DgnJzhQACAQAAiXNlY3AyNTZrMaECwV71mKcns-3R6CyxQS3_RJaWHsbPFsoJnsNE2s7WIoiDdGNwghD2g3VkcIIQ-IV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:4341/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.118.72/tcp/7892/p2p/16Uiu2HAm7gmazDF1m1UEcTiMPeA1Q4M22kwJauctv7i8uqwBLCNg"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7891/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/myapp/1/latest/proto"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:4341/relay/v1/auto/subscriptions" -H "Content-Type: application/json" -d '["/myapp/1/latest/proto"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7891/relay/v1/auto/messages" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7891/relay/v1/auto/messages/%2Fmyapp%2F1%2Flatest%2Fproto" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1774500526464560175,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_2:wakuorg/nwaku:latest can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:4341/relay/v1/auto/messages/%2Fmyapp%2F1%2Flatest%2Fproto" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1774500526464560175,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 59bea6cbc4f4
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 998a75023039
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |