mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-05 10:03:16 +00:00
77 lines
14 KiB
Plaintext
77 lines
14 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_store_sync_range_with_jitter_catches_old_messages with id: 2026-01-30_04-42-42__cd093fa4-8c9c-4afd-9d09-8ff869c7820f
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-01-30_04-42-42__cd093fa4-8c9c-4afd-9d09-8ff869c7820f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-01-30_04-42-42__cd093fa4-8c9c-4afd-9d09-8ff869c7820f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node3_2026-01-30_04-42-42__cd093fa4-8c9c-4afd-9d09-8ff869c7820f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container d5a9d7943b50. Exiting log stream.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.84.240
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['63758', '63759', '63760', '63761', '63762']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 63758:63758 -p 63759:63759 -p 63760:63760 -p 63761:63761 -p 63762:63762 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=63760 --rest-port=63758 --tcp-port=63759 --discv5-udp-port=63761 --rest-address=0.0.0.0 --nat=extip:172.18.84.240 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=66caaf3ac5abc875e5d0c80371e7d9a5ae05ad57f0be997e11d79f0648ddb6c7 --store-sync=true --store=true --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=63762 --metrics-logging=true --relay=true --dns-discovery=false
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.84.240 waku 9537aa2cfb0b0228bca1a541580f46a73b178002e1fa53bab9551db7cf620b53
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 9537aa2cfb0b. Setting up logs at ./log/docker/node1_2026-01-30_04-42-42__cd093fa4-8c9c-4afd-9d09-8ff869c7820f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 63758
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 65a67761195c. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:63758/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Rln Relay":"NOT_MOUNTED"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:63758/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.84.240/tcp/63759/p2p/16Uiu2HAm97qwqGZm9cvQD4qwfZc3ejJc7TuwpLGiydjySuddstYB","/ip4/172.18.84.240/tcp/63760/ws/p2p/16Uiu2HAm97qwqGZm9cvQD4qwfZc3ejJc7TuwpLGiydjySuddstYB"],"enrUri":"enr:-L24QKLZgceUCtykQFTjwXiQv2oO2DSz5uWeh8kO1J95u1KgPMrvSbRtZOUKk5F4RK9uKb8PZURYfk4uzXTPSfVDwMoCgmlkgnY0gmlwhKwSVPCKbXVsdGlhZGRyc5YACASsElTwBvkPAAoErBJU8Ab5EN0DgnJzhQADAQAAiXNlY3AyNTZrMaECy3JuZJMDrbouXj7CzeurleFounLWk9krmSA92gwfCUyDdGNwgvkPg3VkcIL5EYV3YWt1MhM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:63758/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:63758/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:63758/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:63758/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.222.45
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['29937', '29938', '29939', '29940', '29941']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 29937:29937 -p 29938:29938 -p 29939:29939 -p 29940:29940 -p 29941:29941 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=29939 --rest-port=29937 --tcp-port=29938 --discv5-udp-port=29940 --rest-address=0.0.0.0 --nat=extip:172.18.222.45 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=4b2e83fbf6d1656d2aa21f697b274df6476ea0c49a26abaa86e41b7cb4b61bad --store-sync=true --store=true --store-sync-range=20 --store-sync-interval=5 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=29941 --metrics-logging=true --store-sync-relay-jitter=25 --relay=false --dns-discovery=false --discv5-bootstrap-node=enr:-L24QKLZgceUCtykQFTjwXiQv2oO2DSz5uWeh8kO1J95u1KgPMrvSbRtZOUKk5F4RK9uKb8PZURYfk4uzXTPSfVDwMoCgmlkgnY0gmlwhKwSVPCKbXVsdGlhZGRyc5YACASsElTwBvkPAAoErBJU8Ab5EN0DgnJzhQADAQAAiXNlY3AyNTZrMaECy3JuZJMDrbouXj7CzeurleFounLWk9krmSA92gwfCUyDdGNwgvkPg3VkcIL5EYV3YWt1MhM
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.222.45 waku c20bd67aefa5eec49e2e0367f31cd229888157902917cbf96baefc752857cc61
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID c20bd67aefa5. Setting up logs at ./log/docker/node2_2026-01-30_04-42-42__cd093fa4-8c9c-4afd-9d09-8ff869c7820f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 29937
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:29937/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","protocolsHealth":[{"Relay":"NOT_MOUNTED"},{"Rln Relay":"NOT_MOUNTED"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:29937/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.222.45/tcp/29938/p2p/16Uiu2HAmBjTQi213LiUkkF9Shpzci1PB6Au5qTnxvLnaRWVDx7WB","/ip4/172.18.222.45/tcp/29939/ws/p2p/16Uiu2HAmBjTQi213LiUkkF9Shpzci1PB6Au5qTnxvLnaRWVDx7WB"],"enrUri":"enr:-L24QMmZx2WcTtVinoDsGusdqKoY0Yz_HVbzVcit3dCNaqfuPg_u9CHyKvJgMFiYrbFLu6oVgqjKzvnnMx4SgkUCrT4CgmlkgnY0gmlwhKwS3i2KbXVsdGlhZGRyc5YACASsEt4tBnTyAAoErBLeLQZ0890DgnJzhQADAQAAiXNlY3AyNTZrMaEC8kk7YKUA1k8ExThEb4GNDNGugxalKqtLzBE1oGZpFoyDdGNwgnTyg3VkcIJ09IV3YWt1MhI"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29937/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.84.240/tcp/63759/p2p/16Uiu2HAm97qwqGZm9cvQD4qwfZc3ejJc7TuwpLGiydjySuddstYB"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:29937/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&cursor=&pageSize=100&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x1fec4f302931dd6497337c474ad6c040088f8a5d8cd4b234af9a6a68433cd3d4"},{"messageHash":"0xdbe2b78cba169e23b26ada7916b195898efaadabb068eef223838d036a3bffa1"},{"messageHash":"0xbf4e3b6907d30b727fe5a6fad85298bbf3835ab645baeb88ea7fc37051be96e0"}]}'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 9537aa2cfb0b
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id c20bd67aefa5
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |