mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-02 16:43:56 +00:00
77 lines
14 KiB
Plaintext
77 lines
14 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_single_pubsub_topic[/waku/2/rs/999/999] with id: 2026-03-18_09-11-15__cf240117-4562-4d23-98b3-3b4d83492978
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:30 Running fixture setup: sharding_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-03-18_09-11-15__cf240117-4562-4d23-98b3-3b4d83492978__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.59.239
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['50792', '50793', '50794', '50795', '50796']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:72 Cluster id was resolved to: 999
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 50792:50792 -p 50793:50793 -p 50794:50794 -p 50795:50795 -p 50796:50796 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=50794 --rest-port=50792 --tcp-port=50793 --discv5-udp-port=50795 --rest-address=0.0.0.0 --nat=extip:172.18.59.239 --peer-exchange=true --discv5-discovery=true --cluster-id=999 --nodekey=2ff2aebd9dcb063e88bcec84bf1decb24d959f4abdc708bba958bad23bfbb230 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=50796 --metrics-logging=true --relay=true --filter=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.59.239 waku 4f9b86d8f1fca930a8b95edd25e77947a7b44a4d9ed7475c27c0c9b3a9d2ed53
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container bce97b6d72c3. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 4f9b86d8f1fc. Setting up logs at ./log/docker/node1_2026-03-18_09-11-15__cf240117-4562-4d23-98b3-3b4d83492978__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 50792
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 4882ba7cfbce. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:50792/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_READY","desc":"Relay is not ready, filter will not be able to sort out messages"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:50792/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.59.239/tcp/50793/p2p/16Uiu2HAm6iAvAqPnmPqnLe82W3bVVabu7cGsohzMnmVftmYRpKsK","/ip4/172.18.59.239/tcp/50794/ws/p2p/16Uiu2HAm6iAvAqPnmPqnLe82W3bVVabu7cGsohzMnmVftmYRpKsK"],"enrUri":"enr:-L24QBzubHLIkRwTPYNnSJ1xLkhEdv25Gqk896nUBzjfLhXeXdABgZL-y_MVMKq0-O1BfYzrknUKJgGjA5iSNNDT7mACgmlkgnY0gmlwhKwSO--KbXVsdGlhZGRyc5YACASsEjvvBsZpAAoErBI77wbGat0DgnJzhQPnAQAAiXNlY3AyNTZrMaECp6qRo-EO8IgH1Dv0sodLP2lHPk30le5vHTTkoCSTfaaDdGNwgsZpg3VkcILGa4V3YWt1MgU"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-03-18_09-11-15__cf240117-4562-4d23-98b3-3b4d83492978__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.10.165
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['34629', '34630', '34631', '34632', '34633']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:72 Cluster id was resolved to: 999
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:163 Removing pubsub-topic from nwaku args
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 34629:34629 -p 34630:34630 -p 34631:34631 -p 34632:34632 -p 34633:34633 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=34631 --rest-port=34629 --tcp-port=34630 --discv5-udp-port=34632 --rest-address=0.0.0.0 --nat=extip:172.18.10.165 --peer-exchange=true --discv5-discovery=true --cluster-id=999 --nodekey=de7b5b880c2ece2da0aff4ab53b9e278ad54d7daad2d61ddb0b3cfb2e328bebc --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=34633 --metrics-logging=true --relay=true --discv5-bootstrap-node=enr:-L24QBzubHLIkRwTPYNnSJ1xLkhEdv25Gqk896nUBzjfLhXeXdABgZL-y_MVMKq0-O1BfYzrknUKJgGjA5iSNNDT7mACgmlkgnY0gmlwhKwSO--KbXVsdGlhZGRyc5YACASsEjvvBsZpAAoErBI77wbGat0DgnJzhQPnAQAAiXNlY3AyNTZrMaECp6qRo-EO8IgH1Dv0sodLP2lHPk30le5vHTTkoCSTfaaDdGNwgsZpg3VkcILGa4V3YWt1MgU
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.10.165 waku 351c7c272eae75136c799366052be94d2d0421b3efb72c70adde484e1ef142ae
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 351c7c272eae. Setting up logs at ./log/docker/node2_2026-03-18_09-11-15__cf240117-4562-4d23-98b3-3b4d83492978__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 34629
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34629/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34629/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.10.165/tcp/34630/p2p/16Uiu2HAmLNUP5nqD5YcUCNbJxYf2d2WYpmiU3s3TAPMUEn6kG5G7","/ip4/172.18.10.165/tcp/34631/ws/p2p/16Uiu2HAmLNUP5nqD5YcUCNbJxYf2d2WYpmiU3s3TAPMUEn6kG5G7"],"enrUri":"enr:-L24QGdIxCKPtWuJKygdrfsHw_C8--UmrQhtOf3wOr64GEyUZbDDfKxiC58jmfdje2BLeHO6pkLgaiw3zhZFYG1D5w4CgmlkgnY0gmlwhKwSCqWKbXVsdGlhZGRyc5YACASsEgqlBodGAAoErBIKpQaHR90DgnJzhQPnAQAAiXNlY3AyNTZrMaEDcqJQ-nztTlDgJ9-97p3yWtCYZCADSXS5bPN9ZBaVwuSDdGNwgodGg3VkcIKHSIV3YWt1MgE"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:34629/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.59.239/tcp/50793/p2p/16Uiu2HAm6iAvAqPnmPqnLe82W3bVVabu7cGsohzMnmVftmYRpKsK"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:50792/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/999/999"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:34629/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/999/999"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:50792/relay/v1/messages/%2Fwaku%2F2%2Frs%2F999%2F999" -H "Content-Type: application/json" -d '{"payload": "U2hhcmRpbmcgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.1 seconds
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_1:harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:50792/relay/v1/messages/%2Fwaku%2F2%2Frs%2F999%2F999" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773825077799480349,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m src.steps.sharding:sharding.py:158 Checking that peer NODE_2:harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 can find the published message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:34629/relay/v1/messages/%2Fwaku%2F2%2Frs%2F999%2F999" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"payload":"U2hhcmRpbmcgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773825077799480349,"ephemeral":false,"proof":""}]'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 4f9b86d8f1fc
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 351c7c272eae
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |