mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-04-18 17:53:09 +00:00
80 lines
16 KiB
Plaintext
80 lines
16 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_relay_peers_on_shard_contains_connected_peer with id: 2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.filter:filter.py:28 Running fixture setup: filter_setup
|
||
[35mDEBUG [0m src.steps.light_push:light_push.py:28 Running fixture setup: light_push_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node1_2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node2_2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node3_2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/node4_2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.steps.relay:relay.py:28 Running fixture setup: relay_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.127.176
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['31494', '31495', '31496', '31497', '31498']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 31494:31494 -p 31495:31495 -p 31496:31496 -p 31497:31497 -p 31498:31498 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=31496 --rest-port=31494 --tcp-port=31495 --discv5-udp-port=31497 --rest-address=0.0.0.0 --nat=extip:172.18.127.176 --peer-exchange=true --discv5-discovery=false --cluster-id=3 --nodekey=5ef90cfe959c2f19defdfacee0eec1cc73db62ba0bb16de07d81cf2bff9afb02 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=31498 --metrics-logging=true --relay=true --dns-discovery=false
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 2964b6afd147. Exiting log stream.
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 6fbb58f9667f. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.127.176 waku afa558bf553182b4ff4767d5be5c59632bf9bd8d8e761a376a9895f6fff06b70
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID afa558bf5531. Setting up logs at ./log/docker/node1_2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 31494
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31494/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31494/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.127.176/tcp/31495/p2p/16Uiu2HAkxqNYCEWpDHJVsy8aMu8U93SS6H2GLG5dELWeNa3K74zH","/ip4/172.18.127.176/tcp/31496/ws/p2p/16Uiu2HAkxqNYCEWpDHJVsy8aMu8U93SS6H2GLG5dELWeNa3K74zH"],"enrUri":"enr:-La4QKM21hdfMDsy4GbGb4ADsVsTlpuic2KmEAmgEmgG7u3pD9NDFbM62eX74mtvPGI2gy2v74S623MiYbu6A6oSJAsCgmlkgnY0gmlwhKwSf7CKbXVsdGlhZGRyc5YACASsEn-wBnsHAAoErBJ_sAZ7CN0DgnJzhQADAQAAiXNlY3AyNTZrMaECMqVHeLDXjc-wYfRhgZvndQvZfD-WF1IZTlORPe3Z3XaDdGNwgnsHhXdha3UyAQ"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.203.109
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['44633', '44634', '44635', '44636', '44637']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 44633:44633 -p 44634:44634 -p 44635:44635 -p 44636:44636 -p 44637:44637 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=44635 --rest-port=44633 --tcp-port=44634 --discv5-udp-port=44636 --rest-address=0.0.0.0 --nat=extip:172.18.203.109 --peer-exchange=true --discv5-discovery=false --cluster-id=3 --nodekey=f10b5725b466dfe76bb077bc18e5f087dd0f66a5adec28efdf76e6704cd4bac2 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=44637 --metrics-logging=true --relay=true --dns-discovery=false --discv5-bootstrap-node=enr:-La4QKM21hdfMDsy4GbGb4ADsVsTlpuic2KmEAmgEmgG7u3pD9NDFbM62eX74mtvPGI2gy2v74S623MiYbu6A6oSJAsCgmlkgnY0gmlwhKwSf7CKbXVsdGlhZGRyc5YACASsEn-wBnsHAAoErBJ_sAZ7CN0DgnJzhQADAQAAiXNlY3AyNTZrMaECMqVHeLDXjc-wYfRhgZvndQvZfD-WF1IZTlORPe3Z3XaDdGNwgnsHhXdha3UyAQ
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.203.109 waku 5129423a3ffd2b262241ce7074f85a69759fec5067d6ce320f21eabefe6c4522
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 5129423a3ffd. Setting up logs at ./log/docker/node2_2026-02-25_07-25-43__08a674ce-7a76-4c04-aed9-b0806674e522__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 44633
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:44633/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"NOT_MOUNTED"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"NOT_READY","desc":"No Store service peer available yet, neither Store service set up for the node"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:44633/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.203.109/tcp/44634/p2p/16Uiu2HAmJ9rGdLaR7FiuoJH25mnrPsv7tkNY7dDrRGSA8Du5CQef","/ip4/172.18.203.109/tcp/44635/ws/p2p/16Uiu2HAmJ9rGdLaR7FiuoJH25mnrPsv7tkNY7dDrRGSA8Du5CQef"],"enrUri":"enr:-La4QBR_AwUKDKHK2NUPzGACzjJSOEUUi_rc6fOaOEInRdprGBmMbwXBSsox0pmcUpCpOSo9K2LekK7QNHiVw_OAcQACgmlkgnY0gmlwhKwSy22KbXVsdGlhZGRyc5YACASsEsttBq5aAAoErBLLbQauW90DgnJzhQADAQAAiXNlY3AyNTZrMaEDUa8kPfGiVQnVUglG1bZM9FwCz0xxFUvm0rlagb3KgryDdGNwgq5ahXdha3UyAQ"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:31494/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.203.109/tcp/44634/p2p/16Uiu2HAmJ9rGdLaR7FiuoJH25mnrPsv7tkNY7dDrRGSA8Du5CQef"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:44633/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.127.176/tcp/31495/p2p/16Uiu2HAkxqNYCEWpDHJVsy8aMu8U93SS6H2GLG5dELWeNa3K74zH"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31494/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.203.109/tcp/44634/p2p/16Uiu2HAmJ9rGdLaR7FiuoJH25mnrPsv7tkNY7dDrRGSA8Du5CQef","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.38.0-rc.0","origin":"UnknownOrigin"}]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:44633/admin/v1/peers" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'[{"multiaddr":"/ip4/172.18.127.176/tcp/46884/p2p/16Uiu2HAkxqNYCEWpDHJVsy8aMu8U93SS6H2GLG5dELWeNa3K74zH","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.38.0-rc.0","origin":"UnknownOrigin"}]'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:31494/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:44633/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:31494/admin/v1/peers/relay/on/0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"shard":0,"peers":[{"multiaddr":"/ip4/172.18.203.109/tcp/44634/p2p/16Uiu2HAmJ9rGdLaR7FiuoJH25mnrPsv7tkNY7dDrRGSA8Du5CQef","protocols":["/ipfs/id/1.0.0","/libp2p/autonat/1.0.0","/libp2p/circuit/relay/0.2.0/hop","/vac/waku/metadata/1.0.0","/vac/waku/relay/2.0.0","/vac/waku/rendezvous/1.0.0","/ipfs/ping/1.0.0","/vac/waku/filter-push/2.0.0-beta1","/vac/waku/peer-exchange/2.0.0-alpha1"],"shards":[0],"connected":"Connected","agent":"nwaku-v0.38.0-rc.0","origin":"UnknownOrigin","score":0.0}]}'
|
||
[35mDEBUG [0m tests.rest_flags.test_admin_flags:test_admin_flags.py:366 checking shard=0 list: {'shard': 0, 'peers': [{'multiaddr': '/ip4/172.18.203.109/tcp/44634/p2p/16Uiu2HAmJ9rGdLaR7FiuoJH25mnrPsv7tkNY7dDrRGSA8Du5CQef', 'protocols': ['/ipfs/id/1.0.0', '/libp2p/autonat/1.0.0', '/libp2p/circuit/relay/0.2.0/hop', '/vac/waku/metadata/1.0.0', '/vac/waku/relay/2.0.0', '/vac/waku/rendezvous/1.0.0', '/ipfs/ping/1.0.0', '/vac/waku/filter-push/2.0.0-beta1', '/vac/waku/peer-exchange/2.0.0-alpha1'], 'shards': [0], 'connected': 'Connected', 'agent': 'nwaku-v0.38.0-rc.0', 'origin': 'UnknownOrigin', 'score': 0.0}]}
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id afa558bf5531
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 5129423a3ffd
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |