mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-04-29 15:13:48 +00:00
100 lines
18 KiB
Plaintext
100 lines
18 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_store_without_pubsub_topic_and_content_topic with id: 2026-03-10_04-34-39__4ee15bac-f983-460d-9c8e-bad86bcdac4f
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:39 Running fixture setup: node_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/publishing_node1_2026-03-10_04-34-39__4ee15bac-f983-460d-9c8e-bad86bcdac4f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.137.171
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['8128', '8129', '8130', '8131', '8132']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 8128:8128 -p 8129:8129 -p 8130:8130 -p 8131:8131 -p 8132:8132 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=8130 --rest-port=8128 --tcp-port=8129 --discv5-udp-port=8131 --rest-address=0.0.0.0 --nat=extip:172.18.137.171 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=3a1d92c7fe8e99d0cc7a86d796f416b8bf4d77df8d05afddddf9e0cfebeecbde --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=8132 --metrics-logging=true --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.137.171 waku 21bc6776c06ee10dc99624815a2518056225ce07c66ec527484f6ccf1787fafb
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 86da0e11d9a1. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 21bc6776c06e. Setting up logs at ./log/docker/publishing_node1_2026-03-10_04-34-39__4ee15bac-f983-460d-9c8e-bad86bcdac4f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 8128
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 773e1002ed6b. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:8128/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:8128/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.137.171/tcp/8129/p2p/16Uiu2HAmC5SV4ZqpnnuYKwGEbFzPtyPKZhAzfxXeiH79NAEGTw9N","/ip4/172.18.137.171/tcp/8130/ws/p2p/16Uiu2HAmC5SV4ZqpnnuYKwGEbFzPtyPKZhAzfxXeiH79NAEGTw9N"],"enrUri":"enr:-L24QPt6Blf2BFAZWvOMzq61ppY97KpgICFPUXHa0i3eRGaQH6uL2M59lytWIwKtnjXzFSLyYdNsHYQdP88hfK6XOwYCgmlkgnY0gmlwhKwSiauKbXVsdGlhZGRyc5YACASsEomrBh_BAAoErBKJqwYfwt0DgnJzhQADAQAAiXNlY3AyNTZrMaEC92fNZAr7aIdkwlU6Od5Ksv1TgrFx5UTSB2h8LQRMvV2DdGNwgh_Bg3VkcIIfw4V3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/store_node1_2026-03-10_04-34-39__4ee15bac-f983-460d-9c8e-bad86bcdac4f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.164.88
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['52955', '52956', '52957', '52958', '52959']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 52955:52955 -p 52956:52956 -p 52957:52957 -p 52958:52958 -p 52959:52959 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=52957 --rest-port=52955 --tcp-port=52956 --discv5-udp-port=52958 --rest-address=0.0.0.0 --nat=extip:172.18.164.88 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=11268dd2d4bc172d7ecfe9696d4eaedcaed8aba27eece37ac1edb3feae7fcc8c --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=52959 --metrics-logging=true --discv5-bootstrap-node=enr:-L24QPt6Blf2BFAZWvOMzq61ppY97KpgICFPUXHa0i3eRGaQH6uL2M59lytWIwKtnjXzFSLyYdNsHYQdP88hfK6XOwYCgmlkgnY0gmlwhKwSiauKbXVsdGlhZGRyc5YACASsEomrBh_BAAoErBKJqwYfwt0DgnJzhQADAQAAiXNlY3AyNTZrMaEC92fNZAr7aIdkwlU6Od5Ksv1TgrFx5UTSB2h8LQRMvV2DdGNwgh_Bg3VkcIIfw4V3YWt1MgM --storenode=/ip4/172.18.137.171/tcp/8129/p2p/16Uiu2HAmC5SV4ZqpnnuYKwGEbFzPtyPKZhAzfxXeiH79NAEGTw9N --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.164.88 waku b3c6c32de6f0d6dc993146978743e98a370a0827771430efca397bcf35c5f3a8
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID b3c6c32de6f0. Setting up logs at ./log/docker/store_node1_2026-03-10_04-34-39__4ee15bac-f983-460d-9c8e-bad86bcdac4f__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 52955
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52955/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52955/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.164.88/tcp/52956/p2p/16Uiu2HAm5aygVJH9qWDHJ6Q8n2Zi1PispvvAFMEm7KnWX96mYq2o","/ip4/172.18.164.88/tcp/52957/ws/p2p/16Uiu2HAm5aygVJH9qWDHJ6Q8n2Zi1PispvvAFMEm7KnWX96mYq2o"],"enrUri":"enr:-L24QBhWViy4b-FuCmdnRN0OPOSKCwdLh_9tK8k6bfT2YktKAFJ6hIgMD80enCciYUkjmOFmnEbZNBQyTKcC_b81BPoCgmlkgnY0gmlwhKwSpFiKbXVsdGlhZGRyc5YACASsEqRYBs7cAAoErBKkWAbO3d0DgnJzhQADAQAAiXNlY3AyNTZrMaEClvccooQ6GAqy_efSumxiw6bUsfkizTSy7TNYdiCCwOCDdGNwgs7cg3VkcILO3oV3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:52955/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.137.171/tcp/8129/p2p/16Uiu2HAmC5SV4ZqpnnuYKwGEbFzPtyPKZhAzfxXeiH79NAEGTw9N"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:52955/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/waku/2/content/test.js", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/app/22/sometopic/someencoding", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/toychat/2/huilong/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/statusim/1/community/cbor", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/app/27/sometopic/someencoding", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/app/29/sometopic/someencoding", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:8128/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/app/20/sometopic/someencoding", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:8128/store/v3/messages?pageSize=20&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x6a3228b30574ff746d169b9dfa0850f06979474c37b348cc5af2730120835b9c"},{"messageHash":"0xd27960b87358ea0720907f6e00491b3edb4d8f8930e9c5d80f9a66a2014fbf09"},{"messageHash":"0x9d9972b959a405e0367c930463b7ab58a5924100e5703fd6084821080e7c4a66"},{"messageHash":"0x4196cae8702ba08ba5f1a444a4b98bb0829472fa390dc8638e761c5eb6a55a40"},{"messageHash":"0xb29bd9df0d3baa5726c6f9bb0c5dacee9199040a06c28cbc0452f3cd7d2e106f"},{"messageHash":"0xf48d0d13b90d4052ffa2897bd1d8d49586a6a32180c64592e070454b0a4416bc"},{"messageHash":"0xf4c18d546eab899798349c6364b0d2de9130742061c902adc027452e433c7075"},{"messageHash":"0xc7fbeff1bfbad05ad2e89abcc97517b2946323906b32b18ec2f65e76afb0ac9d"}]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:52955/store/v3/messages?pageSize=20&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x6a3228b30574ff746d169b9dfa0850f06979474c37b348cc5af2730120835b9c"},{"messageHash":"0xd27960b87358ea0720907f6e00491b3edb4d8f8930e9c5d80f9a66a2014fbf09"},{"messageHash":"0x9d9972b959a405e0367c930463b7ab58a5924100e5703fd6084821080e7c4a66"},{"messageHash":"0x4196cae8702ba08ba5f1a444a4b98bb0829472fa390dc8638e761c5eb6a55a40"},{"messageHash":"0xb29bd9df0d3baa5726c6f9bb0c5dacee9199040a06c28cbc0452f3cd7d2e106f"},{"messageHash":"0xf48d0d13b90d4052ffa2897bd1d8d49586a6a32180c64592e070454b0a4416bc"},{"messageHash":"0xf4c18d546eab899798349c6364b0d2de9130742061c902adc027452e433c7075"},{"messageHash":"0xc7fbeff1bfbad05ad2e89abcc97517b2946323906b32b18ec2f65e76afb0ac9d"}]}'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 21bc6776c06e
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id b3c6c32de6f0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |