mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-04 09:33:42 +00:00
95 lines
18 KiB
Plaintext
95 lines
18 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_time_filter_end_time_now with id: 2026-03-16_12-36-31__7f4d7a2a-4aeb-429d-ba53-2e6fb8e9b4e1
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:39 Running fixture setup: node_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/publishing_node1_2026-03-16_12-36-31__7f4d7a2a-4aeb-429d-ba53-2e6fb8e9b4e1__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.118.209
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['29782', '29783', '29784', '29785', '29786']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 29782:29782 -p 29783:29783 -p 29784:29784 -p 29785:29785 -p 29786:29786 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=29784 --rest-port=29782 --tcp-port=29783 --discv5-udp-port=29785 --rest-address=0.0.0.0 --nat=extip:172.18.118.209 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=4a30fb7da6be87041d8cd200e93ea536c3c3d277f4da3b8574e19e7de4aedec1 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=29786 --metrics-logging=true --store=true --relay=true
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container defcc321d0a4. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.118.209 waku b316d198ee4bb7549cc4584436339f79d57562cc6446d4baf6147e12f2a05fb4
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID b316d198ee4b. Setting up logs at ./log/docker/publishing_node1_2026-03-16_12-36-31__7f4d7a2a-4aeb-429d-ba53-2e6fb8e9b4e1__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 29782
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[1m[31mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container f67353623f70. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:29782/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:29782/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.118.209/tcp/29783/p2p/16Uiu2HAm4P3kTyUQDaKZzGwLyygPQ6fnCmAoZd2b6XeEHvXBPzjw","/ip4/172.18.118.209/tcp/29784/ws/p2p/16Uiu2HAm4P3kTyUQDaKZzGwLyygPQ6fnCmAoZd2b6XeEHvXBPzjw"],"enrUri":"enr:-L24QPXm2a1ZXq8_N0WXa9AArshxnrKbyE22jmTUxxK9FpipUap-ROZY-lUfOc8xlDQGgW5GVPnqf8MKYKmOh8rAEAwCgmlkgnY0gmlwhKwSdtGKbXVsdGlhZGRyc5YACASsEnbRBnRXAAoErBJ20QZ0WN0DgnJzhQADAQAAiXNlY3AyNTZrMaEChQ0OObuCYb413zqHmob3kl7I8tQbjpS2exioH18_o86DdGNwgnRXg3VkcIJ0WYV3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/store_node1_2026-03-16_12-36-31__7f4d7a2a-4aeb-429d-ba53-2e6fb8e9b4e1__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.179.98
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['42494', '42495', '42496', '42497', '42498']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 42494:42494 -p 42495:42495 -p 42496:42496 -p 42497:42497 -p 42498:42498 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=42496 --rest-port=42494 --tcp-port=42495 --discv5-udp-port=42497 --rest-address=0.0.0.0 --nat=extip:172.18.179.98 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=2c53e157f28dad3fb0cea60cc4e80c005dd928f9a4f5fef7c48fd0fd60e6efeb --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=42498 --metrics-logging=true --discv5-bootstrap-node=enr:-L24QPXm2a1ZXq8_N0WXa9AArshxnrKbyE22jmTUxxK9FpipUap-ROZY-lUfOc8xlDQGgW5GVPnqf8MKYKmOh8rAEAwCgmlkgnY0gmlwhKwSdtGKbXVsdGlhZGRyc5YACASsEnbRBnRXAAoErBJ20QZ0WN0DgnJzhQADAQAAiXNlY3AyNTZrMaEChQ0OObuCYb413zqHmob3kl7I8tQbjpS2exioH18_o86DdGNwgnRXg3VkcIJ0WYV3YWt1MgM --storenode=/ip4/172.18.118.209/tcp/29783/p2p/16Uiu2HAm4P3kTyUQDaKZzGwLyygPQ6fnCmAoZd2b6XeEHvXBPzjw --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.179.98 waku 511da105d6b70e2fc28c433d529dcc6b890d4f74db221456e3c77daaa127be7f
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 511da105d6b7. Setting up logs at ./log/docker/store_node1_2026-03-16_12-36-31__7f4d7a2a-4aeb-429d-ba53-2e6fb8e9b4e1__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 42494
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:42494/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:42494/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.179.98/tcp/42495/p2p/16Uiu2HAm1cvSY6SkYgXzrhZa8ADfrwxUPJfiJnVF7rBSmpSHBPBa","/ip4/172.18.179.98/tcp/42496/ws/p2p/16Uiu2HAm1cvSY6SkYgXzrhZa8ADfrwxUPJfiJnVF7rBSmpSHBPBa"],"enrUri":"enr:-L24QHLBIrh-GIy_-0o7KJiIVjeC_L09nqWwD5cR_mgHag3NfqFAdWO3wlQovHpLXxEvkXpt-G-8UVT-yxFQpyDR2IACgmlkgnY0gmlwhKwSs2KKbXVsdGlhZGRyc5YACASsErNiBqX_AAoErBKzYgamAN0DgnJzhQADAQAAiXNlY3AyNTZrMaECXAfclkb5kcZBdN2OdgisCXdy9QitcYhSJbjzPUTVO22DdGNwgqX_g3VkcIKmAYV3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:42494/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.118.209/tcp/29783/p2p/16Uiu2HAm4P3kTyUQDaKZzGwLyygPQ6fnCmAoZd2b6XeEHvXBPzjw"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:42494/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:29782/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m tests.store.test_time_filter:test_time_filter.py:186 inquering stored messages with start time 1773664590720292096 after end time 1773664594965656064
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:29782/store/v3/messages?includeData=True&pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&startTime=1773664590720292096&endTime=1773664594965656064&pageSize=20&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x350df42ae9c10cf8e856ae9fa68fa05627e6b6bd1bdaacdb6a7986c432bf9a8f","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773664590720292096,"ephemeral":false},"pubsubTopic":"/waku/2/rs/3/0"},{"messageHash":"0xe02faf5cafd15457a3b60b4620f608edfccc6b7488c8c8cfaa402271d8d87eeb","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773664592720299008,"ephemeral":false},"pubsubTopic":"/waku/2/rs/3/0"},{"messageHash":"0x31a2e5fad52f35f80434bf66f9b4e27af767c303218278670763d343ada4b6e7","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773664593620301056,"ephemeral":false},"pubsubTopic":"/waku/2/rs/3/0"}]}'
|
||
[35mDEBUG [0m tests.store.test_time_filter:test_time_filter.py:189 number of messages stored for start time 1773664590720292096 and end time = 1773664594965656064 is 3
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:42494/store/v3/messages?includeData=True&pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&startTime=1773664590720292096&endTime=1773664594965656064&pageSize=20&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x350df42ae9c10cf8e856ae9fa68fa05627e6b6bd1bdaacdb6a7986c432bf9a8f","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773664590720292096,"ephemeral":false},"pubsubTopic":"/waku/2/rs/3/0"},{"messageHash":"0xe02faf5cafd15457a3b60b4620f608edfccc6b7488c8c8cfaa402271d8d87eeb","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773664592720299008,"ephemeral":false},"pubsubTopic":"/waku/2/rs/3/0"},{"messageHash":"0x31a2e5fad52f35f80434bf66f9b4e27af767c303218278670763d343ada4b6e7","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773664593620301056,"ephemeral":false},"pubsubTopic":"/waku/2/rs/3/0"}]}'
|
||
[35mDEBUG [0m tests.store.test_time_filter:test_time_filter.py:189 number of messages stored for start time 1773664590720292096 and end time = 1773664594965656064 is 3
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id b316d198ee4b
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 511da105d6b7
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |