mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-05-05 10:03:16 +00:00
95 lines
18 KiB
Plaintext
95 lines
18 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_time_filter_end_time_now with id: 2026-03-18_09-08-32__a0ffd45d-f28b-4aba-b440-58e65203c6b3
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:39 Running fixture setup: node_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/publishing_node1_2026-03-18_09-08-32__a0ffd45d-f28b-4aba-b440-58e65203c6b3__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.15.38
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['7389', '7390', '7391', '7392', '7393']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 7389:7389 -p 7390:7390 -p 7391:7391 -p 7392:7392 -p 7393:7393 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=7391 --rest-port=7389 --tcp-port=7390 --discv5-udp-port=7392 --rest-address=0.0.0.0 --nat=extip:172.18.15.38 --peer-exchange=true --discv5-discovery=true --cluster-id=198 --nodekey=f8c2a6cc5592ccb78ce0dc6baef9bed4d48f8dad33f6213becdb7cb635b5bbea --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=7393 --metrics-logging=true --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.15.38 waku d20275177f7c8eb3da50aaff85ed9dc6e95b4f4a67b109ad28545b17e58e23c2
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 5753b48378de. Exiting log stream.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID d20275177f7c. Setting up logs at ./log/docker/publishing_node1_2026-03-18_09-08-32__a0ffd45d-f28b-4aba-b440-58e65203c6b3__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 7389
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 3ce76f8fb40e. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7389/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"Disconnected","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7389/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.15.38/tcp/7390/p2p/16Uiu2HAkuhii2PhB9vH4fjK84dZdFvFdzskcCxFqMKvnbry5S7nq","/ip4/172.18.15.38/tcp/7391/ws/p2p/16Uiu2HAkuhii2PhB9vH4fjK84dZdFvFdzskcCxFqMKvnbry5S7nq"],"enrUri":"enr:-L24QLuPU-kVxAE5poEildD0AhX1GMuz5o5hB-KTMjfb6K6pSFVbtcxIVea2snbeBkNxp7BwpVRyQSYvuN3T4K5jHtwCgmlkgnY0gmlwhKwSDyaKbXVsdGlhZGRyc5YACASsEg8mBhzeAAoErBIPJgYc390DgnJzhQDGAQAAiXNlY3AyNTZrMaECBBxgrXP8OMtCrd1UJF3o9QiZ3FWkt18yKV-fokYzZKKDdGNwghzeg3VkcIIc4IV3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/store_node1_2026-03-18_09-08-32__a0ffd45d-f28b-4aba-b440-58e65203c6b3__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.204.66
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['22007', '22008', '22009', '22010', '22011']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 22007:22007 -p 22008:22008 -p 22009:22009 -p 22010:22010 -p 22011:22011 harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0 --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=22009 --rest-port=22007 --tcp-port=22008 --discv5-udp-port=22010 --rest-address=0.0.0.0 --nat=extip:172.18.204.66 --peer-exchange=true --discv5-discovery=true --cluster-id=198 --nodekey=0161d66f3047f1d56edefbdaceb341d1ef08743fcefef02ef05bce85e79ae31d --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=22011 --metrics-logging=true --discv5-bootstrap-node=enr:-L24QLuPU-kVxAE5poEildD0AhX1GMuz5o5hB-KTMjfb6K6pSFVbtcxIVea2snbeBkNxp7BwpVRyQSYvuN3T4K5jHtwCgmlkgnY0gmlwhKwSDyaKbXVsdGlhZGRyc5YACASsEg8mBhzeAAoErBIPJgYc390DgnJzhQDGAQAAiXNlY3AyNTZrMaECBBxgrXP8OMtCrd1UJF3o9QiZ3FWkt18yKV-fokYzZKKDdGNwghzeg3VkcIIc4IV3YWt1MgM --storenode=/ip4/172.18.15.38/tcp/7390/p2p/16Uiu2HAkuhii2PhB9vH4fjK84dZdFvFdzskcCxFqMKvnbry5S7nq --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.204.66 waku f1b92f236bf39e9deed9b503716e92b6265fd100c3eedef59a8447bdbb847cd6
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID f1b92f236bf3. Setting up logs at ./log/docker/store_node1_2026-03-18_09-08-32__a0ffd45d-f28b-4aba-b440-58e65203c6b3__harbor.status.im_wakuorg_nwaku:v0.38.0-rc.0.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image harbor.status.im/wakuorg/nwaku:v0.38.0-rc.0. REST: 22007
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:22007/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","connectionStatus":"PartiallyConnected","protocolsHealth":[{"Relay":"READY"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"},{"Rln Relay":"NOT_MOUNTED"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:22007/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.204.66/tcp/22008/p2p/16Uiu2HAmGQHpLTombo9t86hn7GfoHDmVJQ1Yx2rrN8fqEtFZPWqD","/ip4/172.18.204.66/tcp/22009/ws/p2p/16Uiu2HAmGQHpLTombo9t86hn7GfoHDmVJQ1Yx2rrN8fqEtFZPWqD"],"enrUri":"enr:-L24QGtA2KO6NatUIpE8ukyMAeQLjJaRA2NcWJc9LzFZ9UAtZ9BJcuOC3io3hAG5Yzz34-407J_bTkUUxlUnlFKLqPsCgmlkgnY0gmlwhKwSzEKKbXVsdGlhZGRyc5YACASsEsxCBlX4AAoErBLMQgZV-d0DgnJzhQDGAQAAiXNlY3AyNTZrMaEDN6rI6JZucWDXMIRMw4xOZIAsIHbnwObxzesP8kMoX3CDdGNwglX4g3VkcIJV-oV3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:22007/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.15.38/tcp/7390/p2p/16Uiu2HAkuhii2PhB9vH4fjK84dZdFvFdzskcCxFqMKvnbry5S7nq"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/198/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:22007/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/198/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/messages/%2Fwaku%2F2%2Frs%2F198%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/messages/%2Fwaku%2F2%2Frs%2F198%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/messages/%2Fwaku%2F2%2Frs%2F198%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/messages/%2Fwaku%2F2%2Frs%2F198%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/messages/%2Fwaku%2F2%2Frs%2F198%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:7389/relay/v1/messages/%2Fwaku%2F2%2Frs%2F198%2F0" -H "Content-Type: application/json" -d '{"payload": "U3RvcmUgd29ya3MhIQ==", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m tests.store.test_time_filter:test_time_filter.py:186 inquering stored messages with start time 1773824912181987072 after end time 1773824916426477824
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:7389/store/v3/messages?includeData=True&pubsubTopic=%2Fwaku%2F2%2Frs%2F198%2F0&startTime=1773824912181987072&endTime=1773824916426477824&pageSize=20&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x4262f505028ba61b41657ee5f9a24889b5c364d6bbc3a012a10f9c3ab5d07a9f","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773824912181987072,"ephemeral":false},"pubsubTopic":"/waku/2/rs/198/0"},{"messageHash":"0x239fdf80879ea02ab657c9fe8028dcdd2c1244bcb78c2ccf5969c2398b6f5ad6","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773824914181993984,"ephemeral":false},"pubsubTopic":"/waku/2/rs/198/0"},{"messageHash":"0xc03ddeb89cd6443aa2c5e4ff760d3726ca0332ef54d68a7d2c4afb94e92f16ea","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773824915081996032,"ephemeral":false},"pubsubTopic":"/waku/2/rs/198/0"}]}'
|
||
[35mDEBUG [0m tests.store.test_time_filter:test_time_filter.py:189 number of messages stored for start time 1773824912181987072 and end time = 1773824916426477824 is 3
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:22007/store/v3/messages?includeData=True&pubsubTopic=%2Fwaku%2F2%2Frs%2F198%2F0&startTime=1773824912181987072&endTime=1773824916426477824&pageSize=20&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x4262f505028ba61b41657ee5f9a24889b5c364d6bbc3a012a10f9c3ab5d07a9f","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773824912181987072,"ephemeral":false},"pubsubTopic":"/waku/2/rs/198/0"},{"messageHash":"0x239fdf80879ea02ab657c9fe8028dcdd2c1244bcb78c2ccf5969c2398b6f5ad6","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773824914181993984,"ephemeral":false},"pubsubTopic":"/waku/2/rs/198/0"},{"messageHash":"0xc03ddeb89cd6443aa2c5e4ff760d3726ca0332ef54d68a7d2c4afb94e92f16ea","message":{"payload":"U3RvcmUgd29ya3MhIQ==","contentTopic":"/myapp/1/latest/proto","version":0,"timestamp":1773824915081996032,"ephemeral":false},"pubsubTopic":"/waku/2/rs/198/0"}]}'
|
||
[35mDEBUG [0m tests.store.test_time_filter:test_time_filter.py:189 number of messages stored for start time 1773824912181987072 and end time = 1773824916426477824 is 3
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id d20275177f7c
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id f1b92f236bf3
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |