mirror of
https://github.com/logos-messaging/logos-messaging-interop-tests.git
synced 2026-01-05 23:43:07 +00:00
96 lines
18 KiB
Plaintext
96 lines
18 KiB
Plaintext
[35mDEBUG [0m tests.conftest:conftest.py:51 Running fixture setup: test_id
|
||
[35mDEBUG [0m tests.conftest:conftest.py:57 Running test: test_invalid_pagination_cursor_param with id: 2025-12-10_04-16-22__c44a0b06-3d5b-4bdd-acd7-3ec285a511f5
|
||
[35mDEBUG [0m src.steps.common:common.py:19 Running fixture setup: common_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:31 Running fixture setup: store_setup
|
||
[35mDEBUG [0m src.steps.store:store.py:39 Running fixture setup: node_setup
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/publishing_node1_2025-12-10_04-16-22__c44a0b06-3d5b-4bdd-acd7-3ec285a511f5__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.139.128
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['15979', '15980', '15981', '15982', '15983']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 15979:15979 -p 15980:15980 -p 15981:15981 -p 15982:15982 -p 15983:15983 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=15981 --rest-port=15979 --tcp-port=15980 --discv5-udp-port=15982 --rest-address=0.0.0.0 --nat=extip:172.18.139.128 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=ee2a1f67b2fe3ba3f6e2c7fcc73b1fb0e9fafb22fa9cdbfaf30bbec4db64fcbb --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=15983 --metrics-logging=true --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.139.128 waku dd8a3beee9e66e954059b64ad077bc661eb86fa37afe895eaec8b7ca3059703c
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID dd8a3beee9e6. Setting up logs at ./log/docker/publishing_node1_2025-12-10_04-16-22__c44a0b06-3d5b-4bdd-acd7-3ec285a511f5__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 15979
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 854121c17ff5. Exiting log stream.
|
||
[31m[1mERROR [0m src.node.docker_mananger:docker_mananger.py:89 Max retries reached for container 9ad840a67d5a. Exiting log stream.
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:15979/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","protocolsHealth":[{"Relay":"NOT_READY","desc":"No connected peers"},{"Rln Relay":"NOT_MOUNTED"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"NOT_READY","desc":"No Legacy Store service peers are available yet, neither Store service set up for the node"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:15979/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.139.128/tcp/15980/p2p/16Uiu2HAmHBkK3NAkYTkXT11BYBFRvhFb12ie28Ejf6NkDyf4azL5","/ip4/172.18.139.128/tcp/15981/ws/p2p/16Uiu2HAmHBkK3NAkYTkXT11BYBFRvhFb12ie28Ejf6NkDyf4azL5"],"enrUri":"enr:-L24QE-Kerqnb3BHfTxPAVMvmnBrJ_PJs6jrTQV5f2YDDfg4DwbV_CzUB5WQnSF6_Sgc6C9-7e9roEqon_wN4mmynsUCgmlkgnY0gmlwhKwSi4CKbXVsdGlhZGRyc5YACASsEouABj5sAAoErBKLgAY-bd0DgnJzhQADAQAAiXNlY3AyNTZrMaEDQ0_hDRrXOnaUAsGCEd9qtKkRtDze5Ma-Rd61rSCFum6DdGNwgj5sg3VkcII-boV3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:19 Docker client initialized with image wakuorg/nwaku:latest
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:86 WakuNode instance initialized with log path ./log/docker/store_node1_2025-12-10_04-16-22__c44a0b06-3d5b-4bdd-acd7-3ec285a511f5__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:90 Starting Node...
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:22 Attempting to create or retrieve network waku
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:25 Network waku already exists
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:108 Generated random external IP 172.18.39.56
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:101 Generated ports ['61676', '61677', '61678', '61679', '61680']
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:439 RLN credentials were not set
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:176 RLN credentials not set or credential store not available, starting without RLN
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:178 Using volumes []
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:49 docker run -i -t -p 61676:61676 -p 61677:61677 -p 61678:61678 -p 61679:61679 -p 61680:61680 wakuorg/nwaku:latest --listen-address=0.0.0.0 --rest=true --rest-admin=true --websocket-support=true --log-level=TRACE --rest-relay-cache-capacity=100 --websocket-port=61678 --rest-port=61676 --tcp-port=61677 --discv5-udp-port=61679 --rest-address=0.0.0.0 --nat=extip:172.18.39.56 --peer-exchange=true --discv5-discovery=true --cluster-id=3 --nodekey=4f0c98df57c2348a39750f9f063cbdb35305ad83ee4a497ea50dbad0f3b6bbf0 --shard=0 --metrics-server=true --metrics-server-address=0.0.0.0 --metrics-server-port=61680 --metrics-logging=true --discv5-bootstrap-node=enr:-L24QE-Kerqnb3BHfTxPAVMvmnBrJ_PJs6jrTQV5f2YDDfg4DwbV_CzUB5WQnSF6_Sgc6C9-7e9roEqon_wN4mmynsUCgmlkgnY0gmlwhKwSi4CKbXVsdGlhZGRyc5YACASsEouABj5sAAoErBKLgAY-bd0DgnJzhQADAQAAiXNlY3AyNTZrMaEDQ0_hDRrXOnaUAsGCEd9qtKkRtDze5Ma-Rd61rSCFum6DdGNwgj5sg3VkcII-boV3YWt1MgM --storenode=/ip4/172.18.139.128/tcp/15980/p2p/16Uiu2HAmHBkK3NAkYTkXT11BYBFRvhFb12ie28Ejf6NkDyf4azL5 --store=true --relay=true
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:55 docker network connect --ip 172.18.39.56 waku 7315b3f64c1cdaf4a53e68ff52f9a23e19316a1d6762454a6fb734abf05a0714
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:58 Container started with ID 7315b3f64c1c. Setting up logs at ./log/docker/store_node1_2025-12-10_04-16-22__c44a0b06-3d5b-4bdd-acd7-3ec285a511f5__wakuorg_nwaku:latest.log
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:190 Started container from image wakuorg/nwaku:latest. REST: 61676
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 1 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:61676/health" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"nodeHealth":"READY","protocolsHealth":[{"Relay":"READY"},{"Rln Relay":"NOT_MOUNTED"},{"Lightpush":"NOT_MOUNTED"},{"Legacy Lightpush":"NOT_MOUNTED"},{"Filter":"NOT_MOUNTED"},{"Store":"READY"},{"Legacy Store":"NOT_MOUNTED"},{"Peer Exchange":"READY"},{"Rendezvous":"NOT_READY","desc":"No Rendezvous peers are available yet"},{"Mix":"NOT_MOUNTED"},{"Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Legacy Lightpush Client":"NOT_READY","desc":"No Lightpush service peer available yet"},{"Store Client":"READY"},{"Legacy Store Client":"READY"},{"Filter Client":"NOT_READY","desc":"No Filter service peer available yet"}]}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:287 Node protocols are initialized !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:61676/debug/v1/info" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"listenAddresses":["/ip4/172.18.39.56/tcp/61677/p2p/16Uiu2HAmVq972X99acLnqbn3ijFvFeX5KoBP9JiDGbkaDdLCH9zV","/ip4/172.18.39.56/tcp/61678/ws/p2p/16Uiu2HAmVq972X99acLnqbn3ijFvFeX5KoBP9JiDGbkaDdLCH9zV"],"enrUri":"enr:-L24QHkUVH04TqO8c6w5M_786c6RsLgETNos54oTQ7_2JTeYLYoYacXGRa-IfXvt49_HBYFCptehRKNijCuHZaD0R2cCgmlkgnY0gmlwhKwSJziKbXVsdGlhZGRyc5YACASsEic4BvDtAAoErBInOAbw7t0DgnJzhQADAQAAiXNlY3AyNTZrMaED_zBiVZQr3L8RysJPbRRMOpF1EQNdDX0y0_USvfEb4TaDdGNwgvDtg3VkcILw74V3YWt1MgM"}'
|
||
[32mINFO [0m src.node.waku_node:waku_node.py:292 REST service is ready !!
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:61676/admin/v1/peers" -H "Content-Type: application/json" -d '["/ip4/172.18.139.128/tcp/15980/p2p/16Uiu2HAmHBkK3NAkYTkXT11BYBFRvhFb12ie28Ejf6NkDyf4azL5"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:15979/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:61676/relay/v1/subscriptions" -H "Content-Type: application/json" -d '["/waku/2/rs/3/0"]'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:15979/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "TWVzc2FnZV8w", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:15979/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "TWVzc2FnZV8x", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:15979/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "TWVzc2FnZV8y", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:15979/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "TWVzc2FnZV8z", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[35mDEBUG [0m src.steps.store:store.py:132 Relaying message
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X POST "http://127.0.0.1:15979/relay/v1/messages/%2Fwaku%2F2%2Frs%2F3%2F0" -H "Content-Type: application/json" -d '{"payload": "TWVzc2FnZV80", "contentTopic": "/myapp/1/latest/proto", "timestamp": '$(date +%s%N)'}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'OK'
|
||
[35mDEBUG [0m src.libs.common:common.py:47 Sleeping for 0.2 seconds
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:15979/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&pageSize=3&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x46b26ba5fdd628d1b08662586932760301e89cf13cb43f481e17c6c7dfccca9e"},{"messageHash":"0x0a98f227ff8b4d0c1d7a3377a124aa74bf03ad2607a6765abbfdeb8423c7d0ef"},{"messageHash":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}],"paginationCursor":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:15979/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&pageSize=3&ascending=true&paginationCursor=0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x46b26ba5fdd628d1b08662586932760301e89cf13cb43f481e17c6c7dfccca9e"},{"messageHash":"0x0a98f227ff8b4d0c1d7a3377a124aa74bf03ad2607a6765abbfdeb8423c7d0ef"},{"messageHash":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}],"paginationCursor":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:15979/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&cursor=0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0&pageSize=3&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x00d49984adfa6371a15bfa8e560b6dd9d541ef204da014ce5dacc8d4a8151a34"},{"messageHash":"0xc63f1186a87d7fb66bc1034cc9a470cd8961cd2efdf644ed80e6fdc895c91894"}]}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:61676/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&pageSize=3&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x46b26ba5fdd628d1b08662586932760301e89cf13cb43f481e17c6c7dfccca9e"},{"messageHash":"0x0a98f227ff8b4d0c1d7a3377a124aa74bf03ad2607a6765abbfdeb8423c7d0ef"},{"messageHash":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}],"paginationCursor":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:61676/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&pageSize=3&ascending=true&paginationCursor=0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x46b26ba5fdd628d1b08662586932760301e89cf13cb43f481e17c6c7dfccca9e"},{"messageHash":"0x0a98f227ff8b4d0c1d7a3377a124aa74bf03ad2607a6765abbfdeb8423c7d0ef"},{"messageHash":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}],"paginationCursor":"0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0"}'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:37 curl -v -X GET "http://127.0.0.1:61676/store/v3/messages?pubsubTopic=%2Fwaku%2F2%2Frs%2F3%2F0&cursor=0xedb83c85cda87bdca3b055096c943f94ba33aedc6c3a3bc07651207ea3aecac0&pageSize=3&ascending=true" -H "Content-Type: application/json" -d 'None'
|
||
[32mINFO [0m src.node.api_clients.base_client:base_client.py:22 Response status code: 200. Response content: b'{"requestId":"","statusCode":200,"statusDesc":"OK","messages":[{"messageHash":"0x00d49984adfa6371a15bfa8e560b6dd9d541ef204da014ce5dacc8d4a8151a34"},{"messageHash":"0xc63f1186a87d7fb66bc1034cc9a470cd8961cd2efdf644ed80e6fdc895c91894"}]}'
|
||
[35mDEBUG [0m tests.conftest:conftest.py:59 Running fixture teardown: test_setup
|
||
[35mDEBUG [0m tests.conftest:conftest.py:83 Running fixture teardown: close_open_nodes
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id dd8a3beee9e6
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:234 Stopping container with id 7315b3f64c1c
|
||
[35mDEBUG [0m src.node.waku_node:waku_node.py:241 Container stopped.
|
||
[35mDEBUG [0m tests.conftest:conftest.py:98 Running fixture teardown: check_waku_log_errors
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs.
|
||
[35mDEBUG [0m src.node.docker_mananger:docker_mananger.py:144 No errors found in the waku logs. |