mirror of
https://github.com/logos-messaging/docs.waku.org.git
synced 2026-01-02 12:53:12 +00:00
script revert research changes
This commit is contained in:
parent
a98447e81a
commit
236a915ffe
@ -1,7 +1,6 @@
|
||||
---
|
||||
title: PostgreSQL
|
||||
description: Document that describes why Nim-Waku started to use Postgres and shows some benchmark and comparison results.
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
## Introduction
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
---
|
||||
title: Performance Benchmarks and Test Reports
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
---
|
||||
title: Capped Bandwidth in Waku
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
This post explains i) why The Waku Network requires a capped bandwidth per shard and ii) how to achieve it by rate limiting with RLN v2.
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
---
|
||||
title: Incentivisation
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
Waku is a family of decentralised communication protocols. The Waku Network (TWN) consists of independent nodes running Waku protocols. TWN needs incentivisation (shortened to i13n) to ensure proper node behaviour.
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
---
|
||||
title: Maximum Bandwidth for Global Adoption
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
**TLDR**: This issue aims to **set the maximum bandwidth** in `x Mbps` that each waku shard should consume so that the **maximum amount of people can run a full waku node**. It is up to https://github.com/waku-org/research/issues/22 to specify how this maximum will be enforced.
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
---
|
||||
title: Message Propagation Times With Waku-RLN
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
**TLDR**: We present the results of 1000 `nwaku` nodes running `rln` using different message sizes, in a real network with bandwidth limitations and network delays. The goal is to study the message propagation delay distribution, and how it's affected by i) rln and ii) message size in a real environment. We observe that for messages of `10kB` the average end-to-end propagation delay is `508 ms`. We can also observe that the message propagation delays are severely affected when increasing the message size, which indicates that it is not a good idea to use waku for messages of eg. `500kB`. See simulation parameters.
|
||||
|
||||
@ -1,19 +1,16 @@
|
||||
---
|
||||
title: RLN Key Benchmarks
|
||||
displayed_sidebar: research
|
||||
---
|
||||
|
||||
## Introduction
|
||||
|
||||
Since RLN has been chosen as the spamming protection mechanism for waku, we must understand the practical implications of using it. This issue explains the main differences between `relay` and `rln-relay` and gives some benchmarks after running simulations using `waku-simulator`, in a network with the following characteristics:
|
||||
|
||||
- 100 nwaku nodes, each one with a valid rln membership and publishing a message every 10 seconds to a common topic.
|
||||
- rln contract deployed in Ethereum Sepolia
|
||||
- 10.000 memberships registered in the contract
|
||||
- pure relay (store and light protocols disabled)
|
||||
|
||||
The main deltas `rln` vs `rln-relay` are:
|
||||
|
||||
- New `proof ` field in `WakuMessage` containing 384 extra bytes. This field must be generated and attached to each message.
|
||||
- New validator, that uses `proof` to `Accept` or `Reject` the message. The proof has to be verified.
|
||||
- New dependency on a blockchain, Ethereum, or any EVM chain, to keep track of the members allowed to publish.
|
||||
@ -21,7 +18,6 @@ The main deltas `rln` vs `rln-relay` are:
|
||||
But what are the practical implications of these?
|
||||
|
||||
## TLDR:
|
||||
|
||||
- Proof generation is constant-ish. 0.15 second for each proof
|
||||
- Proof verification is constant-ish, 0.012 seconds. In a network with 10k nodes and D=6 this would add an overhead delay of 0.06 seconds.
|
||||
- Gossipsub scoring drops connections from spammer peers, which acts as the punishment (instead of slashing). Validated in the simulation.
|
||||
@ -37,7 +33,7 @@ Seems that proof generation times stay constant no matter the size of the messag
|
||||
|
||||
On the other hand, rln also adds an overhead in the gossipsub validation process. On average it takes `0.012 seconds` to verify the proof. It seems that when we increase the message size, validation time seems to increase a bit, which can be for any other reason besides rln itself (eg deserializing the message might take longer).
|
||||
|
||||
This number seems reasonable and shouldn't affect that much the average delay of a message. Assuming a d-regular graph, with `10k` nodes and `D=6`, we can have up to `log(total_nodes)/log(D)=5` hops. So in the worst case, rln will add a network latency of `0.012*5 = 0.06 seconds`
|
||||
This number seems reasonable and shouldn't affect that much the average delay of a message. Assuming a d-regular graph, with `10k` nodes and `D=6`, we can have up to `log(total_nodes)/log(D)=5` hops. So in the worst case, rln will add a network latency of `0.012*5 = 0.06 seconds`
|
||||
|
||||

|
||||
|
||||
@ -52,7 +48,6 @@ In the following simulation, we can see `100` nwaku interconnected nodes, where
|
||||
## RLN tree sync
|
||||
|
||||
Using RLN implies that waku should now piggyback on a blockchain (the case study uses Ethereum Sepolia) and has to stay up to date with the latest events emitted by the rln smart contract. These events are used to locally construct a tree that contains all members allowed to create valid proofs to send messages. Some numbers:
|
||||
|
||||
- A tree with 10k members takes `2Mbytes` of space. Negligible.
|
||||
- A tree with 10k members takes `<4` minutes to synchronize. Assumable since it's done just once.
|
||||
- With a block range of 5000 blocks for each request, we would need `520 requests` to synchronize 1 year of historical data from the tree. Assumable since most of the free endpoints out there allow 100k/day.
|
||||
@ -60,7 +55,6 @@ Using RLN implies that waku should now piggyback on a blockchain (the case study
|
||||
## Performance relay vs. rln-relay
|
||||
|
||||
Same simulation with 100 nodes was executed `with rln` and `without rln`:
|
||||
|
||||
- Memory consumption is almost identical
|
||||
|
||||
**with rln**
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user