You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/research/research-and-studies/incentivisation.md
-1Lines changed: 0 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,5 @@
1
1
---
2
2
title: Incentivisation
3
-
displayed_sidebar: research
4
3
---
5
4
6
5
Waku is a family of decentralised communication protocols. The Waku Network (TWN) consists of independent nodes running Waku protocols. TWN needs incentivisation (shortened to i13n) to ensure proper node behaviour.
Copy file name to clipboardExpand all lines: docs/research/research-and-studies/maximum-bandwidth.md
-1Lines changed: 0 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,5 @@
1
1
---
2
2
title: Maximum Bandwidth for Global Adoption
3
-
displayed_sidebar: research
4
3
---
5
4
6
5
**TLDR**: This issue aims to **set the maximum bandwidth** in `x Mbps` that each waku shard should consume so that the **maximum amount of people can run a full waku node**. It is up to https://github.com/waku-org/research/issues/22 to specify how this maximum will be enforced.
Copy file name to clipboardExpand all lines: docs/research/research-and-studies/message-propagation.md
-1Lines changed: 0 additions & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,5 @@
1
1
---
2
2
title: Message Propagation Times With Waku-RLN
3
-
displayed_sidebar: research
4
3
---
5
4
6
5
**TLDR**: We present the results of 1000 `nwaku` nodes running `rln` using different message sizes, in a real network with bandwidth limitations and network delays. The goal is to study the message propagation delay distribution, and how it's affected by i) rln and ii) message size in a real environment. We observe that for messages of `10kB` the average end-to-end propagation delay is `508 ms`. We can also observe that the message propagation delays are severely affected when increasing the message size, which indicates that it is not a good idea to use waku for messages of eg. `500kB`. See simulation parameters.
Copy file name to clipboardExpand all lines: docs/research/research-and-studies/rln-key-benchmarks.md
+1-7Lines changed: 1 addition & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,27 +1,23 @@
1
1
---
2
2
title: RLN Key Benchmarks
3
-
displayed_sidebar: research
4
3
---
5
4
6
5
## Introduction
7
6
8
7
Since RLN has been chosen as the spamming protection mechanism for waku, we must understand the practical implications of using it. This issue explains the main differences between `relay` and `rln-relay` and gives some benchmarks after running simulations using `waku-simulator`, in a network with the following characteristics:
9
-
10
8
- 100 nwaku nodes, each one with a valid rln membership and publishing a message every 10 seconds to a common topic.
11
9
- rln contract deployed in Ethereum Sepolia
12
10
- 10.000 memberships registered in the contract
13
11
- pure relay (store and light protocols disabled)
14
12
15
13
The main deltas `rln` vs `rln-relay` are:
16
-
17
14
- New `proof ` field in `WakuMessage` containing 384 extra bytes. This field must be generated and attached to each message.
18
15
- New validator, that uses `proof` to `Accept` or `Reject` the message. The proof has to be verified.
19
16
- New dependency on a blockchain, Ethereum, or any EVM chain, to keep track of the members allowed to publish.
20
17
21
18
But what are the practical implications of these?
22
19
23
20
## TLDR:
24
-
25
21
- Proof generation is constant-ish. 0.15 second for each proof
26
22
- Proof verification is constant-ish, 0.012 seconds. In a network with 10k nodes and D=6 this would add an overhead delay of 0.06 seconds.
27
23
- Gossipsub scoring drops connections from spammer peers, which acts as the punishment (instead of slashing). Validated in the simulation.
@@ -37,7 +33,7 @@ Seems that proof generation times stay constant no matter the size of the messag
37
33
38
34
On the other hand, rln also adds an overhead in the gossipsub validation process. On average it takes `0.012 seconds` to verify the proof. It seems that when we increase the message size, validation time seems to increase a bit, which can be for any other reason besides rln itself (eg deserializing the message might take longer).
39
35
40
-
This number seems reasonable and shouldn't affect that much the average delay of a message. Assuming a d-regular graph, with `10k` nodes and `D=6`, we can have up to `log(total_nodes)/log(D)=5` hops. So in the worst case, rln will add a network latency of `0.012*5 = 0.06 seconds`
36
+
This number seems reasonable and shouldn't affect that much the average delay of a message. Assuming a d-regular graph, with `10k` nodes and `D=6`, we can have up to `log(total_nodes)/log(D)=5` hops. So in the worst case, rln will add a network latency of `0.012*5 = 0.06 seconds`
@@ -52,15 +48,13 @@ In the following simulation, we can see `100` nwaku interconnected nodes, where
52
48
## RLN tree sync
53
49
54
50
Using RLN implies that waku should now piggyback on a blockchain (the case study uses Ethereum Sepolia) and has to stay up to date with the latest events emitted by the rln smart contract. These events are used to locally construct a tree that contains all members allowed to create valid proofs to send messages. Some numbers:
55
-
56
51
- A tree with 10k members takes `2Mbytes` of space. Negligible.
57
52
- A tree with 10k members takes `<4` minutes to synchronize. Assumable since it's done just once.
58
53
- With a block range of 5000 blocks for each request, we would need `520 requests` to synchronize 1 year of historical data from the tree. Assumable since most of the free endpoints out there allow 100k/day.
59
54
60
55
## Performance relay vs. rln-relay
61
56
62
57
Same simulation with 100 nodes was executed `with rln` and `without rln`:
0 commit comments