2021-04-28 14:20:05 +00:00
|
|
|
# nim-eth
|
2023-05-10 13:50:04 +00:00
|
|
|
# Copyright (c) 2018-2023 Status Research & Development GmbH
|
2021-04-28 14:20:05 +00:00
|
|
|
# Licensed and distributed under either of
|
|
|
|
# * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT).
|
|
|
|
# * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0).
|
|
|
|
# at your option. This file may not be copied, modified, or distributed except according to those terms.
|
|
|
|
|
2023-05-10 13:50:04 +00:00
|
|
|
{.push raises: [].}
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
import
|
2022-09-19 08:47:30 +00:00
|
|
|
std/[tables, hashes, times, algorithm, sets, sequtils],
|
2023-03-16 15:45:12 +00:00
|
|
|
chronos, chronicles, stint, nimcrypto/keccak, metrics,
|
2022-09-19 08:47:30 +00:00
|
|
|
../keys, ./discoveryv5/random2,
|
2021-04-06 11:33:24 +00:00
|
|
|
./enode
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
export sets # TODO: This should not be needed, but compilation fails otherwise
|
|
|
|
|
2023-10-31 17:28:21 +00:00
|
|
|
declareGauge discv4_routing_table_nodes,
|
|
|
|
"Discovery v4 routing table nodes"
|
2023-03-16 15:45:12 +00:00
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
logScope:
|
2022-12-06 13:54:03 +00:00
|
|
|
topics = "eth p2p kademlia"
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
type
|
2022-09-06 15:19:47 +00:00
|
|
|
# 32 bytes NodeId | 16 bytes ip | 1 byte mode
|
|
|
|
TimeKey = array[49, byte]
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
KademliaProtocol* [Wire] = ref object
|
|
|
|
wire: Wire
|
|
|
|
thisNode: Node
|
|
|
|
routing: RoutingTable
|
|
|
|
pongFutures: Table[seq[byte], Future[bool]]
|
|
|
|
pingFutures: Table[Node, Future[bool]]
|
2023-05-10 13:50:04 +00:00
|
|
|
neighboursCallbacks: Table[Node, proc(n: seq[Node]) {.gcsafe, raises: [].}]
|
2022-06-17 20:45:37 +00:00
|
|
|
rng: ref HmacDrbgContext
|
2022-09-06 15:19:47 +00:00
|
|
|
pingPongTime: OrderedTable[TimeKey, int64] # int64 -> unix time
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
NodeId* = UInt256
|
|
|
|
|
|
|
|
Node* = ref object
|
|
|
|
node*: ENode
|
|
|
|
id*: NodeId
|
|
|
|
|
|
|
|
RoutingTable = object
|
|
|
|
thisNode: Node
|
|
|
|
buckets: seq[KBucket]
|
|
|
|
|
|
|
|
KBucket = ref object
|
|
|
|
istart, iend: UInt256
|
|
|
|
nodes: seq[Node]
|
|
|
|
replacementCache: seq[Node]
|
|
|
|
lastUpdated: float # epochTime
|
|
|
|
|
2022-09-06 15:19:47 +00:00
|
|
|
CommandId* = enum
|
|
|
|
cmdPing = 1
|
|
|
|
cmdPong = 2
|
|
|
|
cmdFindNode = 3
|
|
|
|
cmdNeighbours = 4
|
|
|
|
cmdENRRequest = 5
|
|
|
|
cmdENRResponse = 6
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
const
|
|
|
|
BUCKET_SIZE = 16
|
|
|
|
BITS_PER_HOP = 8
|
2020-02-22 18:49:14 +00:00
|
|
|
REQUEST_TIMEOUT = chronos.milliseconds(5000) # timeout of message round trips
|
2019-02-05 15:40:29 +00:00
|
|
|
FIND_CONCURRENCY = 3 # parallel find node lookups
|
|
|
|
ID_SIZE = 256
|
2022-09-06 15:19:47 +00:00
|
|
|
BOND_EXPIRATION = initDuration(hours = 12)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2022-09-07 14:12:32 +00:00
|
|
|
proc len(r: RoutingTable): int
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
proc toNodeId*(pk: PublicKey): NodeId =
|
2020-04-04 16:44:01 +00:00
|
|
|
readUintBE[256](keccak256.digest(pk.toRaw()).data)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc newNode*(pk: PublicKey, address: Address): Node =
|
|
|
|
result.new()
|
2020-04-06 16:24:15 +00:00
|
|
|
result.node = ENode(pubkey: pk, address: address)
|
2019-02-05 15:40:29 +00:00
|
|
|
result.id = pk.toNodeId()
|
|
|
|
|
|
|
|
proc newNode*(uriString: string): Node =
|
|
|
|
result.new()
|
2020-04-06 16:24:15 +00:00
|
|
|
result.node = ENode.fromString(uriString)[]
|
2019-02-05 15:40:29 +00:00
|
|
|
result.id = result.node.pubkey.toNodeId()
|
|
|
|
|
|
|
|
proc newNode*(enode: ENode): Node =
|
|
|
|
result.new()
|
|
|
|
result.node = enode
|
|
|
|
result.id = result.node.pubkey.toNodeId()
|
|
|
|
|
|
|
|
proc distanceTo(n: Node, id: NodeId): UInt256 = n.id xor id
|
|
|
|
|
|
|
|
proc `$`*(n: Node): string =
|
|
|
|
if n == nil:
|
|
|
|
"Node[local]"
|
|
|
|
else:
|
|
|
|
"Node[" & $n.node.address.ip & ":" & $n.node.address.udpPort & "]"
|
|
|
|
|
2023-04-18 11:51:02 +00:00
|
|
|
chronicles.formatIt(Node): $it
|
|
|
|
chronicles.formatIt(seq[Node]): $it
|
|
|
|
|
2020-04-04 16:44:01 +00:00
|
|
|
proc hash*(n: Node): hashes.Hash = hash(n.node.pubkey.toRaw)
|
2021-04-27 09:29:54 +00:00
|
|
|
proc `==`*(a, b: Node): bool = (a.isNil and b.isNil) or
|
|
|
|
(not a.isNil and not b.isNil and a.node.pubkey == b.node.pubkey)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2022-09-06 15:19:47 +00:00
|
|
|
proc timeKey(id: NodeId, ip: IpAddress, cmd: CommandId): TimeKey =
|
|
|
|
result[0..31] = id.toByteArrayBE()[0..31]
|
|
|
|
case ip.family
|
|
|
|
of IpAddressFamily.IPv6:
|
|
|
|
result[32..47] = ip.address_v6[0..15]
|
|
|
|
of IpAddressFamily.IPv4:
|
|
|
|
result[32..35] = ip.address_v4[0..3]
|
|
|
|
result[48] = cmd.byte
|
|
|
|
|
|
|
|
proc ip(n: Node): IpAddress =
|
|
|
|
n.node.address.ip
|
|
|
|
|
|
|
|
proc timeKeyPong(n: Node): TimeKey =
|
|
|
|
timeKey(n.id, n.ip, cmdPong)
|
|
|
|
|
|
|
|
proc timeKeyPing(n: Node): TimeKey =
|
|
|
|
timeKey(n.id, n.ip, cmdPing)
|
|
|
|
|
|
|
|
proc lastPingReceived(k: KademliaProtocol, n: Node): Time =
|
|
|
|
k.pingPongTime.getOrDefault(n.timeKeyPing, 0'i64).fromUnix
|
|
|
|
|
|
|
|
proc lastPongReceived(k: KademliaProtocol, n: Node): Time =
|
|
|
|
k.pingPongTime.getOrDefault(n.timeKeyPong, 0'i64).fromUnix
|
|
|
|
|
|
|
|
proc cmp(x, y: (TimeKey, int64)): int =
|
|
|
|
if x[1] < y[1]: return -1
|
|
|
|
if x[1] > y[1]: return 1
|
|
|
|
0
|
|
|
|
|
|
|
|
proc removeTooOldPingPongTime(k: KademliaProtocol) =
|
|
|
|
const
|
2022-09-07 14:12:32 +00:00
|
|
|
MinEntries = 128
|
|
|
|
MaxRC = MinEntries div 8
|
|
|
|
|
|
|
|
# instead of using fixed limit, we use dynamic limit
|
|
|
|
# with minimum entries = 128.
|
|
|
|
# remove 25% of too old entries if we need more space.
|
|
|
|
# the reason maxEntries is twice routing table because we
|
|
|
|
# store ping and pong time.
|
|
|
|
let
|
|
|
|
maxEntries = max(k.routing.len * 2, MinEntries)
|
|
|
|
maxRemove = maxEntries div 4
|
|
|
|
|
|
|
|
if k.pingPongTime.len < maxEntries:
|
2022-09-06 15:19:47 +00:00
|
|
|
return
|
|
|
|
|
2022-09-07 14:12:32 +00:00
|
|
|
# it is safe to remove this table sort?
|
|
|
|
# because we already using ordered table to store time from
|
|
|
|
# older value to newer value
|
|
|
|
when false:
|
|
|
|
k.pingPongTime.sort(cmp, order = SortOrder.Descending)
|
2022-09-06 15:19:47 +00:00
|
|
|
|
2022-09-07 14:12:32 +00:00
|
|
|
var
|
|
|
|
rci = 0
|
|
|
|
numRemoved = 0
|
|
|
|
rc: array[MaxRC, TimeKey] # 784 bytes(MinEntries/8*sizeof(TimeKey))
|
|
|
|
|
|
|
|
# using fixed size temp on stack possibly
|
|
|
|
# requires multiple iteration to remove
|
|
|
|
# old entries
|
|
|
|
while numRemoved < maxRemove:
|
|
|
|
for v in keys(k.pingPongTime):
|
|
|
|
rc[rci] = v
|
|
|
|
inc rci
|
|
|
|
inc numRemoved
|
|
|
|
if rci >= MaxRC or numRemoved >= maxRemove: break
|
|
|
|
|
|
|
|
for i in 0..<rci:
|
|
|
|
k.pingPongTime.del(rc[i])
|
|
|
|
|
|
|
|
rci = 0
|
2022-09-06 15:19:47 +00:00
|
|
|
|
|
|
|
proc updateLastPingReceived(k: KademliaProtocol, n: Node, t: Time) =
|
|
|
|
k.removeTooOldPingPongTime()
|
|
|
|
k.pingPongTime[n.timeKeyPing] = t.toUnix
|
|
|
|
|
|
|
|
proc updateLastPongReceived(k: KademliaProtocol, n: Node, t: Time) =
|
|
|
|
k.removeTooOldPingPongTime()
|
|
|
|
k.pingPongTime[n.timeKeyPong] = t.toUnix
|
|
|
|
|
|
|
|
# checkBond checks if the given node has a recent enough endpoint proof.
|
|
|
|
proc checkBond(k: KademliaProtocol, n: Node): bool =
|
|
|
|
getTime() - k.lastPongReceived(n) < BOND_EXPIRATION
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
proc newKBucket(istart, iend: NodeId): KBucket =
|
|
|
|
result.new()
|
|
|
|
result.istart = istart
|
|
|
|
result.iend = iend
|
|
|
|
result.nodes = @[]
|
|
|
|
result.replacementCache = @[]
|
|
|
|
|
|
|
|
proc midpoint(k: KBucket): NodeId =
|
|
|
|
k.istart + (k.iend - k.istart) div 2.u256
|
|
|
|
|
|
|
|
proc distanceTo(k: KBucket, id: NodeId): UInt256 = k.midpoint xor id
|
|
|
|
proc nodesByDistanceTo(k: KBucket, id: NodeId): seq[Node] =
|
|
|
|
sortedByIt(k.nodes, it.distanceTo(id))
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc len(k: KBucket): int = k.nodes.len
|
|
|
|
proc head(k: KBucket): Node = k.nodes[0]
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc add(k: KBucket, n: Node): Node =
|
|
|
|
## Try to add the given node to this bucket.
|
|
|
|
|
|
|
|
## If the node is already present, it is moved to the tail of the list, and we return None.
|
|
|
|
|
|
|
|
## If the node is not already present and the bucket has fewer than k entries, it is inserted
|
|
|
|
## at the tail of the list, and we return None.
|
|
|
|
|
|
|
|
## If the bucket is full, we add the node to the bucket's replacement cache and return the
|
|
|
|
## node at the head of the list (i.e. the least recently seen), which should be evicted if it
|
|
|
|
## fails to respond to a ping.
|
|
|
|
k.lastUpdated = epochTime()
|
|
|
|
let nodeIdx = k.nodes.find(n)
|
|
|
|
if nodeIdx != -1:
|
|
|
|
k.nodes.delete(nodeIdx)
|
|
|
|
k.nodes.add(n)
|
|
|
|
elif k.len < BUCKET_SIZE:
|
|
|
|
k.nodes.add(n)
|
2023-10-31 17:28:21 +00:00
|
|
|
discv4_routing_table_nodes.inc()
|
2019-02-05 15:40:29 +00:00
|
|
|
else:
|
|
|
|
k.replacementCache.add(n)
|
|
|
|
return k.head
|
|
|
|
return nil
|
|
|
|
|
|
|
|
proc removeNode(k: KBucket, n: Node) =
|
|
|
|
let i = k.nodes.find(n)
|
2023-03-16 15:45:12 +00:00
|
|
|
if i != -1:
|
2023-10-31 17:28:21 +00:00
|
|
|
discv4_routing_table_nodes.dec()
|
2023-03-16 15:45:12 +00:00
|
|
|
k.nodes.delete(i)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc split(k: KBucket): tuple[lower, upper: KBucket] =
|
|
|
|
## Split at the median id
|
|
|
|
let splitid = k.midpoint
|
|
|
|
result.lower = newKBucket(k.istart, splitid)
|
|
|
|
result.upper = newKBucket(splitid + 1.u256, k.iend)
|
|
|
|
for node in k.nodes:
|
|
|
|
let bucket = if node.id <= splitid: result.lower else: result.upper
|
|
|
|
discard bucket.add(node)
|
|
|
|
for node in k.replacementCache:
|
|
|
|
let bucket = if node.id <= splitid: result.lower else: result.upper
|
|
|
|
bucket.replacementCache.add(node)
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc inRange(k: KBucket, n: Node): bool =
|
2019-02-05 15:40:29 +00:00
|
|
|
k.istart <= n.id and n.id <= k.iend
|
|
|
|
|
|
|
|
proc isFull(k: KBucket): bool = k.len == BUCKET_SIZE
|
|
|
|
|
|
|
|
proc contains(k: KBucket, n: Node): bool = n in k.nodes
|
|
|
|
|
2021-12-20 12:14:50 +00:00
|
|
|
proc binaryGetBucketForNode(buckets: openArray[KBucket], n: Node):
|
2023-05-10 13:50:04 +00:00
|
|
|
KBucket {.raises: [ValueError].} =
|
2019-02-05 15:40:29 +00:00
|
|
|
## Given a list of ordered buckets, returns the bucket for a given node.
|
|
|
|
let bucketPos = lowerBound(buckets, n.id) do(a: KBucket, b: NodeId) -> int:
|
|
|
|
cmp(a.iend, b)
|
|
|
|
# Prevents edge cases where bisect_left returns an out of range index
|
|
|
|
if bucketPos < buckets.len:
|
|
|
|
let bucket = buckets[bucketPos]
|
|
|
|
if bucket.istart <= n.id and n.id <= bucket.iend:
|
|
|
|
result = bucket
|
|
|
|
|
|
|
|
if result.isNil:
|
|
|
|
raise newException(ValueError, "No bucket found for node with id " & $n.id)
|
|
|
|
|
2021-12-20 12:14:50 +00:00
|
|
|
proc computeSharedPrefixBits(nodes: openArray[Node]): int =
|
2019-02-05 15:40:29 +00:00
|
|
|
## Count the number of prefix bits shared by all nodes.
|
|
|
|
if nodes.len < 2:
|
|
|
|
return ID_SIZE
|
|
|
|
|
|
|
|
var mask = zero(UInt256)
|
|
|
|
let one = one(UInt256)
|
|
|
|
|
|
|
|
for i in 1 .. ID_SIZE:
|
|
|
|
mask = mask or (one shl (ID_SIZE - i))
|
|
|
|
let reference = nodes[0].id and mask
|
|
|
|
for j in 1 .. nodes.high:
|
|
|
|
if (nodes[j].id and mask) != reference: return i - 1
|
|
|
|
|
2019-03-13 22:15:26 +00:00
|
|
|
doAssert(false, "Unable to calculate number of shared prefix bits")
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc init(r: var RoutingTable, thisNode: Node) =
|
2019-02-05 15:40:29 +00:00
|
|
|
r.thisNode = thisNode
|
2021-12-20 12:14:50 +00:00
|
|
|
r.buckets = @[newKBucket(0.u256, high(UInt256))]
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc splitBucket(r: var RoutingTable, index: int) =
|
|
|
|
let bucket = r.buckets[index]
|
|
|
|
let (a, b) = bucket.split()
|
|
|
|
r.buckets[index] = a
|
|
|
|
r.buckets.insert(b, index + 1)
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc bucketForNode(r: RoutingTable, n: Node): KBucket
|
2023-05-10 13:50:04 +00:00
|
|
|
{.raises: [ValueError].} =
|
2019-02-05 15:40:29 +00:00
|
|
|
binaryGetBucketForNode(r.buckets, n)
|
|
|
|
|
2023-05-10 13:50:04 +00:00
|
|
|
proc removeNode(r: var RoutingTable, n: Node) {.raises: [ValueError].} =
|
2019-02-05 15:40:29 +00:00
|
|
|
r.bucketForNode(n).removeNode(n)
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc addNode(r: var RoutingTable, n: Node): Node
|
2023-05-10 13:50:04 +00:00
|
|
|
{.raises: [ValueError].} =
|
2019-04-24 23:17:11 +00:00
|
|
|
if n == r.thisNode:
|
|
|
|
warn "Trying to add ourselves to the routing table", node = n
|
|
|
|
return
|
2019-02-05 15:40:29 +00:00
|
|
|
let bucket = r.bucketForNode(n)
|
|
|
|
let evictionCandidate = bucket.add(n)
|
|
|
|
if not evictionCandidate.isNil:
|
|
|
|
# Split if the bucket has the local node in its range or if the depth is not congruent
|
|
|
|
# to 0 mod BITS_PER_HOP
|
|
|
|
|
|
|
|
let depth = computeSharedPrefixBits(bucket.nodes)
|
|
|
|
if bucket.inRange(r.thisNode) or (depth mod BITS_PER_HOP != 0 and depth != ID_SIZE):
|
|
|
|
r.splitBucket(r.buckets.find(bucket))
|
|
|
|
return r.addNode(n) # retry
|
|
|
|
|
|
|
|
# Nothing added, ping evictionCandidate
|
|
|
|
return evictionCandidate
|
|
|
|
|
2023-05-10 13:50:04 +00:00
|
|
|
proc contains(r: RoutingTable, n: Node): bool {.raises: [ValueError].} =
|
2021-04-27 09:29:54 +00:00
|
|
|
n in r.bucketForNode(n)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc bucketsByDistanceTo(r: RoutingTable, id: NodeId): seq[KBucket] =
|
|
|
|
sortedByIt(r.buckets, it.distanceTo(id))
|
|
|
|
|
|
|
|
proc notFullBuckets(r: RoutingTable): seq[KBucket] =
|
|
|
|
r.buckets.filterIt(not it.isFull)
|
|
|
|
|
|
|
|
proc neighbours(r: RoutingTable, id: NodeId, k: int = BUCKET_SIZE): seq[Node] =
|
|
|
|
## Return up to k neighbours of the given node.
|
|
|
|
result = newSeqOfCap[Node](k * 2)
|
|
|
|
for bucket in r.bucketsByDistanceTo(id):
|
|
|
|
for n in bucket.nodesByDistanceTo(id):
|
|
|
|
if n.id != id:
|
|
|
|
result.add(n)
|
|
|
|
if result.len == k * 2:
|
|
|
|
break
|
|
|
|
result = sortedByIt(result, it.distanceTo(id))
|
|
|
|
if result.len > k:
|
|
|
|
result.setLen(k)
|
|
|
|
|
|
|
|
proc len(r: RoutingTable): int =
|
|
|
|
for b in r.buckets: result += b.len
|
|
|
|
|
2020-07-07 08:56:26 +00:00
|
|
|
proc newKademliaProtocol*[Wire](
|
|
|
|
thisNode: Node, wire: Wire, rng = newRng()): KademliaProtocol[Wire] =
|
|
|
|
if rng == nil: raiseAssert "Need an RNG" # doAssert gives compile error on mac
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
result.new()
|
|
|
|
result.thisNode = thisNode
|
|
|
|
result.wire = wire
|
|
|
|
result.routing.init(thisNode)
|
2020-07-07 08:56:26 +00:00
|
|
|
result.rng = rng
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc bond(k: KademliaProtocol, n: Node): Future[bool] {.async.}
|
|
|
|
proc bondDiscard(k: KademliaProtocol, n: Node) {.async.}
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc updateRoutingTable(k: KademliaProtocol, n: Node)
|
2023-05-10 13:50:04 +00:00
|
|
|
{.raises: [ValueError], gcsafe.} =
|
2019-02-05 15:40:29 +00:00
|
|
|
## Update the routing table entry for the given node.
|
|
|
|
let evictionCandidate = k.routing.addNode(n)
|
|
|
|
if not evictionCandidate.isNil:
|
|
|
|
# This means we couldn't add the node because its bucket is full, so schedule a bond()
|
|
|
|
# with the least recently seen node on that bucket. If the bonding fails the node will
|
|
|
|
# be removed from the bucket and a new one will be picked from the bucket's
|
|
|
|
# replacement cache.
|
2021-04-27 09:29:54 +00:00
|
|
|
asyncSpawn k.bondDiscard(evictionCandidate)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2023-05-10 13:50:04 +00:00
|
|
|
proc doSleep(p: proc() {.gcsafe, raises: [].}) {.async.} =
|
2019-02-05 15:40:29 +00:00
|
|
|
await sleepAsync(REQUEST_TIMEOUT)
|
|
|
|
p()
|
|
|
|
|
|
|
|
template onTimeout(b: untyped) =
|
2021-04-27 09:29:54 +00:00
|
|
|
asyncSpawn doSleep() do():
|
2019-02-05 15:40:29 +00:00
|
|
|
b
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc pingId(n: Node, token: seq[byte]): seq[byte] =
|
2020-04-04 16:44:01 +00:00
|
|
|
result = token & @(n.node.pubkey.toRaw)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc waitPong(k: KademliaProtocol, n: Node, pingid: seq[byte]): Future[bool] =
|
2019-03-13 22:15:26 +00:00
|
|
|
doAssert(pingid notin k.pongFutures, "Already waiting for pong from " & $n)
|
2019-02-05 15:40:29 +00:00
|
|
|
result = newFuture[bool]("waitPong")
|
|
|
|
let fut = result
|
|
|
|
k.pongFutures[pingid] = result
|
|
|
|
onTimeout:
|
|
|
|
if not fut.finished:
|
|
|
|
k.pongFutures.del(pingid)
|
|
|
|
fut.complete(false)
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc ping(k: KademliaProtocol, n: Node): seq[byte] =
|
2019-03-13 22:15:26 +00:00
|
|
|
doAssert(n != k.thisNode)
|
2019-02-05 15:40:29 +00:00
|
|
|
result = k.wire.sendPing(n)
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc waitPing(k: KademliaProtocol, n: Node): Future[bool] =
|
2019-02-05 15:40:29 +00:00
|
|
|
result = newFuture[bool]("waitPing")
|
2019-03-13 22:15:26 +00:00
|
|
|
doAssert(n notin k.pingFutures)
|
2019-02-05 15:40:29 +00:00
|
|
|
k.pingFutures[n] = result
|
|
|
|
let fut = result
|
|
|
|
onTimeout:
|
|
|
|
if not fut.finished:
|
|
|
|
k.pingFutures.del(n)
|
|
|
|
fut.complete(false)
|
|
|
|
|
2023-05-10 13:50:04 +00:00
|
|
|
proc waitNeighbours(k: KademliaProtocol, remote: Node): Future[seq[Node]] =
|
2019-03-13 22:15:26 +00:00
|
|
|
doAssert(remote notin k.neighboursCallbacks)
|
2019-02-05 15:40:29 +00:00
|
|
|
result = newFuture[seq[Node]]("waitNeighbours")
|
|
|
|
let fut = result
|
|
|
|
var neighbours = newSeqOfCap[Node](BUCKET_SIZE)
|
2023-05-10 13:50:04 +00:00
|
|
|
k.neighboursCallbacks[remote] = proc(n: seq[Node]) {.gcsafe, raises: [].} =
|
2019-02-05 15:40:29 +00:00
|
|
|
# This callback is expected to be called multiple times because nodes usually
|
|
|
|
# split the neighbours replies into multiple packets, so we only complete the
|
|
|
|
# future event.set() we've received enough neighbours.
|
|
|
|
|
|
|
|
for i in n:
|
|
|
|
if i != k.thisNode:
|
|
|
|
neighbours.add(i)
|
|
|
|
if neighbours.len == BUCKET_SIZE:
|
|
|
|
k.neighboursCallbacks.del(remote)
|
2019-03-13 22:15:26 +00:00
|
|
|
doAssert(not fut.finished)
|
2019-02-05 15:40:29 +00:00
|
|
|
fut.complete(neighbours)
|
|
|
|
|
|
|
|
onTimeout:
|
|
|
|
if not fut.finished:
|
|
|
|
k.neighboursCallbacks.del(remote)
|
|
|
|
fut.complete(neighbours)
|
|
|
|
|
2021-04-06 13:15:56 +00:00
|
|
|
# Exported for test.
|
|
|
|
proc findNode*(k: KademliaProtocol, nodesSeen: ref HashSet[Node],
|
|
|
|
nodeId: NodeId, remote: Node): Future[seq[Node]] {.async.} =
|
|
|
|
if remote in k.neighboursCallbacks:
|
|
|
|
# Sometimes findNode is called while another findNode is already in flight.
|
|
|
|
# It's a bug when this happens, and the logic should probably be fixed
|
|
|
|
# elsewhere. However, this small fix has been tested and proven adequate.
|
|
|
|
debug "Ignoring peer already in k.neighboursCallbacks", peer = remote
|
|
|
|
result = newSeq[Node]()
|
|
|
|
return
|
|
|
|
k.wire.sendFindNode(remote, nodeId)
|
|
|
|
var candidates = await k.waitNeighbours(remote)
|
|
|
|
if candidates.len == 0:
|
|
|
|
trace "Got no candidates from peer, returning", peer = remote
|
|
|
|
result = candidates
|
|
|
|
else:
|
|
|
|
# The following line:
|
|
|
|
# 1. Add new candidates to nodesSeen so that we don't attempt to bond with failing ones
|
|
|
|
# in the future
|
|
|
|
# 2. Removes all previously seen nodes from candidates
|
|
|
|
# 3. Deduplicates candidates
|
|
|
|
candidates.keepItIf(not nodesSeen[].containsOrIncl(it))
|
|
|
|
trace "Got new candidates", count = candidates.len
|
2021-04-27 09:29:54 +00:00
|
|
|
|
|
|
|
var bondedNodes: seq[Future[bool]] = @[]
|
|
|
|
for node in candidates:
|
2022-11-07 13:25:11 +00:00
|
|
|
if node != k.thisNode:
|
|
|
|
bondedNodes.add(k.bond(node))
|
2021-04-27 09:29:54 +00:00
|
|
|
|
|
|
|
await allFutures(bondedNodes)
|
|
|
|
|
|
|
|
for i in 0..<bondedNodes.len:
|
|
|
|
let b = bondedNodes[i]
|
|
|
|
# `bond` will not raise so there should be no failures,
|
|
|
|
# and for cancellation this should be fine to raise for now.
|
|
|
|
doAssert(b.finished() and not(b.failed()))
|
|
|
|
let bonded = b.read()
|
|
|
|
if not bonded: candidates[i] = nil
|
|
|
|
|
2021-04-06 13:15:56 +00:00
|
|
|
candidates.keepItIf(not it.isNil)
|
|
|
|
trace "Bonded with candidates", count = candidates.len
|
|
|
|
result = candidates
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
proc populateNotFullBuckets(k: KademliaProtocol) =
|
|
|
|
## Go through all buckets that are not full and try to fill them.
|
|
|
|
##
|
|
|
|
## For every node in the replacement cache of every non-full bucket, try to bond.
|
|
|
|
## When the bonding succeeds the node is automatically added to the bucket.
|
|
|
|
for bucket in k.routing.notFullBuckets:
|
|
|
|
for node in bucket.replacementCache:
|
2021-04-27 09:29:54 +00:00
|
|
|
asyncSpawn k.bondDiscard(node)
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc bond(k: KademliaProtocol, n: Node): Future[bool] {.async.} =
|
2019-02-05 15:40:29 +00:00
|
|
|
## Bond with the given node.
|
|
|
|
##
|
|
|
|
## Bonding consists of pinging the node, waiting for a pong and maybe a ping as well.
|
|
|
|
## It is necessary to do this at least once before we send findNode requests to a node.
|
2019-02-26 12:47:26 +00:00
|
|
|
trace "Bonding to peer", n
|
2019-02-05 15:40:29 +00:00
|
|
|
if n in k.routing:
|
|
|
|
return true
|
|
|
|
|
|
|
|
let pid = pingId(n, k.ping(n))
|
|
|
|
if pid in k.pongFutures:
|
2019-02-26 12:47:26 +00:00
|
|
|
debug "Bonding failed, already waiting for pong", n
|
2019-02-05 15:40:29 +00:00
|
|
|
return false
|
|
|
|
|
|
|
|
let gotPong = await k.waitPong(n, pid)
|
|
|
|
if not gotPong:
|
2019-04-15 09:50:00 +00:00
|
|
|
trace "Bonding failed, didn't receive pong from", n
|
2019-02-05 15:40:29 +00:00
|
|
|
# Drop the failing node and schedule a populateNotFullBuckets() call to try and
|
|
|
|
# fill its spot.
|
|
|
|
k.routing.removeNode(n)
|
|
|
|
k.populateNotFullBuckets()
|
|
|
|
return false
|
|
|
|
|
|
|
|
# Give the remote node a chance to ping us before we move on and start sending findNode
|
|
|
|
# requests. It is ok for waitPing() to timeout and return false here as that just means
|
|
|
|
# the remote remembers us.
|
|
|
|
if n in k.pingFutures:
|
|
|
|
debug "Bonding failed, already waiting for ping", n
|
|
|
|
return false
|
|
|
|
|
|
|
|
discard await k.waitPing(n)
|
|
|
|
|
2019-02-26 12:47:26 +00:00
|
|
|
trace "Bonding completed successfully", n
|
2019-02-05 15:40:29 +00:00
|
|
|
k.updateRoutingTable(n)
|
|
|
|
return true
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc bondDiscard(k: KademliaProtocol, n: Node) {.async.} =
|
|
|
|
discard (await k.bond(n))
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
proc sortByDistance(nodes: var seq[Node], nodeId: NodeId, maxResults = 0) =
|
|
|
|
nodes = nodes.sortedByIt(it.distanceTo(nodeId))
|
|
|
|
if maxResults != 0 and nodes.len > maxResults:
|
|
|
|
nodes.setLen(maxResults)
|
|
|
|
|
|
|
|
proc lookup*(k: KademliaProtocol, nodeId: NodeId): Future[seq[Node]] {.async.} =
|
|
|
|
## Lookup performs a network search for nodes close to the given target.
|
|
|
|
|
|
|
|
## It approaches the target by querying nodes that are closer to it on each iteration. The
|
|
|
|
## given target does not need to be an actual node identifier.
|
2020-04-18 08:17:59 +00:00
|
|
|
var nodesAsked = initHashSet[Node]()
|
2021-04-06 13:15:56 +00:00
|
|
|
let nodesSeen = new(HashSet[Node])
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc excludeIfAsked(nodes: seq[Node]): seq[Node] =
|
2020-04-18 08:17:59 +00:00
|
|
|
result = toSeq(items(nodes.toHashSet() - nodesAsked))
|
2019-02-05 15:40:29 +00:00
|
|
|
sortByDistance(result, nodeId, FIND_CONCURRENCY)
|
|
|
|
|
|
|
|
var closest = k.routing.neighbours(nodeId)
|
|
|
|
trace "Starting lookup; initial neighbours: ", closest
|
|
|
|
var nodesToAsk = excludeIfAsked(closest)
|
|
|
|
while nodesToAsk.len != 0:
|
|
|
|
trace "Node lookup; querying ", nodesToAsk
|
2020-04-18 08:17:59 +00:00
|
|
|
nodesAsked.incl(nodesToAsk.toHashSet())
|
2021-04-27 09:29:54 +00:00
|
|
|
|
|
|
|
var findNodeRequests: seq[Future[seq[Node]]] = @[]
|
|
|
|
for node in nodesToAsk:
|
|
|
|
findNodeRequests.add(k.findNode(nodesSeen, nodeId, node))
|
|
|
|
|
|
|
|
await allFutures(findNodeRequests)
|
|
|
|
|
|
|
|
for candidates in findNodeRequests:
|
|
|
|
# `findNode` will not raise so there should be no failures,
|
|
|
|
# and for cancellation this should be fine to raise for now.
|
|
|
|
doAssert(candidates.finished() and not(candidates.failed()))
|
|
|
|
closest.add(candidates.read())
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
sortByDistance(closest, nodeId, BUCKET_SIZE)
|
|
|
|
nodesToAsk = excludeIfAsked(closest)
|
|
|
|
|
|
|
|
trace "Kademlia lookup finished", target = nodeId.toHex, closest
|
|
|
|
result = closest
|
|
|
|
|
|
|
|
proc lookupRandom*(k: KademliaProtocol): Future[seq[Node]] =
|
2022-06-17 20:45:37 +00:00
|
|
|
k.lookup(k.rng[].generate(NodeId))
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc resolve*(k: KademliaProtocol, id: NodeId): Future[Node] {.async.} =
|
|
|
|
let closest = await k.lookup(id)
|
|
|
|
for n in closest:
|
|
|
|
if n.id == id: return n
|
|
|
|
|
2019-04-04 12:35:30 +00:00
|
|
|
proc bootstrap*(k: KademliaProtocol, bootstrapNodes: seq[Node], retries = 0) {.async.} =
|
|
|
|
## Bond with bootstrap nodes and do initial lookup. Retry `retries` times
|
|
|
|
## in case of failure, or indefinitely if `retries` is 0.
|
2019-04-10 17:03:58 +00:00
|
|
|
var retryInterval = chronos.milliseconds(2)
|
2019-04-04 12:35:30 +00:00
|
|
|
var numTries = 0
|
2019-04-09 08:02:33 +00:00
|
|
|
if bootstrapNodes.len != 0:
|
|
|
|
while true:
|
2021-04-27 09:29:54 +00:00
|
|
|
var bondedNodes: seq[Future[bool]] = @[]
|
|
|
|
for node in bootstrapNodes:
|
|
|
|
bondedNodes.add(k.bond(node))
|
|
|
|
await allFutures(bondedNodes)
|
|
|
|
|
|
|
|
# `bond` will not raise so there should be no failures,
|
|
|
|
# and for cancellation this should be fine to raise for now.
|
|
|
|
let bonded = bondedNodes.mapIt(it.read())
|
|
|
|
|
2019-04-09 08:02:33 +00:00
|
|
|
if true notin bonded:
|
|
|
|
inc numTries
|
|
|
|
if retries == 0 or numTries < retries:
|
|
|
|
info "Failed to bond with bootstrap nodes, retrying"
|
2019-04-10 17:03:58 +00:00
|
|
|
retryInterval = min(chronos.seconds(10), retryInterval * 2)
|
2019-04-10 07:04:48 +00:00
|
|
|
await sleepAsync(retryInterval)
|
2019-04-09 08:02:33 +00:00
|
|
|
else:
|
|
|
|
info "Failed to bond with bootstrap nodes"
|
|
|
|
return
|
2019-04-04 12:35:30 +00:00
|
|
|
else:
|
2019-04-09 08:02:33 +00:00
|
|
|
break
|
2019-04-10 07:04:48 +00:00
|
|
|
discard await k.lookupRandom() # Prepopulate the routing table
|
2019-04-09 08:02:33 +00:00
|
|
|
else:
|
|
|
|
info "Skipping discovery bootstrap, no bootnodes provided"
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc recvPong*(k: KademliaProtocol, n: Node, token: seq[byte]) =
|
2019-02-26 12:47:26 +00:00
|
|
|
trace "<<< pong from ", n
|
2020-04-04 16:44:01 +00:00
|
|
|
let pingid = token & @(n.node.pubkey.toRaw)
|
2019-02-05 15:40:29 +00:00
|
|
|
var future: Future[bool]
|
|
|
|
if k.pongFutures.take(pingid, future):
|
|
|
|
future.complete(true)
|
2022-09-06 15:19:47 +00:00
|
|
|
k.updateLastPongReceived(n, getTime())
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2022-11-10 15:32:57 +00:00
|
|
|
proc recvPing*(k: KademliaProtocol, n: Node, msgHash: auto)
|
2023-05-10 13:50:04 +00:00
|
|
|
{.raises: [ValueError].} =
|
2019-02-26 12:47:26 +00:00
|
|
|
trace "<<< ping from ", n
|
2019-02-05 15:40:29 +00:00
|
|
|
k.wire.sendPong(n, msgHash)
|
|
|
|
|
2022-09-06 15:19:47 +00:00
|
|
|
if getTime() - k.lastPongReceived(n) > BOND_EXPIRATION:
|
2022-11-07 13:25:11 +00:00
|
|
|
# TODO: It is strange that this would occur, as it means our own node would
|
|
|
|
# have pinged us which should have caused an assert in the first place.
|
|
|
|
if n != k.thisNode:
|
|
|
|
let pingId = pingId(n, k.ping(n))
|
|
|
|
|
|
|
|
let fut = if pingId in k.pongFutures:
|
|
|
|
k.pongFutures[pingId]
|
|
|
|
else:
|
|
|
|
k.waitPong(n, pingId)
|
|
|
|
|
|
|
|
let cb = proc(data: pointer) {.gcsafe.} =
|
|
|
|
# fut.read == true if pingid exists
|
|
|
|
try:
|
|
|
|
if fut.completed and fut.read:
|
|
|
|
k.updateRoutingTable(n)
|
|
|
|
except CatchableError as ex:
|
|
|
|
error "recvPing:WaitPong exception", msg=ex.msg
|
|
|
|
|
|
|
|
fut.addCallback cb
|
2022-09-06 15:19:47 +00:00
|
|
|
else:
|
|
|
|
k.updateRoutingTable(n)
|
|
|
|
|
2019-02-05 15:40:29 +00:00
|
|
|
var future: Future[bool]
|
|
|
|
if k.pingFutures.take(n, future):
|
|
|
|
future.complete(true)
|
2022-09-06 15:19:47 +00:00
|
|
|
k.updateLastPingReceived(n, getTime())
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
proc recvNeighbours*(k: KademliaProtocol, remote: Node, neighbours: seq[Node]) =
|
|
|
|
## Process a neighbours response.
|
|
|
|
##
|
|
|
|
## Neighbours responses should only be received as a reply to a find_node, and that is only
|
|
|
|
## done as part of node lookup, so the actual processing is left to the callback from
|
|
|
|
## neighbours_callbacks, which is added (and removed after it's done or timed out) in
|
|
|
|
## wait_neighbours().
|
2019-02-26 12:47:26 +00:00
|
|
|
trace "Received neighbours", remote, neighbours
|
2019-02-05 15:40:29 +00:00
|
|
|
let cb = k.neighboursCallbacks.getOrDefault(remote)
|
|
|
|
if not cb.isNil:
|
|
|
|
cb(neighbours)
|
|
|
|
else:
|
2019-03-28 15:55:41 +00:00
|
|
|
trace "Unexpected neighbours, probably came too late", remote
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc recvFindNode*(k: KademliaProtocol, remote: Node, nodeId: NodeId)
|
2023-05-10 13:50:04 +00:00
|
|
|
{.raises: [ValueError].} =
|
2019-02-05 15:40:29 +00:00
|
|
|
if remote notin k.routing:
|
|
|
|
# FIXME: This is not correct; a node we've bonded before may have become unavailable
|
|
|
|
# and thus removed from self.routing, but once it's back online we should accept
|
|
|
|
# find_nodes from them.
|
2019-03-28 15:55:41 +00:00
|
|
|
trace "Ignoring find_node request from unknown node ", remote
|
2019-02-05 15:40:29 +00:00
|
|
|
return
|
|
|
|
k.updateRoutingTable(remote)
|
|
|
|
var found = k.routing.neighbours(nodeId)
|
|
|
|
found.sort() do(x, y: Node) -> int: cmp(x.id, y.id)
|
|
|
|
k.wire.sendNeighbours(remote, found)
|
|
|
|
|
|
|
|
proc randomNodes*(k: KademliaProtocol, count: int): seq[Node] =
|
|
|
|
var count = count
|
|
|
|
let sz = k.routing.len
|
|
|
|
if count > sz:
|
2019-03-28 18:21:35 +00:00
|
|
|
debug "Looking for peers", requested = count, present = sz
|
2019-02-05 15:40:29 +00:00
|
|
|
count = sz
|
|
|
|
|
|
|
|
result = newSeqOfCap[Node](count)
|
2020-04-18 08:17:59 +00:00
|
|
|
var seen = initHashSet[Node]()
|
2019-02-05 15:40:29 +00:00
|
|
|
|
2022-11-16 16:44:00 +00:00
|
|
|
# This is a rather inefficient way of randomizing nodes from all buckets, but even if we
|
2019-02-05 15:40:29 +00:00
|
|
|
# iterate over all nodes in the routing table, the time it takes would still be
|
|
|
|
# insignificant compared to the time it takes for the network roundtrips when connecting
|
|
|
|
# to nodes.
|
|
|
|
while len(seen) < count:
|
2022-09-19 08:47:30 +00:00
|
|
|
let bucket = k.rng[].sample(k.routing.buckets)
|
2019-02-05 15:40:29 +00:00
|
|
|
if bucket.nodes.len != 0:
|
2022-09-19 08:47:30 +00:00
|
|
|
let node = k.rng[].sample(bucket.nodes)
|
2019-02-05 15:40:29 +00:00
|
|
|
if node notin seen:
|
|
|
|
result.add(node)
|
|
|
|
seen.incl(node)
|
|
|
|
|
2021-04-27 09:29:54 +00:00
|
|
|
proc nodesDiscovered*(k: KademliaProtocol): int = k.routing.len
|
2019-02-05 15:40:29 +00:00
|
|
|
|
|
|
|
when isMainModule:
|
|
|
|
proc randomNode(): Node =
|
|
|
|
newNode("enode://aa36fdf33dd030378a0168efe6ed7d5cc587fafa3cdd375854fe735a2e11ea3650ba29644e2db48368c46e1f60e716300ba49396cd63778bf8a818c09bded46f@13.93.211.84:30303")
|
|
|
|
|
|
|
|
var nodes = @[randomNode()]
|
|
|
|
doAssert(computeSharedPrefixBits(nodes) == ID_SIZE)
|
|
|
|
nodes.add(randomNode())
|
|
|
|
nodes[0].id = 0b1.u256
|
|
|
|
nodes[1].id = 0b0.u256
|
|
|
|
doAssert(computeSharedPrefixBits(nodes) == ID_SIZE - 1)
|
|
|
|
|
|
|
|
nodes[0].id = 0b010.u256
|
|
|
|
nodes[1].id = 0b110.u256
|
|
|
|
doAssert(computeSharedPrefixBits(nodes) == ID_SIZE - 3)
|