mirror of
https://github.com/logos-storage/logos-storage-nim.git
synced 2026-01-04 06:23:06 +00:00
deps: upgrade libp2p & constantine (#1167)
* upgrade libp2p and constantine * fix libp2p update issues * add missing vendor package * add missing vendor package
This commit is contained in:
parent
9d7b521519
commit
3a312596bf
10
.gitmodules
vendored
10
.gitmodules
vendored
@ -221,3 +221,13 @@
|
|||||||
[submodule "vendor/nph"]
|
[submodule "vendor/nph"]
|
||||||
path = vendor/nph
|
path = vendor/nph
|
||||||
url = https://github.com/arnetheduck/nph.git
|
url = https://github.com/arnetheduck/nph.git
|
||||||
|
[submodule "vendor/nim-quic"]
|
||||||
|
path = vendor/nim-quic
|
||||||
|
url = https://github.com/vacp2p/nim-quic.git
|
||||||
|
ignore = untracked
|
||||||
|
branch = master
|
||||||
|
[submodule "vendor/nim-ngtcp2"]
|
||||||
|
path = vendor/nim-ngtcp2
|
||||||
|
url = https://github.com/vacp2p/nim-ngtcp2.git
|
||||||
|
ignore = untracked
|
||||||
|
branch = master
|
||||||
|
|||||||
@ -678,7 +678,9 @@ proc new*(
|
|||||||
advertiser: advertiser,
|
advertiser: advertiser,
|
||||||
)
|
)
|
||||||
|
|
||||||
proc peerEventHandler(peerId: PeerId, event: PeerEvent) {.async.} =
|
proc peerEventHandler(
|
||||||
|
peerId: PeerId, event: PeerEvent
|
||||||
|
): Future[void] {.gcsafe, async: (raises: [CancelledError]).} =
|
||||||
if event.kind == PeerEventKind.Joined:
|
if event.kind == PeerEventKind.Joined:
|
||||||
await self.setupPeer(peerId)
|
await self.setupPeer(peerId)
|
||||||
else:
|
else:
|
||||||
|
|||||||
@ -323,7 +323,9 @@ method init*(self: BlockExcNetwork) =
|
|||||||
## Perform protocol initialization
|
## Perform protocol initialization
|
||||||
##
|
##
|
||||||
|
|
||||||
proc peerEventHandler(peerId: PeerId, event: PeerEvent) {.async.} =
|
proc peerEventHandler(
|
||||||
|
peerId: PeerId, event: PeerEvent
|
||||||
|
): Future[void] {.gcsafe, async: (raises: [CancelledError]).} =
|
||||||
if event.kind == PeerEventKind.Joined:
|
if event.kind == PeerEventKind.Joined:
|
||||||
self.setupPeer(peerId)
|
self.setupPeer(peerId)
|
||||||
else:
|
else:
|
||||||
@ -332,7 +334,9 @@ method init*(self: BlockExcNetwork) =
|
|||||||
self.switch.addPeerEventHandler(peerEventHandler, PeerEventKind.Joined)
|
self.switch.addPeerEventHandler(peerEventHandler, PeerEventKind.Joined)
|
||||||
self.switch.addPeerEventHandler(peerEventHandler, PeerEventKind.Left)
|
self.switch.addPeerEventHandler(peerEventHandler, PeerEventKind.Left)
|
||||||
|
|
||||||
proc handler(conn: Connection, proto: string) {.async.} =
|
proc handler(
|
||||||
|
conn: Connection, proto: string
|
||||||
|
): Future[void] {.async: (raises: [CancelledError]).} =
|
||||||
let peerId = conn.peerId
|
let peerId = conn.peerId
|
||||||
let blockexcPeer = self.getOrCreatePeer(peerId)
|
let blockexcPeer = self.getOrCreatePeer(peerId)
|
||||||
await blockexcPeer.readLoop(conn) # attach read loop
|
await blockexcPeer.readLoop(conn) # attach read loop
|
||||||
|
|||||||
@ -97,7 +97,7 @@ proc write*(pb: var ProtoBuffer, field: int, value: WantList) =
|
|||||||
pb.write(field, ipb)
|
pb.write(field, ipb)
|
||||||
|
|
||||||
proc write*(pb: var ProtoBuffer, field: int, value: BlockDelivery) =
|
proc write*(pb: var ProtoBuffer, field: int, value: BlockDelivery) =
|
||||||
var ipb = initProtoBuffer(maxSize = MaxBlockSize)
|
var ipb = initProtoBuffer()
|
||||||
ipb.write(1, value.blk.cid.data.buffer)
|
ipb.write(1, value.blk.cid.data.buffer)
|
||||||
ipb.write(2, value.blk.data)
|
ipb.write(2, value.blk.data)
|
||||||
ipb.write(3, value.address)
|
ipb.write(3, value.address)
|
||||||
@ -128,7 +128,7 @@ proc write*(pb: var ProtoBuffer, field: int, value: StateChannelUpdate) =
|
|||||||
pb.write(field, ipb)
|
pb.write(field, ipb)
|
||||||
|
|
||||||
proc protobufEncode*(value: Message): seq[byte] =
|
proc protobufEncode*(value: Message): seq[byte] =
|
||||||
var ipb = initProtoBuffer(maxSize = MaxMessageSize)
|
var ipb = initProtoBuffer()
|
||||||
ipb.write(1, value.wantList)
|
ipb.write(1, value.wantList)
|
||||||
for v in value.payload:
|
for v in value.payload:
|
||||||
ipb.write(3, v)
|
ipb.write(3, v)
|
||||||
@ -254,16 +254,14 @@ proc decode*(
|
|||||||
proc protobufDecode*(_: type Message, msg: seq[byte]): ProtoResult[Message] =
|
proc protobufDecode*(_: type Message, msg: seq[byte]): ProtoResult[Message] =
|
||||||
var
|
var
|
||||||
value = Message()
|
value = Message()
|
||||||
pb = initProtoBuffer(msg, maxSize = MaxMessageSize)
|
pb = initProtoBuffer(msg)
|
||||||
ipb: ProtoBuffer
|
ipb: ProtoBuffer
|
||||||
sublist: seq[seq[byte]]
|
sublist: seq[seq[byte]]
|
||||||
if ?pb.getField(1, ipb):
|
if ?pb.getField(1, ipb):
|
||||||
value.wantList = ?WantList.decode(ipb)
|
value.wantList = ?WantList.decode(ipb)
|
||||||
if ?pb.getRepeatedField(3, sublist):
|
if ?pb.getRepeatedField(3, sublist):
|
||||||
for item in sublist:
|
for item in sublist:
|
||||||
value.payload.add(
|
value.payload.add(?BlockDelivery.decode(initProtoBuffer(item)))
|
||||||
?BlockDelivery.decode(initProtoBuffer(item, maxSize = MaxBlockSize))
|
|
||||||
)
|
|
||||||
if ?pb.getRepeatedField(4, sublist):
|
if ?pb.getRepeatedField(4, sublist):
|
||||||
for item in sublist:
|
for item in sublist:
|
||||||
value.blockPresences.add(?BlockPresence.decode(initProtoBuffer(item)))
|
value.blockPresences.add(?BlockPresence.decode(initProtoBuffer(item)))
|
||||||
|
|||||||
@ -27,11 +27,11 @@ const MaxMerkleTreeSize = 100.MiBs.uint
|
|||||||
const MaxMerkleProofSize = 1.MiBs.uint
|
const MaxMerkleProofSize = 1.MiBs.uint
|
||||||
|
|
||||||
proc encode*(self: CodexTree): seq[byte] =
|
proc encode*(self: CodexTree): seq[byte] =
|
||||||
var pb = initProtoBuffer(maxSize = MaxMerkleTreeSize)
|
var pb = initProtoBuffer()
|
||||||
pb.write(1, self.mcodec.uint64)
|
pb.write(1, self.mcodec.uint64)
|
||||||
pb.write(2, self.leavesCount.uint64)
|
pb.write(2, self.leavesCount.uint64)
|
||||||
for node in self.nodes:
|
for node in self.nodes:
|
||||||
var nodesPb = initProtoBuffer(maxSize = MaxMerkleTreeSize)
|
var nodesPb = initProtoBuffer()
|
||||||
nodesPb.write(1, node)
|
nodesPb.write(1, node)
|
||||||
nodesPb.finish()
|
nodesPb.finish()
|
||||||
pb.write(3, nodesPb)
|
pb.write(3, nodesPb)
|
||||||
@ -40,7 +40,7 @@ proc encode*(self: CodexTree): seq[byte] =
|
|||||||
pb.buffer
|
pb.buffer
|
||||||
|
|
||||||
proc decode*(_: type CodexTree, data: seq[byte]): ?!CodexTree =
|
proc decode*(_: type CodexTree, data: seq[byte]): ?!CodexTree =
|
||||||
var pb = initProtoBuffer(data, maxSize = MaxMerkleTreeSize)
|
var pb = initProtoBuffer(data)
|
||||||
var mcodecCode: uint64
|
var mcodecCode: uint64
|
||||||
var leavesCount: uint64
|
var leavesCount: uint64
|
||||||
discard ?pb.getField(1, mcodecCode).mapFailure
|
discard ?pb.getField(1, mcodecCode).mapFailure
|
||||||
@ -63,13 +63,13 @@ proc decode*(_: type CodexTree, data: seq[byte]): ?!CodexTree =
|
|||||||
CodexTree.fromNodes(mcodec, nodes, leavesCount.int)
|
CodexTree.fromNodes(mcodec, nodes, leavesCount.int)
|
||||||
|
|
||||||
proc encode*(self: CodexProof): seq[byte] =
|
proc encode*(self: CodexProof): seq[byte] =
|
||||||
var pb = initProtoBuffer(maxSize = MaxMerkleProofSize)
|
var pb = initProtoBuffer()
|
||||||
pb.write(1, self.mcodec.uint64)
|
pb.write(1, self.mcodec.uint64)
|
||||||
pb.write(2, self.index.uint64)
|
pb.write(2, self.index.uint64)
|
||||||
pb.write(3, self.nleaves.uint64)
|
pb.write(3, self.nleaves.uint64)
|
||||||
|
|
||||||
for node in self.path:
|
for node in self.path:
|
||||||
var nodesPb = initProtoBuffer(maxSize = MaxMerkleTreeSize)
|
var nodesPb = initProtoBuffer()
|
||||||
nodesPb.write(1, node)
|
nodesPb.write(1, node)
|
||||||
nodesPb.finish()
|
nodesPb.finish()
|
||||||
pb.write(4, nodesPb)
|
pb.write(4, nodesPb)
|
||||||
@ -78,7 +78,7 @@ proc encode*(self: CodexProof): seq[byte] =
|
|||||||
pb.buffer
|
pb.buffer
|
||||||
|
|
||||||
proc decode*(_: type CodexProof, data: seq[byte]): ?!CodexProof =
|
proc decode*(_: type CodexProof, data: seq[byte]): ?!CodexProof =
|
||||||
var pb = initProtoBuffer(data, maxSize = MaxMerkleProofSize)
|
var pb = initProtoBuffer(data)
|
||||||
var mcodecCode: uint64
|
var mcodecCode: uint64
|
||||||
var index: uint64
|
var index: uint64
|
||||||
var nleaves: uint64
|
var nleaves: uint64
|
||||||
|
|||||||
2
vendor/constantine
vendored
2
vendor/constantine
vendored
@ -1 +1 @@
|
|||||||
Subproject commit bc3845aa492b52f7fef047503b1592e830d1a774
|
Subproject commit 8d6a6a38b90fb8ee3ec2230839773e69aab36d80
|
||||||
2
vendor/nim-codex-dht
vendored
2
vendor/nim-codex-dht
vendored
@ -1 +1 @@
|
|||||||
Subproject commit 4bd3a39e0030f8ee269ef217344b6b59ec2be6dc
|
Subproject commit f6eef1ac95c70053b2518f1e3909c909ed8701a6
|
||||||
2
vendor/nim-libp2p
vendored
2
vendor/nim-libp2p
vendored
@ -1 +1 @@
|
|||||||
Subproject commit 036e110a6080fba1a1662c58cfd8c21f9a548021
|
Subproject commit c08d80734989b028b3d1705f2188d783a343aac0
|
||||||
1
vendor/nim-ngtcp2
vendored
Submodule
1
vendor/nim-ngtcp2
vendored
Submodule
@ -0,0 +1 @@
|
|||||||
|
Subproject commit 6834f4756b6af58356ac9c4fef3d71db3c3ae5fe
|
||||||
1
vendor/nim-quic
vendored
Submodule
1
vendor/nim-quic
vendored
Submodule
@ -0,0 +1 @@
|
|||||||
|
Subproject commit ddcb31ffb74b5460ab37fd13547eca90594248bc
|
||||||
Loading…
x
Reference in New Issue
Block a user