# Nimbus # Copyright (c) 2022-2024 Status Research & Development GmbH # Licensed and distributed under either of # * MIT license (license terms in the root directory or at https://opensource.org/licenses/MIT). # * Apache v2 license (license terms in the root directory or at https://www.apache.org/licenses/LICENSE-2.0). # at your option. This file may not be copied, modified, or distributed except according to those terms. {.push raises: [].} import std/sequtils, chronos, stew/[byteutils, leb128, endians2], chronicles, eth/utp/utp_discv5_protocol, # even though utp_discv5_protocol exports this, import is still needed, # perhaps protocol.Protocol type of usage? eth/p2p/discoveryv5/protocol, ./messages export utp_discv5_protocol logScope: topics = "portal_stream" const utpProtocolId = "utp".toBytes() defaultConnectionTimeout = 15.seconds defaultContentReadTimeout = 60.seconds # TalkReq message is used as transport for uTP. It is assumed here that Portal # protocol messages were exchanged before sending uTP over discv5 data. This # means that a session is established and that the discv5 messages send are # discv5 ordinary message packets, for which below calculation applies. talkReqOverhead = getTalkReqOverhead(utpProtocolId) utpHeaderOverhead = 20 maxUtpPayloadSize = maxDiscv5PacketSize - talkReqOverhead - utpHeaderOverhead type ContentRequest = object connectionId: uint16 nodeId: NodeId content: seq[byte] timeout: Moment ContentOffer = object connectionId: uint16 nodeId: NodeId contentKeys: ContentKeysList timeout: Moment PortalStream* = ref object transport: UtpDiscv5Protocol # TODO: # Decide on what's the better collection to use and set some limits in them # on how many uTP transfers allowed to happen concurrently. # Either set some limit, and drop whatever comes next. Unsure how to # communicate that with the peer however. Or have some more async waiting # until a spot becomes free, like with an AsyncQueue. Although the latter # probably can not be used here directly. This system however does needs # some agreement on timeout values of how long a uTP socket may be # "listening" before it times out because of inactivity. # Or, depending on the direction, it might also depend on the time out # values of the discovery v5 talkresp message. # TODO: Should the content key also be stored to be able to validate the # received data? contentRequests: seq[ContentRequest] contentOffers: seq[ContentOffer] connectionTimeout: Duration contentReadTimeout*: Duration rng: ref HmacDrbgContext contentQueue*: AsyncQueue[(Opt[NodeId], ContentKeysList, seq[seq[byte]])] StreamManager* = ref object transport: UtpDiscv5Protocol streams: seq[PortalStream] rng: ref HmacDrbgContext proc pruneAllowedConnections(stream: PortalStream) = # Prune requests and offers that didn't receive a connection request # before `connectionTimeout`. let now = Moment.now() stream.contentRequests.keepIf( proc(x: ContentRequest): bool = x.timeout > now ) stream.contentOffers.keepIf( proc(x: ContentOffer): bool = x.timeout > now ) proc addContentOffer*( stream: PortalStream, nodeId: NodeId, contentKeys: ContentKeysList ): Bytes2 = stream.pruneAllowedConnections() # TODO: Should we check if `NodeId` & `connectionId` combo already exists? # What happens if we get duplicates? var connectionId: Bytes2 stream.rng[].generate(connectionId) # uTP protocol uses BE for all values in the header, incl. connection id. let id = uint16.fromBytesBE(connectionId) debug "Register new incoming offer", contentKeys let contentOffer = ContentOffer( connectionId: id, nodeId: nodeId, contentKeys: contentKeys, timeout: Moment.now() + stream.connectionTimeout, ) stream.contentOffers.add(contentOffer) return connectionId proc addContentRequest*( stream: PortalStream, nodeId: NodeId, content: seq[byte] ): Bytes2 = stream.pruneAllowedConnections() # TODO: Should we check if `NodeId` & `connectionId` combo already exists? # What happens if we get duplicates? var connectionId: Bytes2 stream.rng[].generate(connectionId) # uTP protocol uses BE for all values in the header, incl. connection id. let id = uint16.fromBytesBE(connectionId) let contentRequest = ContentRequest( connectionId: id, nodeId: nodeId, content: content, timeout: Moment.now() + stream.connectionTimeout, ) stream.contentRequests.add(contentRequest) return connectionId proc connectTo*( stream: PortalStream, nodeAddress: NodeAddress, connectionId: uint16 ): Future[Result[UtpSocket[NodeAddress], string]] {.async.} = let connectRes = await stream.transport.connectTo(nodeAddress, connectionId) if connectRes.isErr(): case connectRes.error of SocketAlreadyExists: # This means that there is already a socket to this nodeAddress with given # connection id. This means that a peer sent us a connection id which is # already in use. The connection is failed and an error returned. let msg = "Socket to " & $nodeAddress & "with connection id: " & $connectionId & " already exists" return err(msg) of ConnectionTimedOut: # A time-out here means that a uTP SYN packet was re-sent 3 times and # failed to be acked. This should be enough of indication that the # remote host is not reachable and no new connections are attempted. let msg = "uTP timeout while trying to connect to " & $nodeAddress return err(msg) else: return ok(connectRes.get()) proc writeContentRequest( socket: UtpSocket[NodeAddress], stream: PortalStream, request: ContentRequest ) {.async: (raises: [CancelledError]).} = let dataWritten = await socket.write(request.content) if dataWritten.isErr(): debug "Error writing requested data", error = dataWritten.error await socket.closeWait() proc readVarint( socket: UtpSocket[NodeAddress] ): Future[Opt[uint32]] {.async: (raises: [CancelledError]).} = var buffer: array[5, byte] for i in 0 ..< len(buffer): let dataRead = await socket.read(1) if dataRead.len() == 0: return err() buffer[i] = dataRead[0] let (lenU32, bytesRead) = fromBytes(uint32, buffer.toOpenArray(0, i), Leb128) if bytesRead > 0: return ok(lenU32) elif bytesRead == 0: continue else: return err() proc readContentItem( socket: UtpSocket[NodeAddress] ): Future[Opt[seq[byte]]] {.async: (raises: [CancelledError]).} = let len = await socket.readVarint() if len.isOk(): let contentItem = await socket.read(len.get()) if contentItem.len() == len.get().int: return ok(contentItem) else: return err() else: return err() proc readContentOffer( socket: UtpSocket[NodeAddress], stream: PortalStream, offer: ContentOffer ) {.async: (raises: [CancelledError]).} = # Read number of content items according to amount of ContentKeys accepted. # This will either end with a FIN, or because the read action times out or # because the number of expected items was read (if this happens and no FIN # was received yet, a FIN will be send from this side). # None of this means that the contentItems are valid, further validation is # required. # Socket will be closed when this call ends. # TODO: Currently reading from the socket 1 item at a time, and validating # items at later time. Uncertain what is best approach here (mostly from a # security PoV), e.g. other options such as reading all content from socket at # once, then processing the individual content items. Or reading and # validating one per time. let amount = offer.contentKeys.len() var contentItems: seq[seq[byte]] for i in 0 ..< amount: let contentItemFut = socket.readContentItem() if await contentItemFut.withTimeout(stream.contentReadTimeout): let contentItem = await contentItemFut if contentItem.isOk(): contentItems.add(contentItem.get()) else: # Invalid data, stop reading content, but still process data received # so far. debug "Reading content item failed, content offer failed", contentKeys = offer.contentKeys break else: # Read timed out, stop further reading, but still process data received # so far. debug "Reading data from socket timed out, content offer failed", contentKeys = offer.contentKeys break if socket.atEof(): # Destroy socket and not closing as we already received FIN. Closing would # send also a FIN from our side, see also: # https://github.com/status-im/nim-eth/blob/b2dab4be0839c95ca2564df9eacf81995bf57802/eth/utp/utp_socket.nim#L1223 await socket.destroyWait() else: # This means FIN didn't arrive yet, perhaps it got dropped but it might also # be still in flight. Closing the socket (= sending FIN) ourselves. # Not waiting here for its ACK however, so no `closeWait`. Underneath the # socket will still wait for the FIN-ACK (or timeout) before it destroys the # socket. socket.close() # TODO: This could currently create a backlog of content items to be validated # as `AcceptConnectionCallback` is `asyncSpawn`'ed and there are no limits # on the `contentOffers`. Might move the queue to before the reading of the # socket, and let the specific networks handle that. await stream.contentQueue.put( (Opt.some(offer.nodeId), offer.contentKeys, contentItems) ) proc new( T: type PortalStream, transport: UtpDiscv5Protocol, contentQueue: AsyncQueue[(Opt[NodeId], ContentKeysList, seq[seq[byte]])], connectionTimeout: Duration, contentReadTimeout: Duration, rng: ref HmacDrbgContext, ): T = let stream = PortalStream( transport: transport, connectionTimeout: connectionTimeout, contentReadTimeout: contentReadTimeout, contentQueue: contentQueue, rng: rng, ) stream proc allowedConnection( stream: PortalStream, address: NodeAddress, connectionId: uint16 ): bool = return stream.contentRequests.any( proc(x: ContentRequest): bool = x.connectionId == connectionId and x.nodeId == address.nodeId ) or stream.contentOffers.any( proc(x: ContentOffer): bool = x.connectionId == connectionId and x.nodeId == address.nodeId ) proc handleIncomingConnection( server: UtpRouter[NodeAddress], socket: UtpSocket[NodeAddress] ): Future[void] {.async: (raw: true, raises: []).} = let manager = getUserData[NodeAddress, StreamManager](server) for stream in manager.streams: # Note: Connection id of uTP SYN is different from other packets, it is # actually the peers `send_conn_id`, opposed to `receive_conn_id` for all # other packets. for i, request in stream.contentRequests: if request.connectionId == socket.connectionId and request.nodeId == socket.remoteAddress.nodeId: let fut = socket.writeContentRequest(stream, request) stream.contentRequests.del(i) return noCancel(fut) for i, offer in stream.contentOffers: if offer.connectionId == socket.connectionId and offer.nodeId == socket.remoteAddress.nodeId: let fut = socket.readContentOffer(stream, offer) stream.contentOffers.del(i) return noCancel(fut) # TODO: Is there a scenario where this can happen, # considering `allowRegisteredIdCallback`? If not, doAssert? var fut = newFuture[void]("fluffy.AcceptConnectionCallback") fut.complete() return fut proc allowIncomingConnection( r: UtpRouter[NodeAddress], remoteAddress: NodeAddress, connectionId: uint16 ): bool = let manager = getUserData[NodeAddress, StreamManager](r) for stream in manager.streams: # stream.pruneAllowedConnections() if allowedConnection(stream, remoteAddress, connectionId): return true proc new*(T: type StreamManager, d: protocol.Protocol): T = let socketConfig = SocketConfig.init( # Setting to none means that incoming sockets are in Connected state, which # means they can send and receive data. incomingSocketReceiveTimeout = none(Duration), payloadSize = uint32(maxUtpPayloadSize), ) manager = StreamManager(streams: @[], rng: d.rng) utpOverDiscV5Protocol = UtpDiscv5Protocol.new( d, utpProtocolId, handleIncomingConnection, manager, allowIncomingConnection, socketConfig, ) manager.transport = utpOverDiscV5Protocol return manager proc registerNewStream*( m: StreamManager, contentQueue: AsyncQueue[(Opt[NodeId], ContentKeysList, seq[seq[byte]])], connectionTimeout = defaultConnectionTimeout, contentReadTimeout = defaultContentReadTimeout, ): PortalStream = let s = PortalStream.new( m.transport, contentQueue, connectionTimeout, contentReadTimeout, m.rng ) m.streams.add(s) return s