2018-05-16 08:22:34 +00:00
|
|
|
#
|
|
|
|
# Asyncdispatch2 Stream Transport
|
|
|
|
# (c) Copyright 2018
|
|
|
|
# Status Research & Development GmbH
|
|
|
|
#
|
|
|
|
# Licensed under either of
|
|
|
|
# Apache License, version 2.0, (LICENSE-APACHEv2)
|
|
|
|
# MIT license (LICENSE-MIT)
|
|
|
|
|
2018-06-04 09:57:17 +00:00
|
|
|
import net, nativesockets, os, deques, strutils
|
2018-05-21 21:52:57 +00:00
|
|
|
import ../asyncloop, ../asyncsync, ../handles, ../sendfile
|
2018-05-16 08:22:34 +00:00
|
|
|
import common
|
|
|
|
|
|
|
|
when defined(windows):
|
|
|
|
import winlean
|
|
|
|
else:
|
|
|
|
import posix
|
|
|
|
|
|
|
|
type
|
2018-05-21 21:52:57 +00:00
|
|
|
VectorKind = enum
|
2018-05-22 21:03:13 +00:00
|
|
|
DataBuffer, # Simple buffer pointer/length
|
|
|
|
DataFile # File handle for sendfile/TransmitFile
|
2018-05-21 21:52:57 +00:00
|
|
|
|
|
|
|
type
|
|
|
|
StreamVector = object
|
2018-05-22 21:03:13 +00:00
|
|
|
kind: VectorKind # Writer vector source kind
|
|
|
|
buf: pointer # Writer buffer pointer
|
|
|
|
buflen: int # Writer buffer size
|
|
|
|
offset: uint # Writer vector offset
|
|
|
|
writer: Future[void] # Writer vector completion Future
|
2018-05-21 21:52:57 +00:00
|
|
|
|
2018-05-16 08:22:34 +00:00
|
|
|
TransportKind* {.pure.} = enum
|
2018-05-22 21:03:13 +00:00
|
|
|
Socket, # Socket transport
|
|
|
|
Pipe, # Pipe transport
|
|
|
|
File # File transport
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
type
|
|
|
|
StreamTransport* = ref object of RootRef
|
2018-05-23 11:03:18 +00:00
|
|
|
fd*: AsyncFD # File descriptor
|
2018-05-16 08:22:34 +00:00
|
|
|
state: set[TransportState] # Current Transport state
|
|
|
|
reader: Future[void] # Current reader Future
|
2018-05-29 18:04:11 +00:00
|
|
|
# ZAH: I'm not quite certain, but it seems to me that the intermediate
|
|
|
|
# buffer is not necessary. The receiving code needs to know how to grow
|
|
|
|
# the output buffer of the future attached to the read operation. If this
|
|
|
|
# is the case, the buffering can be replaced with direct writing to this
|
|
|
|
# output buffer. Furthermore, we'll be able to signal additional 'progress'
|
|
|
|
# events for the future to make the API more complete.
|
2018-05-16 08:22:34 +00:00
|
|
|
buffer: seq[byte] # Reading buffer
|
|
|
|
offset: int # Reading buffer offset
|
|
|
|
error: ref Exception # Current error
|
|
|
|
queue: Deque[StreamVector] # Writer queue
|
|
|
|
future: Future[void] # Stream life future
|
2018-05-21 21:52:57 +00:00
|
|
|
transferred: int
|
2018-05-16 08:22:34 +00:00
|
|
|
case kind*: TransportKind
|
|
|
|
of TransportKind.Socket:
|
|
|
|
domain: Domain # Socket transport domain (IPv4/IPv6)
|
|
|
|
local: TransportAddress # Local address
|
|
|
|
remote: TransportAddress # Remote address
|
|
|
|
of TransportKind.Pipe:
|
2018-05-21 21:52:57 +00:00
|
|
|
todo1: int
|
2018-05-16 08:22:34 +00:00
|
|
|
of TransportKind.File:
|
2018-05-21 21:52:57 +00:00
|
|
|
todo2: int
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-05-22 22:29:07 +00:00
|
|
|
StreamCallback* = proc(server: StreamServer,
|
|
|
|
client: StreamTransport,
|
2018-05-16 08:22:34 +00:00
|
|
|
udata: pointer): Future[void] {.gcsafe.}
|
2018-05-28 23:35:15 +00:00
|
|
|
## New remote client connection callback
|
|
|
|
## ``server`` - StreamServer object.
|
|
|
|
## ``client`` - accepted client transport.
|
|
|
|
## ``udata`` - user-defined pointer passed at ``createStreamServer()`` call.
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
StreamServer* = ref object of SocketServer
|
|
|
|
function*: StreamCallback
|
|
|
|
|
|
|
|
proc remoteAddress*(transp: StreamTransport): TransportAddress =
|
|
|
|
## Returns ``transp`` remote socket address.
|
|
|
|
if transp.kind != TransportKind.Socket:
|
|
|
|
raise newException(TransportError, "Socket required!")
|
|
|
|
if transp.remote.port == Port(0):
|
|
|
|
var saddr: Sockaddr_storage
|
|
|
|
var slen = SockLen(sizeof(saddr))
|
|
|
|
if getpeername(SocketHandle(transp.fd), cast[ptr SockAddr](addr saddr),
|
|
|
|
addr slen) != 0:
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
fromSockAddr(saddr, slen, transp.remote.address, transp.remote.port)
|
|
|
|
result = transp.remote
|
|
|
|
|
|
|
|
proc localAddress*(transp: StreamTransport): TransportAddress =
|
|
|
|
## Returns ``transp`` local socket address.
|
|
|
|
if transp.kind != TransportKind.Socket:
|
|
|
|
raise newException(TransportError, "Socket required!")
|
|
|
|
if transp.local.port == Port(0):
|
|
|
|
var saddr: Sockaddr_storage
|
|
|
|
var slen = SockLen(sizeof(saddr))
|
|
|
|
if getsockname(SocketHandle(transp.fd), cast[ptr SockAddr](addr saddr),
|
|
|
|
addr slen) != 0:
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
fromSockAddr(saddr, slen, transp.local.address, transp.local.port)
|
|
|
|
result = transp.local
|
|
|
|
|
|
|
|
template setReadError(t, e: untyped) =
|
|
|
|
(t).state.incl(ReadError)
|
|
|
|
(t).error = newException(TransportOsError, osErrorMsg((e)))
|
|
|
|
|
|
|
|
template setWriteError(t, e: untyped) =
|
|
|
|
(t).state.incl(WriteError)
|
|
|
|
(t).error = newException(TransportOsError, osErrorMsg((e)))
|
|
|
|
|
|
|
|
template finishReader(t: untyped) =
|
|
|
|
var reader = (t).reader
|
|
|
|
reader.complete()
|
|
|
|
|
|
|
|
template checkPending(t: untyped) =
|
|
|
|
if not isNil((t).reader):
|
|
|
|
raise newException(TransportError, "Read operation already pending!")
|
|
|
|
|
|
|
|
template shiftBuffer(t, c: untyped) =
|
|
|
|
if (t).offset > c:
|
|
|
|
moveMem(addr((t).buffer[0]), addr((t).buffer[(c)]), (t).offset - (c))
|
|
|
|
(t).offset = (t).offset - (c)
|
|
|
|
else:
|
|
|
|
(t).offset = 0
|
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
template shiftVectorBuffer(v, o: untyped) =
|
|
|
|
(v).buf = cast[pointer](cast[uint]((v).buf) + uint(o))
|
|
|
|
(v).buflen -= int(o)
|
|
|
|
|
|
|
|
template shiftVectorFile(v, o: untyped) =
|
|
|
|
(v).buf = cast[pointer](cast[uint]((v).buf) - cast[uint](o))
|
|
|
|
(v).offset += cast[uint]((o))
|
|
|
|
|
2018-05-16 08:22:34 +00:00
|
|
|
when defined(windows):
|
|
|
|
import winlean
|
|
|
|
type
|
|
|
|
WindowsStreamTransport = ref object of StreamTransport
|
2018-05-21 21:52:57 +00:00
|
|
|
rwsabuf: TWSABuf # Reader WSABUF
|
|
|
|
wwsabuf: TWSABuf # Writer WSABUF
|
2018-05-16 08:22:34 +00:00
|
|
|
rovl: CustomOverlapped # Reader OVERLAPPED structure
|
|
|
|
wovl: CustomOverlapped # Writer OVERLAPPED structure
|
|
|
|
roffset: int # Pending reading offset
|
|
|
|
|
|
|
|
const SO_UPDATE_CONNECT_CONTEXT = 0x7010
|
|
|
|
|
|
|
|
template finishWriter(t: untyped) =
|
|
|
|
var vv = (t).queue.popFirst()
|
|
|
|
vv.writer.complete()
|
|
|
|
|
|
|
|
template zeroOvelappedOffset(t: untyped) =
|
|
|
|
(t).offset = 0
|
|
|
|
(t).offsetHigh = 0
|
|
|
|
|
|
|
|
template setOverlappedOffset(t, o: untyped) =
|
|
|
|
(t).offset = cast[int32](cast[uint64](o) and 0xFFFFFFFF'u64)
|
|
|
|
(t).offsetHigh = cast[int32](cast[uint64](o) shr 32)
|
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
template getFileSize(v: untyped): uint =
|
|
|
|
cast[uint]((v).buf)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
template getFileHandle(v: untyped): Handle =
|
|
|
|
cast[Handle]((v).buflen)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
template slideBuffer(t, o: untyped) =
|
2018-05-21 21:52:57 +00:00
|
|
|
(t).wwsabuf.buf = cast[cstring](cast[uint]((t).wwsabuf.buf) + uint(o))
|
|
|
|
(t).wwsabuf.len -= int32(o)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
template setReaderWSABuffer(t: untyped) =
|
|
|
|
(t).rwsabuf.buf = cast[cstring](
|
2018-05-16 08:22:34 +00:00
|
|
|
cast[uint](addr t.buffer[0]) + uint((t).roffset))
|
2018-05-21 21:52:57 +00:00
|
|
|
(t).rwsabuf.len = int32(len((t).buffer) - (t).roffset)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
template setWriterWSABuffer(t, v: untyped) =
|
|
|
|
(t).wwsabuf.buf = cast[cstring](v.buf)
|
|
|
|
(t).wwsabuf.len = cast[int32](v.buflen)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-05-22 16:27:20 +00:00
|
|
|
proc writeStreamLoop(udata: pointer) {.gcsafe, nimcall.} =
|
2018-05-16 08:22:34 +00:00
|
|
|
var bytesCount: int32
|
2018-05-21 21:52:57 +00:00
|
|
|
var ovl = cast[PtrCustomOverlapped](udata)
|
2018-05-16 08:22:34 +00:00
|
|
|
var transp = cast[WindowsStreamTransport](ovl.data.udata)
|
|
|
|
|
|
|
|
while len(transp.queue) > 0:
|
|
|
|
if WritePending in transp.state:
|
|
|
|
## Continuation
|
|
|
|
transp.state.excl(WritePending)
|
|
|
|
let err = transp.wovl.data.errCode
|
|
|
|
if err == OSErrorCode(-1):
|
|
|
|
bytesCount = transp.wovl.data.bytesCount
|
|
|
|
var vector = transp.queue.popFirst()
|
|
|
|
if bytesCount == 0:
|
|
|
|
vector.writer.complete()
|
|
|
|
else:
|
|
|
|
if transp.kind == TransportKind.Socket:
|
|
|
|
if vector.kind == VectorKind.DataBuffer:
|
2018-05-21 21:52:57 +00:00
|
|
|
if bytesCount < transp.wwsabuf.len:
|
|
|
|
vector.shiftVectorBuffer(bytesCount)
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.queue.addFirst(vector)
|
|
|
|
else:
|
|
|
|
vector.writer.complete()
|
|
|
|
else:
|
|
|
|
if uint(bytesCount) < getFileSize(vector):
|
2018-05-21 21:52:57 +00:00
|
|
|
vector.shiftVectorFile(bytesCount)
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.queue.addFirst(vector)
|
|
|
|
else:
|
|
|
|
vector.writer.complete()
|
2018-05-28 23:35:15 +00:00
|
|
|
elif int(err) == ERROR_OPERATION_ABORTED:
|
|
|
|
# CancelIO() interrupt
|
|
|
|
transp.finishWriter()
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
|
|
|
transp.setWriteError(err)
|
|
|
|
transp.finishWriter()
|
|
|
|
else:
|
|
|
|
## Initiation
|
|
|
|
transp.state.incl(WritePending)
|
|
|
|
if transp.kind == TransportKind.Socket:
|
|
|
|
let sock = SocketHandle(transp.wovl.data.fd)
|
2018-05-16 15:28:23 +00:00
|
|
|
var vector = transp.queue.popFirst()
|
|
|
|
if vector.kind == VectorKind.DataBuffer:
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.wovl.zeroOvelappedOffset()
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.setWriterWSABuffer(vector)
|
|
|
|
let ret = WSASend(sock, addr transp.wwsabuf, 1,
|
2018-05-16 08:22:34 +00:00
|
|
|
addr bytesCount, DWORD(0),
|
|
|
|
cast[POVERLAPPED](addr transp.wovl), nil)
|
|
|
|
if ret != 0:
|
|
|
|
let err = osLastError()
|
2018-05-16 15:28:23 +00:00
|
|
|
if int(err) == ERROR_OPERATION_ABORTED:
|
2018-05-28 23:35:15 +00:00
|
|
|
# CancelIO() interrupt
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.state.excl(WritePending)
|
2018-05-16 15:28:23 +00:00
|
|
|
transp.state.incl(WritePaused)
|
|
|
|
elif int(err) == ERROR_IO_PENDING:
|
|
|
|
transp.queue.addFirst(vector)
|
|
|
|
else:
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.state.excl(WritePending)
|
|
|
|
transp.setWriteError(err)
|
2018-05-21 21:52:57 +00:00
|
|
|
vector.writer.complete()
|
2018-05-16 15:28:23 +00:00
|
|
|
else:
|
|
|
|
transp.queue.addFirst(vector)
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
|
|
|
let loop = getGlobalDispatcher()
|
|
|
|
var size: int32
|
|
|
|
var flags: int32
|
|
|
|
|
|
|
|
if getFileSize(vector) > 2_147_483_646'u:
|
|
|
|
size = 2_147_483_646
|
|
|
|
else:
|
|
|
|
size = int32(getFileSize(vector))
|
|
|
|
|
|
|
|
transp.wovl.setOverlappedOffset(vector.offset)
|
|
|
|
var ret = loop.transmitFile(sock, getFileHandle(vector), size, 0,
|
|
|
|
cast[POVERLAPPED](addr transp.wovl),
|
|
|
|
nil, flags)
|
|
|
|
if ret == 0:
|
|
|
|
let err = osLastError()
|
2018-05-16 15:28:23 +00:00
|
|
|
if int(err) == ERROR_OPERATION_ABORTED:
|
2018-05-28 23:35:15 +00:00
|
|
|
# CancelIO() interrupt
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.state.excl(WritePending)
|
2018-05-16 15:28:23 +00:00
|
|
|
transp.state.incl(WritePaused)
|
|
|
|
elif int(err) == ERROR_IO_PENDING:
|
|
|
|
transp.queue.addFirst(vector)
|
|
|
|
else:
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.state.excl(WritePending)
|
|
|
|
transp.setWriteError(err)
|
2018-05-21 21:52:57 +00:00
|
|
|
vector.writer.complete()
|
2018-05-16 15:28:23 +00:00
|
|
|
else:
|
|
|
|
transp.queue.addFirst(vector)
|
2018-05-16 08:22:34 +00:00
|
|
|
break
|
|
|
|
|
|
|
|
if len(transp.queue) == 0:
|
|
|
|
transp.state.incl(WritePaused)
|
|
|
|
|
2018-05-22 16:27:20 +00:00
|
|
|
proc readStreamLoop(udata: pointer) {.gcsafe, nimcall.} =
|
2018-05-21 21:52:57 +00:00
|
|
|
var ovl = cast[PtrCustomOverlapped](udata)
|
2018-05-16 08:22:34 +00:00
|
|
|
var transp = cast[WindowsStreamTransport](ovl.data.udata)
|
|
|
|
|
|
|
|
while true:
|
|
|
|
if ReadPending in transp.state:
|
|
|
|
## Continuation
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.state.excl(ReadPending)
|
2018-05-16 08:22:34 +00:00
|
|
|
if ReadClosed in transp.state:
|
|
|
|
break
|
|
|
|
let err = transp.rovl.data.errCode
|
|
|
|
if err == OSErrorCode(-1):
|
|
|
|
let bytesCount = transp.rovl.data.bytesCount
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.transferred += bytesCount
|
2018-05-16 08:22:34 +00:00
|
|
|
if bytesCount == 0:
|
|
|
|
transp.state.incl(ReadEof)
|
|
|
|
transp.state.incl(ReadPaused)
|
|
|
|
else:
|
|
|
|
if transp.offset != transp.roffset:
|
|
|
|
moveMem(addr transp.buffer[transp.offset],
|
|
|
|
addr transp.buffer[transp.roffset],
|
|
|
|
bytesCount)
|
|
|
|
transp.offset += bytesCount
|
|
|
|
transp.roffset = transp.offset
|
|
|
|
if transp.offset == len(transp.buffer):
|
|
|
|
transp.state.incl(ReadPaused)
|
2018-05-28 23:35:15 +00:00
|
|
|
elif int(err) == ERROR_OPERATION_ABORTED:
|
|
|
|
# CancelIO() interrupt
|
|
|
|
discard
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
|
|
|
transp.setReadError(err)
|
|
|
|
if not isNil(transp.reader):
|
|
|
|
transp.finishReader()
|
2018-05-21 21:52:57 +00:00
|
|
|
if ReadPaused in transp.state:
|
|
|
|
# Transport buffer is full, so we will not continue on reading.
|
|
|
|
break
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
|
|
|
## Initiation
|
2018-05-21 21:52:57 +00:00
|
|
|
if transp.state * {ReadEof, ReadClosed, ReadError} == {}:
|
2018-05-16 08:22:34 +00:00
|
|
|
var flags = DWORD(0)
|
|
|
|
var bytesCount: int32 = 0
|
|
|
|
transp.state.excl(ReadPaused)
|
|
|
|
transp.state.incl(ReadPending)
|
|
|
|
if transp.kind == TransportKind.Socket:
|
|
|
|
let sock = SocketHandle(transp.rovl.data.fd)
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.roffset = transp.offset
|
|
|
|
transp.setReaderWSABuffer()
|
|
|
|
let ret = WSARecv(sock, addr transp.rwsabuf, 1,
|
2018-05-16 08:22:34 +00:00
|
|
|
addr bytesCount, addr flags,
|
|
|
|
cast[POVERLAPPED](addr transp.rovl), nil)
|
|
|
|
if ret != 0:
|
|
|
|
let err = osLastError()
|
|
|
|
if int(err) == ERROR_OPERATION_ABORTED:
|
2018-05-28 23:35:15 +00:00
|
|
|
# CancelIO() interrupt
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.state.excl(ReadPending)
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.state.incl(ReadPaused)
|
2018-06-04 09:57:17 +00:00
|
|
|
elif int32(err) in {WSAECONNRESET, WSAENETRESET}:
|
|
|
|
if not isNil(transp.reader):
|
|
|
|
transp.state = {ReadEof, ReadPaused}
|
|
|
|
transp.finishReader()
|
2018-05-16 08:22:34 +00:00
|
|
|
elif int32(err) != ERROR_IO_PENDING:
|
2018-06-04 09:57:17 +00:00
|
|
|
transp.state.excl(ReadPending)
|
|
|
|
transp.state.incl(ReadPaused)
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.setReadError(err)
|
|
|
|
if not isNil(transp.reader):
|
|
|
|
transp.finishReader()
|
|
|
|
## Finish Loop
|
|
|
|
break
|
|
|
|
|
|
|
|
proc newStreamSocketTransport(sock: AsyncFD, bufsize: int): StreamTransport =
|
2018-05-21 21:52:57 +00:00
|
|
|
var transp = WindowsStreamTransport(kind: TransportKind.Socket)
|
|
|
|
transp.fd = sock
|
|
|
|
transp.rovl.data = CompletionData(fd: sock, cb: readStreamLoop,
|
|
|
|
udata: cast[pointer](transp))
|
|
|
|
transp.wovl.data = CompletionData(fd: sock, cb: writeStreamLoop,
|
|
|
|
udata: cast[pointer](transp))
|
|
|
|
transp.buffer = newSeq[byte](bufsize)
|
|
|
|
transp.state = {ReadPaused, WritePaused}
|
|
|
|
transp.queue = initDeque[StreamVector]()
|
|
|
|
transp.future = newFuture[void]("stream.socket.transport")
|
2018-05-29 18:04:11 +00:00
|
|
|
# ZAH: If these objects are going to be manually managed, why do we bother
|
|
|
|
# with using the GC at all? It's better to rely on a destructor. If someone
|
|
|
|
# wants to share a Transport reference, they can still create a GC-managed
|
|
|
|
# wrapping object.
|
2018-05-21 21:52:57 +00:00
|
|
|
GC_ref(transp)
|
|
|
|
result = cast[StreamTransport](transp)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc bindToDomain(handle: AsyncFD, domain: Domain): bool =
|
|
|
|
result = true
|
|
|
|
if domain == Domain.AF_INET6:
|
|
|
|
var saddr: Sockaddr_in6
|
|
|
|
saddr.sin6_family = int16(toInt(domain))
|
|
|
|
if bindAddr(SocketHandle(handle), cast[ptr SockAddr](addr(saddr)),
|
|
|
|
sizeof(saddr).SockLen) != 0'i32:
|
|
|
|
result = false
|
|
|
|
else:
|
|
|
|
var saddr: Sockaddr_in
|
|
|
|
saddr.sin_family = int16(toInt(domain))
|
|
|
|
if bindAddr(SocketHandle(handle), cast[ptr SockAddr](addr(saddr)),
|
|
|
|
sizeof(saddr).SockLen) != 0'i32:
|
|
|
|
result = false
|
|
|
|
|
|
|
|
proc connect*(address: TransportAddress,
|
|
|
|
bufferSize = DefaultStreamBufferSize): Future[StreamTransport] =
|
2018-05-27 05:49:47 +00:00
|
|
|
## Open new connection to remote peer with address ``address`` and create
|
|
|
|
## new transport object ``StreamTransport`` for established connection.
|
2018-05-28 23:35:15 +00:00
|
|
|
## ``bufferSize`` is size of internal buffer for transport.
|
2018-05-16 08:22:34 +00:00
|
|
|
let loop = getGlobalDispatcher()
|
|
|
|
var
|
|
|
|
saddr: Sockaddr_storage
|
|
|
|
slen: SockLen
|
|
|
|
sock: AsyncFD
|
|
|
|
povl: RefCustomOverlapped
|
|
|
|
|
|
|
|
var retFuture = newFuture[StreamTransport]("stream.transport.connect")
|
|
|
|
toSockAddr(address.address, address.port, saddr, slen)
|
|
|
|
sock = createAsyncSocket(address.address.getDomain(), SockType.SOCK_STREAM,
|
|
|
|
Protocol.IPPROTO_TCP)
|
|
|
|
if sock == asyncInvalidSocket:
|
|
|
|
result.fail(newException(OSError, osErrorMsg(osLastError())))
|
|
|
|
|
|
|
|
if not bindToDomain(sock, address.address.getDomain()):
|
|
|
|
sock.closeAsyncSocket()
|
|
|
|
result.fail(newException(OSError, osErrorMsg(osLastError())))
|
|
|
|
|
|
|
|
proc continuation(udata: pointer) =
|
2018-05-22 16:27:20 +00:00
|
|
|
var ovl = cast[RefCustomOverlapped](udata)
|
2018-05-16 08:22:34 +00:00
|
|
|
if not retFuture.finished:
|
|
|
|
if ovl.data.errCode == OSErrorCode(-1):
|
|
|
|
if setsockopt(SocketHandle(sock), cint(SOL_SOCKET),
|
|
|
|
cint(SO_UPDATE_CONNECT_CONTEXT), nil,
|
|
|
|
SockLen(0)) != 0'i32:
|
|
|
|
sock.closeAsyncSocket()
|
|
|
|
retFuture.fail(newException(OSError, osErrorMsg(osLastError())))
|
|
|
|
else:
|
|
|
|
retFuture.complete(newStreamSocketTransport(povl.data.fd,
|
|
|
|
bufferSize))
|
|
|
|
else:
|
|
|
|
sock.closeAsyncSocket()
|
|
|
|
retFuture.fail(newException(OSError, osErrorMsg(ovl.data.errCode)))
|
2018-05-22 16:27:20 +00:00
|
|
|
GC_unref(ovl)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
povl = RefCustomOverlapped()
|
2018-05-22 16:27:20 +00:00
|
|
|
GC_ref(povl)
|
2018-05-16 08:22:34 +00:00
|
|
|
povl.data = CompletionData(fd: sock, cb: continuation)
|
|
|
|
var res = loop.connectEx(SocketHandle(sock),
|
|
|
|
cast[ptr SockAddr](addr saddr),
|
|
|
|
DWORD(slen), nil, 0, nil,
|
|
|
|
cast[POVERLAPPED](povl))
|
|
|
|
# We will not process immediate completion, to avoid undefined behavior.
|
|
|
|
if not res:
|
|
|
|
let err = osLastError()
|
|
|
|
if int32(err) != ERROR_IO_PENDING:
|
2018-05-22 16:27:20 +00:00
|
|
|
GC_unref(povl)
|
2018-05-16 08:22:34 +00:00
|
|
|
sock.closeAsyncSocket()
|
|
|
|
retFuture.fail(newException(OSError, osErrorMsg(err)))
|
|
|
|
return retFuture
|
|
|
|
|
2018-06-04 09:57:17 +00:00
|
|
|
proc acceptLoop(udata: pointer) {.gcsafe, nimcall.} =
|
|
|
|
var ovl = cast[PtrCustomOverlapped](udata)
|
|
|
|
var server = cast[StreamServer](ovl.data.udata)
|
|
|
|
var loop = getGlobalDispatcher()
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-06-04 09:57:17 +00:00
|
|
|
while true:
|
|
|
|
if server.apending:
|
|
|
|
## Continuation
|
|
|
|
server.apending = false
|
|
|
|
if server.status == ServerStatus.Stopped:
|
|
|
|
server.asock.closeAsyncSocket()
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
|
|
|
if ovl.data.errCode == OSErrorCode(-1):
|
2018-06-04 09:57:17 +00:00
|
|
|
if setsockopt(SocketHandle(server.asock), cint(SOL_SOCKET),
|
2018-05-16 08:22:34 +00:00
|
|
|
cint(SO_UPDATE_ACCEPT_CONTEXT),
|
2018-06-04 09:57:17 +00:00
|
|
|
addr server.sock,
|
2018-05-16 08:22:34 +00:00
|
|
|
SockLen(sizeof(SocketHandle))) != 0'i32:
|
2018-06-04 09:57:17 +00:00
|
|
|
server.asock.closeAsyncSocket()
|
|
|
|
raiseOsError(osLastError())
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
2018-06-04 09:57:17 +00:00
|
|
|
discard server.function(server,
|
|
|
|
newStreamSocketTransport(server.asock, server.bufferSize),
|
|
|
|
server.udata)
|
2018-05-23 11:03:18 +00:00
|
|
|
elif int32(ovl.data.errCode) == ERROR_OPERATION_ABORTED:
|
2018-05-28 23:35:15 +00:00
|
|
|
# CancelIO() interrupt
|
2018-06-04 09:57:17 +00:00
|
|
|
server.asock.closeAsyncSocket()
|
|
|
|
break
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
2018-06-04 09:57:17 +00:00
|
|
|
server.asock.closeAsyncSocket()
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
else:
|
|
|
|
## Initiation
|
|
|
|
server.apending = true
|
|
|
|
server.asock = createAsyncSocket(server.domain, SockType.SOCK_STREAM,
|
|
|
|
Protocol.IPPROTO_TCP)
|
|
|
|
if server.asock == asyncInvalidSocket:
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
|
|
|
|
var dwBytesReceived = DWORD(0)
|
|
|
|
let dwReceiveDataLength = DWORD(0)
|
|
|
|
let dwLocalAddressLength = DWORD(sizeof(Sockaddr_in6) + 16)
|
|
|
|
let dwRemoteAddressLength = DWORD(sizeof(Sockaddr_in6) + 16)
|
|
|
|
|
|
|
|
let res = loop.acceptEx(SocketHandle(server.sock),
|
|
|
|
SocketHandle(server.asock),
|
|
|
|
addr server.abuffer[0],
|
|
|
|
dwReceiveDataLength, dwLocalAddressLength,
|
|
|
|
dwRemoteAddressLength, addr dwBytesReceived,
|
|
|
|
cast[POVERLAPPED](addr server.aovl))
|
|
|
|
if not res:
|
|
|
|
let err = osLastError()
|
|
|
|
if int32(err) == ERROR_OPERATION_ABORTED:
|
|
|
|
server.apending = false
|
|
|
|
break
|
|
|
|
elif int32(err) == ERROR_IO_PENDING:
|
|
|
|
discard
|
|
|
|
else:
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
break
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc resumeRead(transp: StreamTransport) {.inline.} =
|
|
|
|
var wtransp = cast[WindowsStreamTransport](transp)
|
|
|
|
wtransp.state.excl(ReadPaused)
|
|
|
|
readStreamLoop(cast[pointer](addr wtransp.rovl))
|
|
|
|
|
|
|
|
proc resumeWrite(transp: StreamTransport) {.inline.} =
|
|
|
|
var wtransp = cast[WindowsStreamTransport](transp)
|
|
|
|
wtransp.state.excl(WritePaused)
|
|
|
|
writeStreamLoop(cast[pointer](addr wtransp.wovl))
|
|
|
|
|
2018-06-04 09:57:17 +00:00
|
|
|
proc pauseAccept(server: SocketServer) {.inline.} =
|
|
|
|
if server.apending:
|
|
|
|
discard cancelIO(Handle(server.sock))
|
|
|
|
|
|
|
|
proc resumeAccept(server: SocketServer) {.inline.} =
|
|
|
|
if not server.apending:
|
|
|
|
acceptLoop(cast[pointer](addr server.aovl))
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
|
|
|
import posix
|
|
|
|
|
|
|
|
type
|
|
|
|
UnixStreamTransport* = ref object of StreamTransport
|
|
|
|
|
|
|
|
template getVectorBuffer(v: untyped): pointer =
|
|
|
|
cast[pointer](cast[uint]((v).buf) + uint((v).boffset))
|
|
|
|
|
|
|
|
template getVectorLength(v: untyped): int =
|
|
|
|
cast[int]((v).buflen - int((v).boffset))
|
|
|
|
|
|
|
|
template initBufferStreamVector(v, p, n, t: untyped) =
|
|
|
|
(v).kind = DataBuffer
|
|
|
|
(v).buf = cast[pointer]((p))
|
|
|
|
(v).buflen = int(n)
|
|
|
|
(v).writer = (t)
|
|
|
|
|
|
|
|
proc writeStreamLoop(udata: pointer) {.gcsafe.} =
|
|
|
|
var cdata = cast[ptr CompletionData](udata)
|
2018-05-29 11:22:53 +00:00
|
|
|
if not isNil(cdata) and int(cdata.fd) == 0:
|
2018-05-29 09:59:39 +00:00
|
|
|
# Transport was closed earlier, exiting
|
|
|
|
return
|
2018-05-16 08:22:34 +00:00
|
|
|
var transp = cast[UnixStreamTransport](cdata.udata)
|
|
|
|
let fd = SocketHandle(cdata.fd)
|
2018-05-28 23:35:15 +00:00
|
|
|
if len(transp.queue) > 0:
|
|
|
|
var vector = transp.queue.popFirst()
|
|
|
|
while true:
|
|
|
|
if transp.kind == TransportKind.Socket:
|
|
|
|
if vector.kind == VectorKind.DataBuffer:
|
|
|
|
let res = posix.send(fd, vector.buf, vector.buflen, MSG_NOSIGNAL)
|
|
|
|
if res >= 0:
|
|
|
|
if vector.buflen - res == 0:
|
|
|
|
vector.writer.complete()
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
2018-05-28 23:35:15 +00:00
|
|
|
vector.shiftVectorBuffer(res)
|
|
|
|
transp.queue.addFirst(vector)
|
2018-05-16 08:22:34 +00:00
|
|
|
else:
|
2018-05-28 23:35:15 +00:00
|
|
|
let err = osLastError()
|
|
|
|
if int(err) == EINTR:
|
|
|
|
continue
|
2018-05-21 21:52:57 +00:00
|
|
|
else:
|
2018-05-28 23:35:15 +00:00
|
|
|
transp.setWriteError(err)
|
|
|
|
vector.writer.complete()
|
|
|
|
else:
|
|
|
|
let res = sendfile(int(fd), cast[int](vector.buflen),
|
|
|
|
int(vector.offset),
|
|
|
|
cast[int](vector.buf))
|
|
|
|
if res >= 0:
|
|
|
|
if cast[int](vector.buf) - res == 0:
|
|
|
|
vector.writer.complete()
|
|
|
|
else:
|
|
|
|
vector.shiftVectorFile(res)
|
|
|
|
transp.queue.addFirst(vector)
|
|
|
|
else:
|
|
|
|
let err = osLastError()
|
|
|
|
if int(err) == EINTR:
|
|
|
|
continue
|
|
|
|
else:
|
|
|
|
transp.setWriteError(err)
|
|
|
|
vector.writer.complete()
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
transp.state.incl(WritePaused)
|
|
|
|
transp.fd.removeWriter()
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc readStreamLoop(udata: pointer) {.gcsafe.} =
|
|
|
|
var cdata = cast[ptr CompletionData](udata)
|
2018-05-29 11:22:53 +00:00
|
|
|
if not isNil(cdata) and int(cdata.fd) == 0:
|
2018-05-29 09:59:39 +00:00
|
|
|
# Transport was closed earlier, exiting
|
|
|
|
return
|
2018-05-16 08:22:34 +00:00
|
|
|
var transp = cast[UnixStreamTransport](cdata.udata)
|
|
|
|
let fd = SocketHandle(cdata.fd)
|
2018-05-28 23:35:15 +00:00
|
|
|
while true:
|
|
|
|
var res = posix.recv(fd, addr transp.buffer[transp.offset],
|
|
|
|
len(transp.buffer) - transp.offset, cint(0))
|
|
|
|
if res < 0:
|
|
|
|
let err = osLastError()
|
|
|
|
if int(err) == EINTR:
|
|
|
|
continue
|
|
|
|
elif int(err) in {ECONNRESET}:
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.state.incl(ReadEof)
|
|
|
|
transp.state.incl(ReadPaused)
|
|
|
|
cdata.fd.removeReader()
|
|
|
|
else:
|
2018-05-28 23:35:15 +00:00
|
|
|
transp.setReadError(err)
|
|
|
|
cdata.fd.removeReader()
|
|
|
|
elif res == 0:
|
|
|
|
transp.state.incl(ReadEof)
|
|
|
|
transp.state.incl(ReadPaused)
|
|
|
|
cdata.fd.removeReader()
|
|
|
|
else:
|
|
|
|
transp.offset += res
|
|
|
|
if transp.offset == len(transp.buffer):
|
|
|
|
transp.state.incl(ReadPaused)
|
|
|
|
cdata.fd.removeReader()
|
|
|
|
if not isNil(transp.reader):
|
|
|
|
transp.finishReader()
|
|
|
|
break
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc newStreamSocketTransport(sock: AsyncFD, bufsize: int): StreamTransport =
|
2018-05-21 21:52:57 +00:00
|
|
|
var transp = UnixStreamTransport(kind: TransportKind.Socket)
|
|
|
|
transp.fd = sock
|
|
|
|
transp.buffer = newSeq[byte](bufsize)
|
|
|
|
transp.state = {ReadPaused, WritePaused}
|
|
|
|
transp.queue = initDeque[StreamVector]()
|
|
|
|
transp.future = newFuture[void]("socket.stream.transport")
|
|
|
|
GC_ref(transp)
|
|
|
|
result = cast[StreamTransport](transp)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc connect*(address: TransportAddress,
|
|
|
|
bufferSize = DefaultStreamBufferSize): Future[StreamTransport] =
|
2018-05-27 05:49:47 +00:00
|
|
|
## Open new connection to remote peer with address ``address`` and create
|
|
|
|
## new transport object ``StreamTransport`` for established connection.
|
|
|
|
## ``bufferSize`` - size of internal buffer for transport.
|
2018-05-16 08:22:34 +00:00
|
|
|
var
|
|
|
|
saddr: Sockaddr_storage
|
|
|
|
slen: SockLen
|
|
|
|
sock: AsyncFD
|
|
|
|
var retFuture = newFuture[StreamTransport]("transport.connect")
|
|
|
|
toSockAddr(address.address, address.port, saddr, slen)
|
|
|
|
sock = createAsyncSocket(address.address.getDomain(), SockType.SOCK_STREAM,
|
|
|
|
Protocol.IPPROTO_TCP)
|
|
|
|
if sock == asyncInvalidSocket:
|
|
|
|
result.fail(newException(OSError, osErrorMsg(osLastError())))
|
|
|
|
|
|
|
|
proc continuation(udata: pointer) =
|
|
|
|
var data = cast[ptr CompletionData](udata)
|
|
|
|
var err = 0
|
2018-05-23 11:03:18 +00:00
|
|
|
let fd = data.fd
|
|
|
|
if not fd.getSocketError(err):
|
|
|
|
fd.closeAsyncSocket()
|
2018-05-16 08:22:34 +00:00
|
|
|
retFuture.fail(newException(OSError, osErrorMsg(osLastError())))
|
|
|
|
return
|
|
|
|
if err != 0:
|
2018-05-23 11:03:18 +00:00
|
|
|
fd.closeAsyncSocket()
|
2018-05-16 08:22:34 +00:00
|
|
|
retFuture.fail(newException(OSError, osErrorMsg(OSErrorCode(err))))
|
|
|
|
return
|
2018-05-23 11:03:18 +00:00
|
|
|
fd.removeWriter()
|
|
|
|
retFuture.complete(newStreamSocketTransport(fd, bufferSize))
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
while true:
|
|
|
|
var res = posix.connect(SocketHandle(sock),
|
|
|
|
cast[ptr SockAddr](addr saddr), slen)
|
|
|
|
if res == 0:
|
|
|
|
retFuture.complete(newStreamSocketTransport(sock, bufferSize))
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
let err = osLastError()
|
|
|
|
if int(err) == EINTR:
|
|
|
|
continue
|
|
|
|
elif int(err) == EINPROGRESS:
|
|
|
|
sock.addWriter(continuation)
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
sock.closeAsyncSocket()
|
|
|
|
retFuture.fail(newException(OSError, osErrorMsg(err)))
|
|
|
|
break
|
|
|
|
return retFuture
|
|
|
|
|
|
|
|
proc serverCallback(udata: pointer) =
|
|
|
|
var
|
|
|
|
saddr: Sockaddr_storage
|
|
|
|
slen: SockLen
|
|
|
|
var server = cast[StreamServer](cast[ptr CompletionData](udata).udata)
|
|
|
|
while true:
|
|
|
|
let res = posix.accept(SocketHandle(server.sock),
|
|
|
|
cast[ptr SockAddr](addr saddr), addr slen)
|
|
|
|
if int(res) > 0:
|
|
|
|
let sock = wrapAsyncSocket(res)
|
|
|
|
if sock != asyncInvalidSocket:
|
2018-05-22 22:29:07 +00:00
|
|
|
discard server.function(server,
|
2018-05-16 08:22:34 +00:00
|
|
|
newStreamSocketTransport(sock, server.bufferSize),
|
|
|
|
server.udata)
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
let err = osLastError()
|
|
|
|
if int(err) == EINTR:
|
|
|
|
continue
|
2018-06-05 05:51:59 +00:00
|
|
|
else:
|
2018-05-16 08:22:34 +00:00
|
|
|
## Critical unrecoverable error
|
|
|
|
raiseOsError(err)
|
|
|
|
|
2018-06-04 09:57:17 +00:00
|
|
|
proc resumeAccept(server: SocketServer) =
|
|
|
|
addReader(server.sock, serverCallback, cast[pointer](server))
|
|
|
|
|
|
|
|
proc pauseAccept(server: SocketServer) =
|
|
|
|
removeReader(server.sock)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc resumeRead(transp: StreamTransport) {.inline.} =
|
|
|
|
transp.state.excl(ReadPaused)
|
|
|
|
addReader(transp.fd, readStreamLoop, cast[pointer](transp))
|
|
|
|
|
|
|
|
proc resumeWrite(transp: StreamTransport) {.inline.} =
|
|
|
|
transp.state.excl(WritePaused)
|
|
|
|
addWriter(transp.fd, writeStreamLoop, cast[pointer](transp))
|
|
|
|
|
|
|
|
proc start*(server: SocketServer) =
|
2018-05-21 21:52:57 +00:00
|
|
|
## Starts ``server``.
|
2018-06-04 09:57:17 +00:00
|
|
|
if server.status == ServerStatus.Starting:
|
|
|
|
server.resumeAccept()
|
|
|
|
server.status = ServerStatus.Running
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc stop*(server: SocketServer) =
|
2018-05-22 21:03:13 +00:00
|
|
|
## Stops ``server``.
|
2018-06-04 09:57:17 +00:00
|
|
|
if server.status == ServerStatus.Running:
|
|
|
|
server.pauseAccept()
|
|
|
|
server.status = ServerStatus.Stopped
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc join*(server: SocketServer) {.async.} =
|
2018-06-04 09:57:17 +00:00
|
|
|
## Waits until ``server`` is not closed.
|
2018-05-21 21:52:57 +00:00
|
|
|
if not server.loopFuture.finished:
|
|
|
|
await server.loopFuture
|
|
|
|
|
|
|
|
proc close*(server: SocketServer) =
|
|
|
|
## Release ``server`` resources.
|
2018-06-04 09:57:17 +00:00
|
|
|
if server.status == ServerStatus.Stopped:
|
|
|
|
closeAsyncSocket(server.sock)
|
|
|
|
server.status = Closed
|
|
|
|
server.loopFuture.complete()
|
|
|
|
GC_unref(server)
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc createStreamServer*(host: TransportAddress,
|
|
|
|
cbproc: StreamCallback,
|
2018-05-31 08:03:58 +00:00
|
|
|
flags: set[ServerFlags] = {},
|
2018-05-16 08:22:34 +00:00
|
|
|
sock: AsyncFD = asyncInvalidSocket,
|
|
|
|
backlog: int = 100,
|
|
|
|
bufferSize: int = DefaultStreamBufferSize,
|
|
|
|
udata: pointer = nil): StreamServer =
|
2018-05-28 23:35:15 +00:00
|
|
|
## Create new TCP stream server.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-27 05:49:47 +00:00
|
|
|
## ``host`` - address to which server will be bound.
|
|
|
|
## ``flags`` - flags to apply to server socket.
|
|
|
|
## ``cbproc`` - callback function which will be called, when new client
|
|
|
|
## connection will be established.
|
2018-05-28 23:35:15 +00:00
|
|
|
## ``sock`` - user-driven socket to use.
|
|
|
|
## ``backlog`` - number of outstanding connections in the socket's listen
|
|
|
|
## queue.
|
|
|
|
## ``bufferSize`` - size of internal buffer for transport.
|
|
|
|
## ``udata`` - user-defined pointer.
|
2018-05-16 08:22:34 +00:00
|
|
|
var
|
|
|
|
saddr: Sockaddr_storage
|
|
|
|
slen: SockLen
|
|
|
|
serverSocket: AsyncFD
|
|
|
|
if sock == asyncInvalidSocket:
|
|
|
|
serverSocket = createAsyncSocket(host.address.getDomain(),
|
|
|
|
SockType.SOCK_STREAM,
|
|
|
|
Protocol.IPPROTO_TCP)
|
|
|
|
if serverSocket == asyncInvalidSocket:
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
else:
|
|
|
|
if not setSocketBlocking(SocketHandle(sock), false):
|
|
|
|
raiseOsError(osLastError())
|
|
|
|
register(sock)
|
|
|
|
serverSocket = sock
|
|
|
|
|
|
|
|
if ServerFlags.ReuseAddr in flags:
|
|
|
|
if not setSockOpt(serverSocket, SOL_SOCKET, SO_REUSEADDR, 1):
|
|
|
|
let err = osLastError()
|
|
|
|
if sock == asyncInvalidSocket:
|
|
|
|
closeAsyncSocket(serverSocket)
|
|
|
|
raiseOsError(err)
|
|
|
|
|
|
|
|
toSockAddr(host.address, host.port, saddr, slen)
|
|
|
|
if bindAddr(SocketHandle(serverSocket), cast[ptr SockAddr](addr saddr),
|
|
|
|
slen) != 0:
|
|
|
|
let err = osLastError()
|
|
|
|
if sock == asyncInvalidSocket:
|
|
|
|
closeAsyncSocket(serverSocket)
|
|
|
|
raiseOsError(err)
|
2018-05-21 21:52:57 +00:00
|
|
|
|
2018-05-16 08:22:34 +00:00
|
|
|
if nativesockets.listen(SocketHandle(serverSocket), cint(backlog)) != 0:
|
|
|
|
let err = osLastError()
|
|
|
|
if sock == asyncInvalidSocket:
|
|
|
|
closeAsyncSocket(serverSocket)
|
|
|
|
raiseOsError(err)
|
|
|
|
|
|
|
|
result = StreamServer()
|
|
|
|
result.sock = serverSocket
|
|
|
|
result.function = cbproc
|
|
|
|
result.bufferSize = bufferSize
|
|
|
|
result.status = Starting
|
2018-06-04 09:57:17 +00:00
|
|
|
result.loopFuture = newFuture[void]("stream.server")
|
2018-05-16 08:22:34 +00:00
|
|
|
result.udata = udata
|
|
|
|
result.local = host
|
2018-06-04 09:57:17 +00:00
|
|
|
|
|
|
|
when defined(windows):
|
|
|
|
result.aovl.data = CompletionData(fd: serverSocket, cb: acceptLoop,
|
|
|
|
udata: cast[pointer](result))
|
|
|
|
result.domain = host.address.getDomain()
|
|
|
|
result.apending = false
|
2018-05-21 21:52:57 +00:00
|
|
|
GC_ref(result)
|
2018-06-04 09:57:17 +00:00
|
|
|
result.resumeAccept()
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc write*(transp: StreamTransport, pbytes: pointer,
|
|
|
|
nbytes: int): Future[int] {.async.} =
|
2018-05-27 05:49:47 +00:00
|
|
|
## Write data from buffer ``pbytes`` with size ``nbytes`` using transport
|
2018-05-22 21:03:13 +00:00
|
|
|
## ``transp``.
|
2018-05-16 08:22:34 +00:00
|
|
|
checkClosed(transp)
|
|
|
|
var waitFuture = newFuture[void]("transport.write")
|
2018-05-21 21:52:57 +00:00
|
|
|
var vector = StreamVector(kind: DataBuffer, writer: waitFuture,
|
|
|
|
buf: pbytes, buflen: nbytes)
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.queue.addLast(vector)
|
|
|
|
if WritePaused in transp.state:
|
|
|
|
transp.resumeWrite()
|
|
|
|
await vector.writer
|
|
|
|
if WriteError in transp.state:
|
|
|
|
raise transp.getError()
|
|
|
|
result = nbytes
|
|
|
|
|
2018-06-04 16:42:54 +00:00
|
|
|
template write*(transp: StreamTransport, msg: var string): untyped =
|
|
|
|
## Write string ``msg`` using transport ``transp``.
|
|
|
|
write(transp, addr msg[0], len(msg))
|
|
|
|
|
|
|
|
template write*(transp: StreamTransport, msg: var seq[byte]): untyped =
|
|
|
|
## Write seq[byte] ``msg`` using transport ``transp``.
|
|
|
|
write(transp, addr msg[0], len(msg))
|
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
proc writeFile*(transp: StreamTransport, handle: int,
|
|
|
|
offset: uint = 0,
|
|
|
|
size: int = 0): Future[void] {.async.} =
|
2018-05-22 21:03:13 +00:00
|
|
|
## Write data from file descriptor ``handle`` to transport ``transp``.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-22 21:03:13 +00:00
|
|
|
## You can specify starting ``offset`` in opened file and number of bytes
|
|
|
|
## to transfer from file to transport via ``size``.
|
2018-05-21 21:52:57 +00:00
|
|
|
if transp.kind != TransportKind.Socket:
|
|
|
|
raise newException(TransportError, "You can transmit files only to sockets")
|
|
|
|
checkClosed(transp)
|
|
|
|
var waitFuture = newFuture[void]("transport.writeFile")
|
|
|
|
var vector = StreamVector(kind: DataFile, writer: waitFuture,
|
|
|
|
buf: cast[pointer](size), offset: offset,
|
|
|
|
buflen: handle)
|
|
|
|
transp.queue.addLast(vector)
|
|
|
|
if WritePaused in transp.state:
|
|
|
|
transp.resumeWrite()
|
|
|
|
await vector.writer
|
|
|
|
if WriteError in transp.state:
|
|
|
|
raise transp.getError()
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-06-04 16:42:54 +00:00
|
|
|
proc atEof*(transp: StreamTransport): bool {.inline.} =
|
|
|
|
## Returns ``true`` if ``transp`` is at EOF.
|
|
|
|
result = (transp.offset == 0) and (ReadEof in transp.state) and
|
|
|
|
(ReadPaused in transp.state)
|
|
|
|
|
2018-05-16 08:22:34 +00:00
|
|
|
proc readExactly*(transp: StreamTransport, pbytes: pointer,
|
2018-05-21 21:52:57 +00:00
|
|
|
nbytes: int) {.async.} =
|
2018-05-22 21:03:13 +00:00
|
|
|
## Read exactly ``nbytes`` bytes from transport ``transp`` and store it to
|
|
|
|
## ``pbytes``.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-22 21:03:13 +00:00
|
|
|
## If EOF is received and ``nbytes`` is not yet readed, the procedure
|
|
|
|
## will raise ``TransportIncompleteError``.
|
2018-05-16 08:22:34 +00:00
|
|
|
checkClosed(transp)
|
|
|
|
checkPending(transp)
|
|
|
|
var index = 0
|
|
|
|
while true:
|
|
|
|
if transp.offset == 0:
|
|
|
|
if (ReadError in transp.state):
|
|
|
|
raise transp.getError()
|
2018-06-05 05:51:59 +00:00
|
|
|
if (ReadClosed in transp.state) or transp.atEof():
|
2018-05-16 08:22:34 +00:00
|
|
|
raise newException(TransportIncompleteError, "Data incomplete!")
|
|
|
|
|
|
|
|
if transp.offset >= (nbytes - index):
|
|
|
|
copyMem(cast[pointer](cast[uint](pbytes) + uint(index)),
|
|
|
|
addr(transp.buffer[0]), nbytes - index)
|
|
|
|
transp.shiftBuffer(nbytes - index)
|
|
|
|
break
|
|
|
|
else:
|
2018-05-21 21:52:57 +00:00
|
|
|
if transp.offset != 0:
|
|
|
|
copyMem(cast[pointer](cast[uint](pbytes) + uint(index)),
|
|
|
|
addr(transp.buffer[0]), transp.offset)
|
|
|
|
index += transp.offset
|
|
|
|
|
|
|
|
transp.reader = newFuture[void]("stream.transport.readExactly")
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.offset = 0
|
|
|
|
if ReadPaused in transp.state:
|
|
|
|
transp.resumeRead()
|
|
|
|
await transp.reader
|
2018-06-04 10:40:41 +00:00
|
|
|
# we need to clear transp.reader to avoid double completion of this
|
|
|
|
# Future[T], because readLoop continues working.
|
|
|
|
transp.reader = nil
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc readOnce*(transp: StreamTransport, pbytes: pointer,
|
|
|
|
nbytes: int): Future[int] {.async.} =
|
|
|
|
## Perform one read operation on transport ``transp``.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-27 05:49:47 +00:00
|
|
|
## If internal buffer is not empty, ``nbytes`` bytes will be transferred from
|
|
|
|
## internal buffer, otherwise it will wait until some bytes will be received.
|
2018-05-16 08:22:34 +00:00
|
|
|
checkClosed(transp)
|
|
|
|
checkPending(transp)
|
|
|
|
while true:
|
|
|
|
if transp.offset == 0:
|
|
|
|
if (ReadError in transp.state):
|
|
|
|
raise transp.getError()
|
2018-06-05 05:51:59 +00:00
|
|
|
if (ReadClosed in transp.state) or transp.atEof():
|
2018-05-16 08:22:34 +00:00
|
|
|
result = 0
|
|
|
|
break
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.reader = newFuture[void]("stream.transport.readOnce")
|
2018-05-16 08:22:34 +00:00
|
|
|
if ReadPaused in transp.state:
|
|
|
|
transp.resumeRead()
|
|
|
|
await transp.reader
|
2018-06-04 10:40:41 +00:00
|
|
|
# we need to clear transp.reader to avoid double completion of this
|
|
|
|
# Future[T], because readLoop continues working.
|
2018-05-16 08:22:34 +00:00
|
|
|
transp.reader = nil
|
|
|
|
else:
|
|
|
|
if transp.offset > nbytes:
|
|
|
|
copyMem(pbytes, addr(transp.buffer[0]), nbytes)
|
|
|
|
transp.shiftBuffer(nbytes)
|
|
|
|
result = nbytes
|
|
|
|
else:
|
|
|
|
copyMem(pbytes, addr(transp.buffer[0]), transp.offset)
|
|
|
|
result = transp.offset
|
|
|
|
break
|
|
|
|
|
|
|
|
proc readUntil*(transp: StreamTransport, pbytes: pointer, nbytes: int,
|
|
|
|
sep: seq[byte]): Future[int] {.async.} =
|
2018-05-22 21:03:13 +00:00
|
|
|
## Read data from the transport ``transp`` until separator ``sep`` is found.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-22 21:03:13 +00:00
|
|
|
## On success, the data and separator will be removed from the internal
|
|
|
|
## buffer (consumed). Returned data will NOT include the separator at the end.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-28 23:35:15 +00:00
|
|
|
## If EOF is received, and `sep` was not found, procedure will raise
|
|
|
|
## ``TransportIncompleteError``.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-28 23:35:15 +00:00
|
|
|
## If ``nbytes`` bytes has been received and `sep` was not found, procedure
|
|
|
|
## will raise ``TransportLimitError``.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-28 23:35:15 +00:00
|
|
|
## Procedure returns actual number of bytes read.
|
2018-05-16 08:22:34 +00:00
|
|
|
checkClosed(transp)
|
|
|
|
checkPending(transp)
|
|
|
|
|
|
|
|
var dest = cast[ptr UncheckedArray[byte]](pbytes)
|
|
|
|
var state = 0
|
|
|
|
var k = 0
|
|
|
|
var index = 0
|
|
|
|
|
|
|
|
while true:
|
2018-06-05 05:51:59 +00:00
|
|
|
if ReadError in transp.state:
|
|
|
|
raise transp.getError()
|
|
|
|
if (ReadClosed in transp.state) or transp.atEof():
|
|
|
|
raise newException(TransportIncompleteError, "Data incomplete!")
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
index = 0
|
|
|
|
while index < transp.offset:
|
|
|
|
let ch = transp.buffer[index]
|
|
|
|
if sep[state] == ch:
|
|
|
|
inc(state)
|
|
|
|
else:
|
|
|
|
state = 0
|
|
|
|
if k < nbytes:
|
|
|
|
dest[k] = ch
|
|
|
|
inc(k)
|
|
|
|
else:
|
|
|
|
raise newException(TransportLimitError, "Limit reached!")
|
|
|
|
if state == len(sep):
|
|
|
|
break
|
|
|
|
inc(index)
|
|
|
|
|
|
|
|
if state == len(sep):
|
2018-06-05 05:51:59 +00:00
|
|
|
transp.shiftBuffer(index + 1)
|
2018-05-16 08:22:34 +00:00
|
|
|
result = k
|
|
|
|
break
|
|
|
|
else:
|
2018-06-05 05:51:59 +00:00
|
|
|
transp.shiftBuffer(transp.offset)
|
|
|
|
transp.reader = newFuture[void]("stream.transport.readUntil")
|
|
|
|
if ReadPaused in transp.state:
|
|
|
|
transp.resumeRead()
|
|
|
|
await transp.reader
|
|
|
|
# we need to clear transp.reader to avoid double completion of this
|
|
|
|
# Future[T], because readLoop continues working.
|
|
|
|
transp.reader = nil
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc readLine*(transp: StreamTransport, limit = 0,
|
|
|
|
sep = "\r\n"): Future[string] {.async.} =
|
2018-05-22 21:03:13 +00:00
|
|
|
## Read one line from transport ``transp``, where "line" is a sequence of
|
|
|
|
## bytes ending with ``sep`` (default is "\r\n").
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-22 21:03:13 +00:00
|
|
|
## If EOF is received, and ``sep`` was not found, the method will return the
|
|
|
|
## partial read bytes.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-22 21:03:13 +00:00
|
|
|
## If the EOF was received and the internal buffer is empty, return an
|
|
|
|
## empty string.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-22 21:03:13 +00:00
|
|
|
## If ``limit`` more then 0, then read is limited to ``limit`` bytes.
|
2018-05-16 08:22:34 +00:00
|
|
|
checkClosed(transp)
|
|
|
|
checkPending(transp)
|
|
|
|
|
|
|
|
result = ""
|
|
|
|
var lim = if limit <= 0: -1 else: limit
|
|
|
|
var state = 0
|
|
|
|
var index = 0
|
|
|
|
|
|
|
|
while true:
|
2018-06-05 05:51:59 +00:00
|
|
|
if (ReadError in transp.state):
|
|
|
|
raise transp.getError()
|
|
|
|
if (ReadClosed in transp.state) or transp.atEof():
|
|
|
|
break
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
index = 0
|
|
|
|
while index < transp.offset:
|
|
|
|
let ch = char(transp.buffer[index])
|
|
|
|
if sep[state] == ch:
|
|
|
|
inc(state)
|
|
|
|
if state == len(sep):
|
|
|
|
transp.shiftBuffer(index + 1)
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
state = 0
|
|
|
|
result.add(ch)
|
|
|
|
if len(result) == lim:
|
|
|
|
transp.shiftBuffer(index + 1)
|
|
|
|
break
|
|
|
|
inc(index)
|
|
|
|
|
|
|
|
if (state == len(sep)) or (lim == len(result)):
|
|
|
|
break
|
|
|
|
else:
|
2018-06-05 05:51:59 +00:00
|
|
|
transp.shiftBuffer(transp.offset)
|
|
|
|
transp.reader = newFuture[void]("stream.transport.readLine")
|
|
|
|
if ReadPaused in transp.state:
|
|
|
|
transp.resumeRead()
|
|
|
|
await transp.reader
|
|
|
|
# we need to clear transp.reader to avoid double completion of this
|
|
|
|
# Future[T], because readLoop continues working.
|
|
|
|
transp.reader = nil
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc read*(transp: StreamTransport, n = -1): Future[seq[byte]] {.async.} =
|
2018-06-05 05:51:59 +00:00
|
|
|
## Read all bytes (n == -1) or exactly `n` bytes from transport ``transp``.
|
2018-06-04 09:57:17 +00:00
|
|
|
##
|
2018-05-28 23:35:15 +00:00
|
|
|
## This procedure allocates buffer seq[byte] and return it as result.
|
2018-05-16 08:22:34 +00:00
|
|
|
checkClosed(transp)
|
|
|
|
checkPending(transp)
|
|
|
|
result = newSeq[byte]()
|
|
|
|
while true:
|
|
|
|
if (ReadError in transp.state):
|
|
|
|
raise transp.getError()
|
2018-06-05 05:51:59 +00:00
|
|
|
if (ReadClosed in transp.state) or transp.atEof():
|
2018-05-16 08:22:34 +00:00
|
|
|
break
|
2018-06-05 05:51:59 +00:00
|
|
|
|
2018-05-16 08:22:34 +00:00
|
|
|
if transp.offset > 0:
|
|
|
|
let s = len(result)
|
|
|
|
let o = s + transp.offset
|
|
|
|
if n == -1:
|
|
|
|
# grabbing all incoming data, until EOF
|
|
|
|
result.setLen(o)
|
|
|
|
copyMem(cast[pointer](addr result[s]), addr(transp.buffer[0]),
|
|
|
|
transp.offset)
|
|
|
|
transp.offset = 0
|
|
|
|
else:
|
2018-06-05 05:51:59 +00:00
|
|
|
let left = n - s
|
|
|
|
if transp.offset >= left:
|
2018-05-16 08:22:34 +00:00
|
|
|
# size of buffer data is more then we need, grabbing only part
|
|
|
|
result.setLen(n)
|
|
|
|
copyMem(cast[pointer](addr result[s]), addr(transp.buffer[0]),
|
2018-06-05 05:51:59 +00:00
|
|
|
left)
|
|
|
|
transp.shiftBuffer(left)
|
2018-05-16 08:22:34 +00:00
|
|
|
break
|
|
|
|
else:
|
|
|
|
# there not enough data in buffer, grabbing all
|
|
|
|
result.setLen(o)
|
|
|
|
copyMem(cast[pointer](addr result[s]), addr(transp.buffer[0]),
|
|
|
|
transp.offset)
|
|
|
|
transp.offset = 0
|
|
|
|
|
2018-05-21 21:52:57 +00:00
|
|
|
transp.reader = newFuture[void]("stream.transport.read")
|
2018-05-16 08:22:34 +00:00
|
|
|
if ReadPaused in transp.state:
|
|
|
|
transp.resumeRead()
|
|
|
|
await transp.reader
|
2018-06-04 10:40:41 +00:00
|
|
|
# we need to clear transp.reader to avoid double completion of this
|
|
|
|
# Future[T], because readLoop continues working.
|
|
|
|
transp.reader = nil
|
2018-05-16 08:22:34 +00:00
|
|
|
|
2018-06-04 16:42:54 +00:00
|
|
|
proc consume*(transp: StreamTransport, n = -1): Future[int] {.async.} =
|
|
|
|
## Consume all bytes (n == -1) or ``n`` bytes from transport ``transp``.
|
|
|
|
##
|
|
|
|
## Return number of bytes actually consumed
|
|
|
|
checkClosed(transp)
|
|
|
|
checkPending(transp)
|
|
|
|
result = 0
|
|
|
|
while true:
|
|
|
|
if (ReadError in transp.state):
|
|
|
|
raise transp.getError()
|
|
|
|
if ReadClosed in transp.state or transp.atEof():
|
|
|
|
break
|
2018-06-05 05:51:59 +00:00
|
|
|
|
2018-06-04 16:42:54 +00:00
|
|
|
if transp.offset > 0:
|
|
|
|
if n == -1:
|
|
|
|
# consume all incoming data, until EOF
|
|
|
|
result += transp.offset
|
|
|
|
transp.offset = 0
|
|
|
|
else:
|
|
|
|
let left = n - result
|
|
|
|
if transp.offset >= left:
|
|
|
|
# size of buffer data is more then we need, consume only part
|
|
|
|
result += left
|
|
|
|
transp.shiftBuffer(left)
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
# there not enough data in buffer, consume all
|
|
|
|
result += transp.offset
|
|
|
|
transp.offset = 0
|
|
|
|
|
|
|
|
transp.reader = newFuture[void]("stream.transport.consume")
|
|
|
|
if ReadPaused in transp.state:
|
|
|
|
transp.resumeRead()
|
|
|
|
await transp.reader
|
|
|
|
# we need to clear transp.reader to avoid double completion of this
|
|
|
|
# Future[T], because readLoop continues working.
|
|
|
|
transp.reader = nil
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc join*(transp: StreamTransport) {.async.} =
|
|
|
|
## Wait until ``transp`` will not be closed.
|
2018-05-21 21:52:57 +00:00
|
|
|
if not transp.future.finished:
|
|
|
|
await transp.future
|
2018-05-16 08:22:34 +00:00
|
|
|
|
|
|
|
proc close*(transp: StreamTransport) =
|
|
|
|
## Closes and frees resources of transport ``transp``.
|
2018-06-04 09:57:17 +00:00
|
|
|
if {ReadClosed, WriteClosed} * transp.state == {}:
|
2018-05-16 08:22:34 +00:00
|
|
|
when defined(windows):
|
|
|
|
discard cancelIo(Handle(transp.fd))
|
|
|
|
closeAsyncSocket(transp.fd)
|
|
|
|
transp.state.incl(WriteClosed)
|
|
|
|
transp.state.incl(ReadClosed)
|
|
|
|
transp.future.complete()
|
2018-05-21 21:52:57 +00:00
|
|
|
GC_unref(transp)
|