nim-chronos/chronos/streams/chunkstream.nim

284 lines
10 KiB
Nim
Raw Normal View History

#
# Chronos Asynchronous Chunked-Encoding Stream
# (c) Copyright 2019-Present
# Status Research & Development GmbH
#
# Licensed under either of
# Apache License, version 2.0, (LICENSE-APACHEv2)
# MIT license (LICENSE-MIT)
## This module implements HTTP/1.1 chunked-encoded stream reading and writing.
{.push raises: [].}
import ../[asyncloop, timer, bipbuffer, config]
import asyncstream, ../transports/[stream, common]
2023-11-13 09:56:19 +00:00
import results
export asyncloop, asyncstream, stream, timer, common, results
const
ChunkBufferSize = chronosStreamDefaultBufferSize
MaxChunkHeaderSize = 1024
ChunkHeaderValueSize = 8
2021-02-17 00:03:12 +00:00
# This is limit for chunk size to 8 hexadecimal digits, so maximum
# chunk size for this implementation become:
# 2^32 == FFFF_FFFF'u32 == 4,294,967,295 bytes.
CRLF = @[byte(0x0D), byte(0x0A)]
type
ChunkedStreamReader* = ref object of AsyncStreamReader
ChunkedStreamWriter* = ref object of AsyncStreamWriter
ChunkedStreamError* = object of AsyncStreamError
ChunkedStreamProtocolError* = object of ChunkedStreamError
ChunkedStreamIncompleteError* = object of ChunkedStreamError
proc `-`(x: uint32): uint32 {.inline.} =
result = (0xFFFF_FFFF'u32 - x) + 1'u32
proc LT(x, y: uint32): uint32 {.inline.} =
let z = x - y
(z xor ((y xor x) and (y xor z))) shr 31
2021-02-18 14:39:36 +00:00
proc hexValue*(c: byte): int =
# This is nim adaptation of
# https://github.com/pornin/CTTK/blob/master/src/hex.c#L28-L52
let x = uint32(c) - 0x30'u32
let y = uint32(c) - 0x41'u32
let z = uint32(c) - 0x61'u32
let r = ((x + 1'u32) and -LT(x, 10)) or
((y + 11'u32) and -LT(y, 6)) or
((z + 11'u32) and -LT(z, 6))
int(r) - 1
proc getChunkSize(buffer: openArray[byte]): Result[uint64, cstring] =
# We using `uint64` representation, but allow only 2^32 chunk size,
# ChunkHeaderValueSize.
var res = 0'u64
for i in 0 ..< min(len(buffer), ChunkHeaderValueSize + 1):
let value = hexValue(buffer[i])
2021-02-17 00:03:12 +00:00
if value < 0:
if buffer[i] == byte(';'):
# chunk-extension is present, so chunk size is already decoded in res.
return ok(res)
else:
return err("Incorrect chunk size encoding")
else:
if i >= ChunkHeaderValueSize:
return err("The chunk size exceeds the limit")
res = (res shl 4) or uint64(value)
2021-02-17 00:03:12 +00:00
ok(res)
proc setChunkSize(buffer: var openArray[byte], length: int64): int =
# Store length as chunk header size (hexadecimal value) with CRLF.
# Maximum stored value is ``0xFFFF_FFFF``.
# Buffer ``buffer`` length must be at least 10 octets.
doAssert(length <= int64(uint32.high))
var n = 0xF000_0000'i64
var i = 32
var c = 0
if length == 0:
buffer[0] = byte('0')
buffer[1] = byte(0x0D)
buffer[2] = byte(0x0A)
3
else:
while n != 0:
var v = length and n
if v != 0 or c != 0:
let digit = byte((length and n) shr (i - 4))
var ch = digit + byte('0')
if ch > byte('9'):
ch = ch + 0x07'u8
buffer[c] = ch
inc(c)
n = n shr 4
i = i - 4
buffer[c] = byte(0x0D)
buffer[c + 1] = byte(0x0A)
2021-02-17 00:03:12 +00:00
(c + 2)
proc chunkedReadLoop(stream: AsyncStreamReader) {.async: (raises: []).} =
2021-02-17 00:03:12 +00:00
var rstream = ChunkedStreamReader(stream)
var buffer = newSeq[byte](MaxChunkHeaderSize)
rstream.state = AsyncStreamState.Running
while true:
try:
# Reading chunk size
2021-02-17 00:03:12 +00:00
let res = await rstream.rsource.readUntil(addr buffer[0], len(buffer),
CRLF)
let cres = getChunkSize(buffer.toOpenArray(0, res - len(CRLF) - 1))
2019-06-06 11:00:47 +00:00
2021-02-17 00:03:12 +00:00
if cres.isErr():
if rstream.state == AsyncStreamState.Running:
rstream.error = newException(ChunkedStreamProtocolError, $cres.error)
rstream.state = AsyncStreamState.Error
else:
2021-02-17 00:03:12 +00:00
var chunksize = cres.get()
if chunksize > 0'u64:
while chunksize > 0'u64:
let
(data, rsize) = rstream.buffer.backend.reserve()
toRead = int(min(chunksize, uint64(rsize)))
await rstream.rsource.readExactly(data, toRead)
rstream.buffer.backend.commit(toRead)
2021-02-17 00:03:12 +00:00
await rstream.buffer.transfer()
chunksize = chunksize - uint64(toRead)
if rstream.state == AsyncStreamState.Running:
# Reading chunk trailing CRLF
await rstream.rsource.readExactly(addr buffer[0], 2)
if buffer[0] != CRLF[0] or buffer[1] != CRLF[1]:
if rstream.state == AsyncStreamState.Running:
rstream.error = newException(ChunkedStreamProtocolError,
"Unexpected trailing bytes")
rstream.state = AsyncStreamState.Error
2021-02-17 00:03:12 +00:00
else:
# Reading trailing line for last chunk
discard await rstream.rsource.readUntil(addr buffer[0],
len(buffer), CRLF)
if rstream.state == AsyncStreamState.Running:
rstream.state = AsyncStreamState.Finished
await rstream.buffer.transfer()
except CancelledError:
if rstream.state == AsyncStreamState.Running:
rstream.state = AsyncStreamState.Stopped
except AsyncStreamLimitError:
if rstream.state == AsyncStreamState.Running:
rstream.state = AsyncStreamState.Error
rstream.error = newException(ChunkedStreamProtocolError,
"Chunk header exceeds maximum size")
except AsyncStreamIncompleteError:
if rstream.state == AsyncStreamState.Running:
rstream.state = AsyncStreamState.Error
rstream.error = newException(ChunkedStreamIncompleteError,
"Incomplete chunk received")
except AsyncStreamReadError as exc:
if rstream.state == AsyncStreamState.Running:
rstream.state = AsyncStreamState.Error
rstream.error = exc
except AsyncStreamError as exc:
if rstream.state == AsyncStreamState.Running:
rstream.state = AsyncStreamState.Error
rstream.error = exc
if rstream.state != AsyncStreamState.Running:
# We need to notify consumer about error/close, but we do not care about
# incoming data anymore.
rstream.buffer.forget()
break
proc chunkedWriteLoop(stream: AsyncStreamWriter) {.async: (raises: []).} =
2021-02-17 00:03:12 +00:00
var wstream = ChunkedStreamWriter(stream)
var buffer: array[16, byte]
var error: ref AsyncStreamError
wstream.state = AsyncStreamState.Running
while true:
var item: WriteItem
# Getting new item from stream's queue.
try:
item = await wstream.queue.get()
# `item.size == 0` is marker of stream finish, while `item.size != 0` is
# data's marker.
if item.size > 0:
let length = setChunkSize(buffer, int64(item.size))
# Writing chunk header <length>CRLF.
await wstream.wsource.write(addr buffer[0], length)
# Writing chunk data.
case item.kind
of WriteType.Pointer:
await wstream.wsource.write(item.dataPtr, item.size)
of WriteType.Sequence:
await wstream.wsource.write(addr item.dataSeq[0], item.size)
of WriteType.String:
await wstream.wsource.write(addr item.dataStr[0], item.size)
# Writing chunk footer CRLF.
await wstream.wsource.write(CRLF)
# Everything is fine, completing queue item's future.
item.future.complete()
else:
let length = setChunkSize(buffer, 0'i64)
# Write finish chunk `0`.
await wstream.wsource.write(addr buffer[0], length)
# Write trailing CRLF.
await wstream.wsource.write(CRLF)
# Everything is fine, completing queue item's future.
item.future.complete()
# Set stream state to Finished.
if wstream.state == AsyncStreamState.Running:
wstream.state = AsyncStreamState.Finished
except CancelledError:
if wstream.state == AsyncStreamState.Running:
wstream.state = AsyncStreamState.Stopped
error = newAsyncStreamUseClosedError()
except AsyncStreamError as exc:
if wstream.state == AsyncStreamState.Running:
wstream.state = AsyncStreamState.Error
error = exc
if wstream.state != AsyncStreamState.Running:
if wstream.state == AsyncStreamState.Finished:
error = newAsyncStreamUseClosedError()
else:
if not(isNil(item.future)):
if not(item.future.finished()):
item.future.fail(error)
while not(wstream.queue.empty()):
let pitem =
try:
wstream.queue.popFirstNoWait()
except AsyncQueueEmptyError:
raiseAssert "AsyncQueue should not be empty at this moment"
if not(pitem.future.finished()):
pitem.future.fail(error)
break
proc init*[T](child: ChunkedStreamReader, rsource: AsyncStreamReader,
2019-06-06 11:00:47 +00:00
bufferSize = ChunkBufferSize, udata: ref T) =
2021-02-17 00:03:12 +00:00
init(AsyncStreamReader(child), rsource, chunkedReadLoop, bufferSize,
udata)
proc init*(child: ChunkedStreamReader, rsource: AsyncStreamReader,
2019-06-06 11:00:47 +00:00
bufferSize = ChunkBufferSize) =
2021-02-17 00:03:12 +00:00
init(AsyncStreamReader(child), rsource, chunkedReadLoop, bufferSize)
proc newChunkedStreamReader*[T](rsource: AsyncStreamReader,
bufferSize = AsyncStreamDefaultBufferSize,
udata: ref T): ChunkedStreamReader =
var res = ChunkedStreamReader()
res.init(rsource, bufferSize, udata)
res
proc newChunkedStreamReader*(rsource: AsyncStreamReader,
bufferSize = AsyncStreamDefaultBufferSize,
): ChunkedStreamReader =
var res = ChunkedStreamReader()
res.init(rsource, bufferSize)
res
proc init*[T](child: ChunkedStreamWriter, wsource: AsyncStreamWriter,
queueSize = AsyncStreamDefaultQueueSize, udata: ref T) =
2021-02-17 00:03:12 +00:00
init(AsyncStreamWriter(child), wsource, chunkedWriteLoop, queueSize,
udata)
proc init*(child: ChunkedStreamWriter, wsource: AsyncStreamWriter,
queueSize = AsyncStreamDefaultQueueSize) =
2021-02-17 00:03:12 +00:00
init(AsyncStreamWriter(child), wsource, chunkedWriteLoop, queueSize)
proc newChunkedStreamWriter*[T](wsource: AsyncStreamWriter,
queueSize = AsyncStreamDefaultQueueSize,
udata: ref T): ChunkedStreamWriter =
var res = ChunkedStreamWriter()
res.init(wsource, queueSize, udata)
res
proc newChunkedStreamWriter*(wsource: AsyncStreamWriter,
queueSize = AsyncStreamDefaultQueueSize,
): ChunkedStreamWriter =
var res = ChunkedStreamWriter()
res.init(wsource, queueSize)
res