mirror of
https://github.com/status-im/nim-dagger.git
synced 2025-01-10 22:55:47 +00:00
f24ded0f76
The initial goal of this patch was to allow to download of a file via REST API in exactly the same size as it was uploaded, which required adding fields Chunker.offset and Manifest.originalBytes to keep that size. On top of that, we added more integrity checks to operations on Manifest, and reorganized TestNode.nim to test the actual interaction between node.store and node.retrieve operations. Note that the wire format of Manifest was changed, so we need to recreate all BlockStores. * Download without padding * Fixed chunker tests * Chunker: get rid of RabinChunker * Verify offset in the chunker tests * Use manifest.originalBytesPadded in StoreStream.size * StoreStream: replace emptyBlock with zeroMem * Manifest.bytes: compute how many bytes corresponding StoreStream(Manifest, pad) will return * Manifest: verify originalBytes and originalLen on new/encode/decode Also set originalBytes in each Manifest creation/update scenario * Manifest: comments, split code into sections * Reordered parameters to deal with int64 size in 32-bit builds * TestNode.nim: combine Store and Retrieve tests 1. Instead of copy-pasting code from node.nim, new test calls node.store() and node.retrieve() in order to check that they can correctly store and then retrieve data 2. New test compares only file contents, manifest contents considered an implementation detail 3. New test chunks at odd chunkSize=BlockSize/1.618 in order to ensure that data retrieved correctly even when buffer sizes mismatch * TestNode.nim: code refactoring * Manifest.add: one more test * Manifest.verify: return Result instead of raising Defect * Node.store: added blockSize parameter
194 lines
5.7 KiB
Nim
194 lines
5.7 KiB
Nim
## Nim-Codex
|
|
## Copyright (c) 2022 Status Research & Development GmbH
|
|
## Licensed under either of
|
|
## * Apache License, version 2.0, ([LICENSE-APACHE](LICENSE-APACHE))
|
|
## * MIT license ([LICENSE-MIT](LICENSE-MIT))
|
|
## at your option.
|
|
## This file may not be copied, modified, or distributed except according to
|
|
## those terms.
|
|
|
|
# This module implements serialization and deserialization of Manifest
|
|
|
|
import pkg/upraises
|
|
|
|
push: {.upraises: [].}
|
|
|
|
import std/tables
|
|
|
|
import pkg/libp2p
|
|
import pkg/questionable
|
|
import pkg/questionable/results
|
|
import pkg/chronicles
|
|
import pkg/chronos
|
|
|
|
import ./manifest
|
|
import ../errors
|
|
import ../blocktype
|
|
import ./types
|
|
|
|
func encode*(_: DagPBCoder, manifest: Manifest): ?!seq[byte] =
|
|
## Encode the manifest into a ``ManifestCodec``
|
|
## multicodec container (Dag-pb) for now
|
|
##
|
|
|
|
? manifest.verify()
|
|
var pbNode = initProtoBuffer()
|
|
|
|
for c in manifest.blocks:
|
|
var pbLink = initProtoBuffer()
|
|
pbLink.write(1, c.data.buffer) # write Cid links
|
|
pbLink.finish()
|
|
pbNode.write(2, pbLink)
|
|
|
|
# NOTE: The `Data` field in the the `dag-pb`
|
|
# contains the following protobuf `Message`
|
|
#
|
|
# ```protobuf
|
|
# Message ErasureInfo {
|
|
# optional uint32 K = 1; # number of encoded blocks
|
|
# optional uint32 M = 2; # number of parity blocks
|
|
# optional bytes cid = 3; # cid of the original dataset
|
|
# optional uint32 original = 4; # number of original blocks
|
|
# }
|
|
# Message Header {
|
|
# optional bytes rootHash = 1; # the root (tree) hash
|
|
# optional uint32 blockSize = 2; # size of a single block
|
|
# optional uint32 blocksLen = 3; # total amount of blocks
|
|
# optional ErasureInfo erasure = 4; # erasure coding info
|
|
# optional uint64 originalBytes = 5;# exact file size
|
|
# }
|
|
# ```
|
|
#
|
|
|
|
let cid = !manifest.rootHash
|
|
var header = initProtoBuffer()
|
|
header.write(1, cid.data.buffer)
|
|
header.write(2, manifest.blockSize.uint32)
|
|
header.write(3, manifest.len.uint32)
|
|
header.write(5, manifest.originalBytes.uint64)
|
|
if manifest.protected:
|
|
var erasureInfo = initProtoBuffer()
|
|
erasureInfo.write(1, manifest.K.uint32)
|
|
erasureInfo.write(2, manifest.M.uint32)
|
|
erasureInfo.write(3, manifest.originalCid.data.buffer)
|
|
erasureInfo.write(4, manifest.originalLen.uint32)
|
|
erasureInfo.finish()
|
|
|
|
header.write(4, erasureInfo)
|
|
|
|
pbNode.write(1, header) # set the rootHash Cid as the data field
|
|
pbNode.finish()
|
|
|
|
return pbNode.buffer.success
|
|
|
|
func decode*(_: DagPBCoder, data: openArray[byte]): ?!Manifest =
|
|
## Decode a manifest from a data blob
|
|
##
|
|
|
|
var
|
|
pbNode = initProtoBuffer(data)
|
|
pbHeader: ProtoBuffer
|
|
pbErasureInfo: ProtoBuffer
|
|
rootHash: seq[byte]
|
|
originalCid: seq[byte]
|
|
originalBytes: uint64
|
|
blockSize: uint32
|
|
blocksLen: uint32
|
|
originalLen: uint32
|
|
K, M: uint32
|
|
blocks: seq[Cid]
|
|
|
|
# Decode `Header` message
|
|
if pbNode.getField(1, pbHeader).isErr:
|
|
return failure("Unable to decode `Header` from dag-pb manifest!")
|
|
|
|
# Decode `Header` contents
|
|
if pbHeader.getField(1, rootHash).isErr:
|
|
return failure("Unable to decode `rootHash` from manifest!")
|
|
|
|
if pbHeader.getField(2, blockSize).isErr:
|
|
return failure("Unable to decode `blockSize` from manifest!")
|
|
|
|
if pbHeader.getField(3, blocksLen).isErr:
|
|
return failure("Unable to decode `blocksLen` from manifest!")
|
|
|
|
if pbHeader.getField(5, originalBytes).isErr:
|
|
return failure("Unable to decode `originalBytes` from manifest!")
|
|
|
|
if pbHeader.getField(4, pbErasureInfo).isErr:
|
|
return failure("Unable to decode `erasureInfo` from manifest!")
|
|
|
|
if pbErasureInfo.buffer.len > 0:
|
|
if pbErasureInfo.getField(1, K).isErr:
|
|
return failure("Unable to decode `K` from manifest!")
|
|
|
|
if pbErasureInfo.getField(2, M).isErr:
|
|
return failure("Unable to decode `M` from manifest!")
|
|
|
|
if pbErasureInfo.getField(3, originalCid).isErr:
|
|
return failure("Unable to decode `originalCid` from manifest!")
|
|
|
|
if pbErasureInfo.getField(4, originalLen).isErr:
|
|
return failure("Unable to decode `originalLen` from manifest!")
|
|
|
|
let rootHashCid = ? Cid.init(rootHash).mapFailure
|
|
var linksBuf: seq[seq[byte]]
|
|
if pbNode.getRepeatedField(2, linksBuf).isOk:
|
|
for pbLinkBuf in linksBuf:
|
|
var
|
|
blocksBuf: seq[seq[byte]]
|
|
blockBuf: seq[byte]
|
|
pbLink = initProtoBuffer(pbLinkBuf)
|
|
|
|
if pbLink.getField(1, blockBuf).isOk:
|
|
blocks.add(? Cid.init(blockBuf).mapFailure)
|
|
|
|
if blocksLen.int != blocks.len:
|
|
return failure("Total blocks and length of blocks in header don't match!")
|
|
|
|
var
|
|
self = Manifest(
|
|
rootHash: rootHashCid.some,
|
|
originalBytes: originalBytes.int,
|
|
blockSize: blockSize.int,
|
|
blocks: blocks,
|
|
hcodec: (? rootHashCid.mhash.mapFailure).mcodec,
|
|
codec: rootHashCid.mcodec,
|
|
version: rootHashCid.cidver,
|
|
protected: pbErasureInfo.buffer.len > 0)
|
|
|
|
if self.protected:
|
|
self.K = K.int
|
|
self.M = M.int
|
|
self.originalCid = ? Cid.init(originalCid).mapFailure
|
|
self.originalLen = originalLen.int
|
|
|
|
? self.verify()
|
|
self.success
|
|
|
|
proc encode*(
|
|
self: Manifest,
|
|
encoder = ManifestContainers[$DagPBCodec]): ?!seq[byte] =
|
|
## Encode a manifest using `encoder`
|
|
##
|
|
|
|
if self.rootHash.isNone:
|
|
? self.makeRoot()
|
|
|
|
encoder.encode(self)
|
|
|
|
func decode*(
|
|
_: type Manifest,
|
|
data: openArray[byte],
|
|
decoder = ManifestContainers[$DagPBCodec]): ?!Manifest =
|
|
## Decode a manifest using `decoder`
|
|
##
|
|
|
|
decoder.decode(data)
|
|
|
|
func decode*(_: type Manifest, blk: Block): ?!Manifest =
|
|
without contentType =? blk.cid.contentType() and
|
|
containerType =? ManifestContainers.?[$contentType]:
|
|
return failure "CID has invalid content type for manifest"
|
|
Manifest.decode(blk.data, containerType)
|