Merge branch 'rest/api-constraints' into fix/iter-bug
This commit is contained in:
commit
00366466ee
|
@ -27,6 +27,7 @@ import ../blocktype as bt
|
|||
import ../utils
|
||||
import ../utils/asynciter
|
||||
import ../indexingstrategy
|
||||
import ../errors
|
||||
|
||||
import pkg/stew/byteutils
|
||||
|
||||
|
@ -82,6 +83,9 @@ type
|
|||
blocksCount: Natural
|
||||
strategy: StrategyType
|
||||
|
||||
ErasureError* = object of CodexError
|
||||
InsufficientBlocksError* = object of ErasureError
|
||||
|
||||
func indexToPos(steps, idx, step: int): int {.inline.} =
|
||||
## Convert an index to a position in the encoded
|
||||
## dataset
|
||||
|
@ -236,11 +240,12 @@ proc init*(
|
|||
ecK: Natural, ecM: Natural,
|
||||
strategy: StrategyType): ?!EncodingParams =
|
||||
if ecK > manifest.blocksCount:
|
||||
return failure(
|
||||
let exc = newException(InsufficientBlocksError,
|
||||
"Unable to encode manifest, not enough blocks, ecK = " &
|
||||
$ecK &
|
||||
", blocksCount = " &
|
||||
$manifest.blocksCount)
|
||||
return failure(exc)
|
||||
|
||||
let
|
||||
rounded = roundUp(manifest.blocksCount, ecK)
|
||||
|
|
|
@ -265,9 +265,6 @@ proc streamEntireDataset(
|
|||
return failure(exc.msg)
|
||||
# --------------------------------------------------------------------------
|
||||
|
||||
GC_fullCollect()
|
||||
echo "Collect 6"
|
||||
|
||||
# Retrieve all blocks of the dataset sequentially from the local store or network
|
||||
trace "Creating store stream for manifest", manifestCid
|
||||
LPStream(StoreStream.new(self.networkStore, manifest, pad = false)).success
|
||||
|
|
|
@ -32,6 +32,7 @@ import ../node
|
|||
import ../blocktype
|
||||
import ../conf
|
||||
import ../contracts
|
||||
import ../erasure/erasure
|
||||
import ../manifest
|
||||
import ../streams/asyncstreamwrapper
|
||||
import ../stores
|
||||
|
@ -432,8 +433,16 @@ proc initPurchasingApi(node: CodexNodeRef, router: var RestRouter) =
|
|||
let nodes = params.nodes |? 1
|
||||
let tolerance = params.tolerance |? 0
|
||||
|
||||
if (nodes - tolerance) < 1:
|
||||
return RestApiResponse.error(Http400, "Tolerance cannot be greater or equal than nodes (nodes - tolerance)")
|
||||
# prevent underflow
|
||||
if tolerance > nodes:
|
||||
return RestApiResponse.error(Http400, "Invalid parameters: `tolerance` cannot be greater than `nodes`")
|
||||
|
||||
let ecK = nodes - tolerance
|
||||
let ecM = tolerance # for readability
|
||||
|
||||
# ensure leopard constrainst of 1 < K ≥ M
|
||||
if ecK <= 1 or ecK < ecM:
|
||||
return RestApiResponse.error(Http400, "Invalid parameters: parameters must satify `1 < (nodes - tolerance) ≥ tolerance`")
|
||||
|
||||
without expiry =? params.expiry:
|
||||
return RestApiResponse.error(Http400, "Expiry required")
|
||||
|
@ -451,6 +460,9 @@ proc initPurchasingApi(node: CodexNodeRef, router: var RestRouter) =
|
|||
params.collateral,
|
||||
expiry), error:
|
||||
|
||||
if error of InsufficientBlocksError:
|
||||
return RestApiResponse.error(Http400, "Insufficient blocks, increased dataset size required")
|
||||
|
||||
return RestApiResponse.error(Http500, error.msg)
|
||||
|
||||
return RestApiResponse.response(purchaseId.toHex)
|
||||
|
|
|
@ -96,7 +96,7 @@ proc requestStorageRaw*(
|
|||
proofProbability: UInt256,
|
||||
collateral: UInt256,
|
||||
expiry: uint = 0,
|
||||
nodes: uint = 1,
|
||||
nodes: uint = 2,
|
||||
tolerance: uint = 0
|
||||
): Response =
|
||||
|
||||
|
@ -125,7 +125,7 @@ proc requestStorage*(
|
|||
proofProbability: UInt256,
|
||||
expiry: uint,
|
||||
collateral: UInt256,
|
||||
nodes: uint = 1,
|
||||
nodes: uint = 2,
|
||||
tolerance: uint = 0
|
||||
): ?!PurchaseId =
|
||||
## Call request storage REST endpoint
|
||||
|
|
|
@ -4,14 +4,10 @@ import ./marketplacesuite
|
|||
import ./nodeconfigs
|
||||
import ./hardhatconfig
|
||||
|
||||
marketplacesuite "EC bug":
|
||||
marketplacesuite "Bug #821 - node crashes during erasure coding":
|
||||
|
||||
test "should be able to create storage request and download dataset",
|
||||
NodeConfigs(
|
||||
# Uncomment to start Hardhat automatically, typically so logs can be
|
||||
# inspected locally
|
||||
hardhat: HardhatConfig().withLogFile().some,
|
||||
|
||||
clients:
|
||||
CodexConfigs.init(nodes=1)
|
||||
# .debug() # uncomment to enable console log output.debug()
|
||||
|
@ -38,7 +34,6 @@ marketplacesuite "EC bug":
|
|||
|
||||
var requestId = none RequestId
|
||||
proc onStorageRequested(event: StorageRequested) {.raises:[].} =
|
||||
echo "storage requested event"
|
||||
requestId = event.requestId.some
|
||||
|
||||
let subscription = await marketplace.subscribe(StorageRequested, onStorageRequested)
|
||||
|
|
|
@ -8,7 +8,8 @@ import ../examples
|
|||
twonodessuite "Purchasing", debug1 = false, debug2 = false:
|
||||
|
||||
test "node handles storage request":
|
||||
let cid = client1.upload("some file contents").get
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
let id1 = client1.requestStorage(cid, duration=100.u256, reward=2.u256, proofProbability=3.u256, expiry=10, collateral=200.u256).get
|
||||
let id2 = client1.requestStorage(cid, duration=400.u256, reward=5.u256, proofProbability=6.u256, expiry=10, collateral=201.u256).get
|
||||
check id1 != id2
|
||||
|
@ -26,7 +27,7 @@ twonodessuite "Purchasing", debug1 = false, debug2 = false:
|
|||
proofProbability=3.u256,
|
||||
expiry=30,
|
||||
collateral=200.u256,
|
||||
nodes=2,
|
||||
nodes=3,
|
||||
tolerance=1).get
|
||||
|
||||
let request = client1.getPurchase(id).get.request.get
|
||||
|
@ -35,7 +36,7 @@ twonodessuite "Purchasing", debug1 = false, debug2 = false:
|
|||
check request.ask.proofProbability == 3.u256
|
||||
check request.expiry == 30
|
||||
check request.ask.collateral == 200.u256
|
||||
check request.ask.slots == 2'u64
|
||||
check request.ask.slots == 3'u64
|
||||
check request.ask.maxSlotLoss == 1'u64
|
||||
|
||||
# TODO: We currently do not support encoding single chunks
|
||||
|
@ -52,7 +53,8 @@ twonodessuite "Purchasing", debug1 = false, debug2 = false:
|
|||
# check request.ask.maxSlotLoss == 1'u64
|
||||
|
||||
test "node remembers purchase status after restart":
|
||||
let cid = client1.upload("some file contents").get
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
let id = client1.requestStorage(cid,
|
||||
duration=100.u256,
|
||||
reward=2.u256,
|
||||
|
@ -71,25 +73,12 @@ twonodessuite "Purchasing", debug1 = false, debug2 = false:
|
|||
check request.ask.proofProbability == 3.u256
|
||||
check request.expiry == 30
|
||||
check request.ask.collateral == 200.u256
|
||||
check request.ask.slots == 1'u64
|
||||
check request.ask.slots == 2'u64
|
||||
check request.ask.maxSlotLoss == 0'u64
|
||||
|
||||
test "request storage fails if nodes and tolerance aren't correct":
|
||||
let cid = client1.upload("some file contents").get
|
||||
let responseBefore = client1.requestStorageRaw(cid,
|
||||
duration=100.u256,
|
||||
reward=2.u256,
|
||||
proofProbability=3.u256,
|
||||
expiry=30,
|
||||
collateral=200.u256,
|
||||
nodes=1,
|
||||
tolerance=1)
|
||||
|
||||
check responseBefore.status == "400 Bad Request"
|
||||
check responseBefore.body == "Tolerance cannot be greater or equal than nodes (nodes - tolerance)"
|
||||
|
||||
test "node requires expiry and its value to be in future":
|
||||
let cid = client1.upload("some file contents").get
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
|
||||
let responseMissing = client1.requestStorageRaw(cid, duration=1.u256, reward=2.u256, proofProbability=3.u256, collateral=200.u256)
|
||||
check responseMissing.status == "400 Bad Request"
|
||||
|
|
|
@ -1,7 +1,9 @@
|
|||
import std/httpclient
|
||||
import std/sequtils
|
||||
from pkg/libp2p import `==`
|
||||
import pkg/codex/units
|
||||
import ./twonodes
|
||||
import ../examples
|
||||
|
||||
twonodessuite "REST API", debug1 = false, debug2 = false:
|
||||
|
||||
|
@ -36,3 +38,93 @@ twonodessuite "REST API", debug1 = false, debug2 = false:
|
|||
|
||||
check:
|
||||
[cid1, cid2].allIt(it in list.content.mapIt(it.cid))
|
||||
|
||||
test "request storage fails for datasets that are too small":
|
||||
let cid = client1.upload("some file contents").get
|
||||
let response = client1.requestStorageRaw(cid, duration=10.u256, reward=2.u256, proofProbability=3.u256, collateral=200.u256, expiry=9)
|
||||
|
||||
check:
|
||||
response.status == "400 Bad Request"
|
||||
response.body == "Insufficient blocks, increased dataset size required"
|
||||
|
||||
test "request storage succeeds for sufficiently sized datasets":
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
let response = client1.requestStorageRaw(cid, duration=10.u256, reward=2.u256, proofProbability=3.u256, collateral=200.u256, expiry=9)
|
||||
|
||||
check:
|
||||
response.status == "200 OK"
|
||||
|
||||
test "request storage fails if nodes and tolerance aren't correct":
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
let duration = 100.u256
|
||||
let reward = 2.u256
|
||||
let proofProbability = 3.u256
|
||||
let expiry = 30.uint
|
||||
let collateral = 200.u256
|
||||
let ecParams = @[(1, 0), (1, 1), (2, 1), (3, 2), (3, 3)]
|
||||
|
||||
for ecParam in ecParams:
|
||||
let (nodes, tolerance) = ecParam
|
||||
|
||||
var responseBefore = client1.requestStorageRaw(cid,
|
||||
duration,
|
||||
reward,
|
||||
proofProbability,
|
||||
collateral,
|
||||
expiry,
|
||||
nodes.uint,
|
||||
tolerance.uint)
|
||||
|
||||
check responseBefore.status == "400 Bad Request"
|
||||
check responseBefore.body == "Invalid parameters: parameters must satify `1 < (nodes - tolerance) ≥ tolerance`"
|
||||
|
||||
test "request storage fails if tolerance > nodes (underflow protection)":
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
let duration = 100.u256
|
||||
let reward = 2.u256
|
||||
let proofProbability = 3.u256
|
||||
let expiry = 30.uint
|
||||
let collateral = 200.u256
|
||||
let ecParams = @[(0, 1), (1, 2), (2, 3)]
|
||||
|
||||
for ecParam in ecParams:
|
||||
let (nodes, tolerance) = ecParam
|
||||
|
||||
var responseBefore = client1.requestStorageRaw(cid,
|
||||
duration,
|
||||
reward,
|
||||
proofProbability,
|
||||
collateral,
|
||||
expiry,
|
||||
nodes.uint,
|
||||
tolerance.uint)
|
||||
|
||||
check responseBefore.status == "400 Bad Request"
|
||||
check responseBefore.body == "Invalid parameters: `tolerance` cannot be greater than `nodes`"
|
||||
|
||||
test "request storage succeeds if nodes and tolerance within range":
|
||||
let data = await RandomChunker.example(blocks=2)
|
||||
let cid = client1.upload(data).get
|
||||
let duration = 100.u256
|
||||
let reward = 2.u256
|
||||
let proofProbability = 3.u256
|
||||
let expiry = 30.uint
|
||||
let collateral = 200.u256
|
||||
let ecParams = @[(2, 0), (3, 1), (5, 2)]
|
||||
|
||||
for ecParam in ecParams:
|
||||
let (nodes, tolerance) = ecParam
|
||||
|
||||
var responseBefore = client1.requestStorageRaw(cid,
|
||||
duration,
|
||||
reward,
|
||||
proofProbability,
|
||||
collateral,
|
||||
expiry,
|
||||
nodes.uint,
|
||||
tolerance.uint)
|
||||
|
||||
check responseBefore.status == "200 OK"
|
||||
|
|
Loading…
Reference in New Issue