review and refactor validator integration tests
This commit is contained in:
parent
2404b75f0d
commit
60142ea061
|
@ -1,4 +1,5 @@
|
|||
from std/times import inMilliseconds, initDuration, inSeconds, fromUnix
|
||||
import std/sets
|
||||
import pkg/codex/logutils
|
||||
import ../contracts/time
|
||||
import ../contracts/deployment
|
||||
|
@ -13,188 +14,50 @@ logScope:
|
|||
topics = "integration test validation"
|
||||
|
||||
marketplacesuite "Validaton":
|
||||
let nodes = 3
|
||||
let tolerance = 1
|
||||
var slotsFilled: seq[SlotId]
|
||||
var slotsFreed: seq[SlotId]
|
||||
|
||||
test "validator uses historical state to mark missing proofs", NodeConfigs(
|
||||
# Uncomment to start Hardhat automatically, typically so logs can be inspected locally
|
||||
hardhat:
|
||||
HardhatConfig.none,
|
||||
|
||||
clients:
|
||||
CodexConfigs.init(nodes=1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("node", "marketplace", "clock")
|
||||
.some,
|
||||
|
||||
providers:
|
||||
CodexConfigs.init(nodes=1)
|
||||
.withSimulateProofFailures(idx=0, failEveryNProofs=1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("marketplace", "sales", "reservations", "node", "clock", "slotsbuilder")
|
||||
.some
|
||||
):
|
||||
let client0 = clients()[0].client
|
||||
let expiry = 5.periods
|
||||
let duration = expiry + 10.periods
|
||||
|
||||
let data = await RandomChunker.example(blocks=8)
|
||||
createAvailabilities(data.len * 2, duration) # TODO: better value for data.len
|
||||
|
||||
let cid = client0.upload(data).get
|
||||
|
||||
let purchaseId = await client0.requestStorage(
|
||||
cid,
|
||||
expiry=expiry,
|
||||
duration=duration,
|
||||
nodes=3,
|
||||
tolerance=1,
|
||||
proofProbability=1
|
||||
)
|
||||
let requestId = client0.requestId(purchaseId).get
|
||||
|
||||
check eventually(client0.purchaseStateIs(purchaseId, "started"),
|
||||
timeout = expiry.int * 1000)
|
||||
|
||||
var validators = CodexConfigs.init(nodes=2)
|
||||
.withValidationGroups(groups = 2)
|
||||
.withValidationGroupIndex(idx = 0, groupIndex = 0)
|
||||
.withValidationGroupIndex(idx = 1, groupIndex = 1)
|
||||
.debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
.withLogTopics("validator") # each topic as a separate string argument
|
||||
|
||||
failAndTeardownOnError "failed to start validator nodes":
|
||||
for config in validators.configs.mitems:
|
||||
let node = await startValidatorNode(config)
|
||||
running.add RunningNode(
|
||||
role: Role.Validator,
|
||||
node: node
|
||||
)
|
||||
|
||||
var slotWasFreed = false
|
||||
proc onSlotFreed(event: SlotFreed) =
|
||||
if event.requestId == requestId:
|
||||
slotWasFreed = true
|
||||
|
||||
let subscription = await marketplace.subscribe(SlotFreed, onSlotFreed)
|
||||
|
||||
check eventually(slotWasFreed, timeout=(duration - expiry).int * 1000)
|
||||
|
||||
await subscription.unsubscribe()
|
||||
|
||||
test "validator only looks 30 days back for historical state", NodeConfigs(
|
||||
# Uncomment to start Hardhat automatically, typically so logs can be inspected locally
|
||||
hardhat:
|
||||
HardhatConfig.none,
|
||||
|
||||
clients:
|
||||
CodexConfigs.init(nodes=1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("node", "marketplace", "clock")
|
||||
.some,
|
||||
|
||||
providers:
|
||||
CodexConfigs.init(nodes=1)
|
||||
.withSimulateProofFailures(idx=0, failEveryNProofs=1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("marketplace", "sales", "reservations", "node", "clock", "slotsbuilder")
|
||||
.some
|
||||
):
|
||||
let client0 = clients()[0].client
|
||||
let expiry = 5.periods
|
||||
let duration30days = initDuration(days = 30)
|
||||
let duration = expiry + duration30days.inSeconds.uint64 + 10.periods
|
||||
|
||||
let data = await RandomChunker.example(blocks=8)
|
||||
createAvailabilities(data.len * 2, duration) # TODO: better value for data.len
|
||||
|
||||
let cid = client0.upload(data).get
|
||||
|
||||
var currentTime = await ethProvider.currentTime()
|
||||
|
||||
let expiryEndTime = currentTime.truncate(uint64) + expiry
|
||||
let requestEndTime = currentTime.truncate(uint64) + duration
|
||||
debug "test validator: ", currentTime = currentTime.truncate(uint64),
|
||||
requestEndTime = requestEndTime, expiryEndTime = expiryEndTime
|
||||
debug "test validator: ", currentTime = currentTime.truncate(int64).fromUnix,
|
||||
requestEndTime = requestEndTime.int64.fromUnix,
|
||||
expiryEndTime = expiryEndTime.int64.fromUnix
|
||||
|
||||
proc trackSlotsFilled(marketplace: Marketplace):
|
||||
Future[provider.Subscription] {.async.} =
|
||||
slotsFilled = newSeq[SlotId]()
|
||||
proc onSlotFilled(event: SlotFilled) =
|
||||
let slotId = slotId(event.requestId, event.slotIndex)
|
||||
slotsFilled.add(slotId)
|
||||
debug "SlotFilled", requestId = event.requestId, slotIndex = event.slotIndex,
|
||||
slotId = slotId
|
||||
|
||||
let subscriptionOnSlotFilled = await marketplace.subscribe(SlotFilled, onSlotFilled)
|
||||
|
||||
let purchaseId = await client0.requestStorage(
|
||||
cid,
|
||||
expiry=expiry,
|
||||
duration=duration,
|
||||
nodes=3,
|
||||
tolerance=1,
|
||||
reward=1.u256,
|
||||
proofProbability=1
|
||||
)
|
||||
let requestId = client0.requestId(purchaseId).get
|
||||
|
||||
check eventually(client0.purchaseStateIs(purchaseId, "started"),
|
||||
timeout = expiry.int * 1000)
|
||||
|
||||
currentTime = await ethProvider.currentTime()
|
||||
var waitTime = (expiryEndTime - currentTime.truncate(uint64)).int.seconds
|
||||
debug "test validation - waiting till end of expiry", waitTime = waitTime
|
||||
await sleepAsync(waitTime)
|
||||
|
||||
discard await ethProvider.send("evm_mine")
|
||||
|
||||
await ethProvider.advanceTimeTo(
|
||||
expiryEndTime.u256 + duration30days.inSeconds.u256)
|
||||
debug "test validator[after advance]: ", currentTime = currentTime.truncate(SecondsSince1970)
|
||||
debug "test validator[after advance]: ", currentTime =
|
||||
currentTime.truncate(SecondsSince1970).fromUnix
|
||||
|
||||
var validators = CodexConfigs.init(nodes=1)
|
||||
.debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
.withLogTopics("validator", "clock", "market") # each topic as a separate string argument
|
||||
|
||||
failAndTeardownOnError "failed to start validator nodes":
|
||||
for config in validators.configs.mitems:
|
||||
let node = await startValidatorNode(config)
|
||||
running.add RunningNode(
|
||||
role: Role.Validator,
|
||||
node: node
|
||||
)
|
||||
|
||||
var slotWasFreed = false
|
||||
let subscription = await marketplace.subscribe(SlotFilled, onSlotFilled)
|
||||
subscription
|
||||
|
||||
proc trackSlotsFreed(requestId: RequestId, marketplace: Marketplace):
|
||||
Future[provider.Subscription] {.async.} =
|
||||
slotsFreed = newSeq[SlotId]()
|
||||
proc onSlotFreed(event: SlotFreed) =
|
||||
if event.requestId == requestId:
|
||||
slotWasFreed = true
|
||||
let slotId = slotId(event.requestId, event.slotIndex)
|
||||
slotsFreed.add(slotId)
|
||||
debug "onSlotFreed", requestId = requestId, slotIndex = event.slotIndex,
|
||||
slotId = slotId, slotsFreed = slotsFreed.len
|
||||
|
||||
let subscription = await marketplace.subscribe(SlotFreed, onSlotFreed)
|
||||
subscription
|
||||
|
||||
# check not freed
|
||||
currentTime = await ethProvider.currentTime()
|
||||
if requestEndTime > currentTime.truncate(uint64):
|
||||
waitTime = (requestEndTime - currentTime.truncate(uint64)).int.seconds
|
||||
debug "test validation - waiting for request end", waitTime = waitTime
|
||||
await sleepAsync(waitTime)
|
||||
proc checkSlotsFailed(slotsFilled: seq[SlotId],
|
||||
slotsFreed: seq[SlotId], marketplace: Marketplace) {.async.} =
|
||||
let slotsNotFreed = slotsFilled.filter(
|
||||
slotId => not slotsFreed.contains(slotId)
|
||||
).toHashSet
|
||||
var slotsFailed = initHashSet[SlotId]()
|
||||
for slotId in slotsFilled:
|
||||
let state = await marketplace.slotState(slotId)
|
||||
if state == SlotState.Failed:
|
||||
slotsFailed.incl(slotId)
|
||||
|
||||
debug "test validation - request ended"
|
||||
|
||||
check not slotWasFreed
|
||||
|
||||
# check eventually(client0.purchaseStateIs(purchaseId, "finished"),
|
||||
# timeout = 60 * 1000)
|
||||
|
||||
await subscription.unsubscribe()
|
||||
await subscriptionOnSlotFilled.unsubscribe()
|
||||
debug "test validation - unsubscribed"
|
||||
|
||||
debug "slots failed", slotsFailed = slotsFailed, slotsNotFreed = slotsNotFreed
|
||||
check slotsNotFreed == slotsFailed
|
||||
|
||||
test "validator marks proofs as missing when using validation groups", NodeConfigs(
|
||||
# Uncomment to start Hardhat automatically, typically so logs can be inspected locally
|
||||
hardhat:
|
||||
|
@ -220,9 +83,9 @@ marketplacesuite "Validaton":
|
|||
.withValidationGroups(groups = 2)
|
||||
.withValidationGroupIndex(idx = 0, groupIndex = 0)
|
||||
.withValidationGroupIndex(idx = 1, groupIndex = 1)
|
||||
.debug() # uncomment to enable console log output
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
.withLogTopics("validator")
|
||||
# .withLogTopics("validator")
|
||||
# .withLogTopics("validator", "integration", "ethers", "clock")
|
||||
.some
|
||||
):
|
||||
|
@ -231,7 +94,12 @@ marketplacesuite "Validaton":
|
|||
let duration = expiry + 10.periods
|
||||
|
||||
let data = await RandomChunker.example(blocks=8)
|
||||
createAvailabilities(data.len * 2, duration) # TODO: better value for data.len
|
||||
|
||||
# TODO: better value for data.len below. This TODO is also present in
|
||||
# testproofs.nim - we may want to address it or remove the comment.
|
||||
createAvailabilities(data.len * 2, duration)
|
||||
|
||||
let slotFilledSubscription = await trackSlotsFilled(marketplace)
|
||||
|
||||
let cid = client0.upload(data).get
|
||||
|
||||
|
@ -239,21 +107,109 @@ marketplacesuite "Validaton":
|
|||
cid,
|
||||
expiry=expiry,
|
||||
duration=duration,
|
||||
nodes=3,
|
||||
tolerance=1,
|
||||
nodes=nodes,
|
||||
tolerance=tolerance,
|
||||
proofProbability=1
|
||||
)
|
||||
let requestId = client0.requestId(purchaseId).get
|
||||
|
||||
check eventually(client0.purchaseStateIs(purchaseId, "started"), timeout = expiry.int * 1000)
|
||||
check eventually(client0.purchaseStateIs(purchaseId, "started"),
|
||||
timeout = expiry.int * 1000)
|
||||
|
||||
let slotFreedSubscription =
|
||||
await trackSlotsFreed(requestId, marketplace)
|
||||
|
||||
var slotWasFreed = false
|
||||
proc onSlotFreed(event: SlotFreed) =
|
||||
if event.requestId == requestId:
|
||||
slotWasFreed = true
|
||||
let expectedSlotsFreed = nodes - tolerance
|
||||
check eventually((slotsFreed.len == expectedSlotsFreed),
|
||||
timeout=(duration - expiry).int * 1000)
|
||||
|
||||
# Because of erasure coding, if e.g. 2 out of 3 nodes are freed, the last
|
||||
# node will not be freed but marked as "Failed" because the whole request
|
||||
# will fail. For this reason we need an extra check:
|
||||
await checkSlotsFailed(slotsFilled, slotsFreed, marketplace)
|
||||
|
||||
await slotFilledSubscription.unsubscribe()
|
||||
await slotFreedSubscription.unsubscribe()
|
||||
|
||||
test "validator uses historical state to mark missing proofs", NodeConfigs(
|
||||
# Uncomment to start Hardhat automatically, typically so logs can be inspected locally
|
||||
hardhat:
|
||||
HardhatConfig.none,
|
||||
|
||||
let subscription = await marketplace.subscribe(SlotFreed, onSlotFreed)
|
||||
clients:
|
||||
CodexConfigs.init(nodes=1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("node", "marketplace", "clock")
|
||||
.some,
|
||||
|
||||
check eventually(slotWasFreed, timeout=(duration - expiry).int * 1000)
|
||||
providers:
|
||||
CodexConfigs.init(nodes=1)
|
||||
.withSimulateProofFailures(idx=0, failEveryNProofs=1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("marketplace", "sales", "reservations", "node", "clock", "slotsbuilder")
|
||||
.some
|
||||
):
|
||||
let client0 = clients()[0].client
|
||||
let expiry = 5.periods
|
||||
let duration = expiry + 10.periods
|
||||
|
||||
await subscription.unsubscribe()
|
||||
let data = await RandomChunker.example(blocks=8)
|
||||
|
||||
# TODO: better value for data.len below. This TODO is also present in
|
||||
# testproofs.nim - we may want to address it or remove the comment.
|
||||
createAvailabilities(data.len * 2, duration)
|
||||
|
||||
let slotFilledSubscription = await trackSlotsFilled(marketplace)
|
||||
|
||||
let cid = client0.upload(data).get
|
||||
|
||||
let purchaseId = await client0.requestStorage(
|
||||
cid,
|
||||
expiry=expiry,
|
||||
duration=duration,
|
||||
nodes=nodes,
|
||||
tolerance=tolerance,
|
||||
proofProbability=1
|
||||
)
|
||||
let requestId = client0.requestId(purchaseId).get
|
||||
|
||||
check eventually(client0.purchaseStateIs(purchaseId, "started"),
|
||||
timeout = expiry.int * 1000)
|
||||
|
||||
# just to make sure we have a mined block that separates us
|
||||
# from the block containing the last SlotFilled event
|
||||
discard await ethProvider.send("evm_mine")
|
||||
|
||||
var validators = CodexConfigs.init(nodes=2)
|
||||
.withValidationGroups(groups = 2)
|
||||
.withValidationGroupIndex(idx = 0, groupIndex = 0)
|
||||
.withValidationGroupIndex(idx = 1, groupIndex = 1)
|
||||
# .debug() # uncomment to enable console log output
|
||||
# .withLogFile() # uncomment to output log file to:
|
||||
# tests/integration/logs/<start_datetime> <suite_name>/<test_name>/<node_role>_<node_idx>.log
|
||||
# .withLogTopics("validator") # each topic as a separate string argument
|
||||
|
||||
failAndTeardownOnError "failed to start validator nodes":
|
||||
for config in validators.configs.mitems:
|
||||
let node = await startValidatorNode(config)
|
||||
running.add RunningNode(
|
||||
role: Role.Validator,
|
||||
node: node
|
||||
)
|
||||
|
||||
let slotFreedSubscription =
|
||||
await trackSlotsFreed(requestId, marketplace)
|
||||
|
||||
let expectedSlotsFreed = nodes - tolerance
|
||||
check eventually((slotsFreed.len == expectedSlotsFreed),
|
||||
timeout=(duration - expiry).int * 1000)
|
||||
|
||||
# Because of erasure coding, if e.g. 2 out of 3 nodes are freed, the last
|
||||
# node will not be freed but marked as "Failed" because the whole request
|
||||
# will fail. For this reason we need an extra check:
|
||||
await checkSlotsFailed(slotsFilled, slotsFreed, marketplace)
|
||||
|
||||
await slotFilledSubscription.unsubscribe()
|
||||
await slotFreedSubscription.unsubscribe()
|
||||
|
|
Loading…
Reference in New Issue