mirror of
https://github.com/logos-storage/logos-storage-proofs-circuits.git
synced 2026-01-04 22:43:07 +00:00
it almost compiles (but does not link/builds) after the big hack refactoring
This commit is contained in:
parent
948f6c2c75
commit
14bab04b19
2
reference/nim/proof_input/.gitignore
vendored
2
reference/nim/proof_input/.gitignore
vendored
@ -2,4 +2,4 @@
|
|||||||
testmain
|
testmain
|
||||||
cli
|
cli
|
||||||
*.json
|
*.json
|
||||||
json/
|
json_tmp/
|
||||||
@ -4,9 +4,11 @@ author = "Balazs Komuves"
|
|||||||
description = "reference implementation for generating the proof inputs"
|
description = "reference implementation for generating the proof inputs"
|
||||||
license = "MIT or Apache-2.0"
|
license = "MIT or Apache-2.0"
|
||||||
srcDir = "src"
|
srcDir = "src"
|
||||||
bin = @["cli","testmain"]
|
bin = @["cli"]
|
||||||
|
#bin = @["cli","testmain"]
|
||||||
|
|
||||||
requires "nim >= 1.6.0"
|
requires "nim >= 1.6.0"
|
||||||
requires "https://github.com/mratsim/constantine#ab6fa6ae1bbbd1b10071a92ec209b381b5d82511"
|
requires "https://github.com/mratsim/constantine#ab6fa6ae1bbbd1b10071a92ec209b381b5d82511"
|
||||||
requires "https://github.com/codex-storage/nim-poseidon2#8a54c69032a741160bbc097d009e45a8b5e4d718"
|
requires "https://github.com/codex-storage/nim-poseidon2#8a54c69032a741160bbc097d009e45a8b5e4d718"
|
||||||
requires "https://github.com/codex-storage/nim-goldilocks-hash"
|
requires "https://github.com/codex-storage/nim-goldilocks-hash#0d12b1429345c3f359df47171332e8966dc94ed3"
|
||||||
|
#requires "../../../../nim-goldilocks-hash/"
|
||||||
|
|||||||
@ -5,18 +5,28 @@ import std/sequtils
|
|||||||
#import poseidon2/types
|
#import poseidon2/types
|
||||||
import poseidon2/io
|
import poseidon2/io
|
||||||
import poseidon2/sponge
|
import poseidon2/sponge
|
||||||
import poseidon2/merkle
|
#import poseidon2/merkle
|
||||||
|
|
||||||
import types
|
import ../types
|
||||||
import merkle
|
import ../types/bn254
|
||||||
|
#import ../merkle
|
||||||
|
import ../merkle/bn254
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
func hashCellOpen( globcfg: GlobalConfig, cellData: openArray[byte] ): Hash =
|
func merkleTree*( hashcfg: HashConfig, what: openarray[Hash]): MerkleTree[Hash] =
|
||||||
|
assert( hashcfg.combo == BN254_Poseidon2 )
|
||||||
|
return merkleTreeBN254( what )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func hashCellOpen( hashcfg: HashConfig, globcfg: GlobalConfig, cellData: openArray[byte] ): Hash =
|
||||||
|
assert( hashcfg.field == BN254 )
|
||||||
|
assert( hashcfg.hashFun == Poseidon2 )
|
||||||
assert( cellData.len == globcfg.cellSize , ("cells are expected to be exactly " & $globcfg.cellSize & " bytes") )
|
assert( cellData.len == globcfg.cellSize , ("cells are expected to be exactly " & $globcfg.cellSize & " bytes") )
|
||||||
return Sponge.digest( cellData, rate=2 )
|
return Sponge.digest( cellData, rate=2 )
|
||||||
|
|
||||||
func hashCell*( globcfg: GlobalConfig, cellData: Cell): Hash = hashCellOpen(globcfg, cellData)
|
func hashCell*(hashcfg: HashConfig, globcfg: GlobalConfig, cellData: Cell): Hash = hashCellOpen(hashcfg, globcfg, cellData)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
@ -36,23 +46,24 @@ func splitBlockIntoCells( globcfg: GlobalConfig, blockData: openArray[byte] ): s
|
|||||||
|
|
||||||
# returns the special hash of a network block (this is a Merkle root built on the
|
# returns the special hash of a network block (this is a Merkle root built on the
|
||||||
# top of the hashes of the 32 cells inside the block)
|
# top of the hashes of the 32 cells inside the block)
|
||||||
func hashNetworkBlockOpen( globcfg: GlobalConfig, blockData: openArray[byte] ): Hash =
|
func hashNetworkBlockOpen( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: openArray[byte] ): Hash =
|
||||||
let cells = splitBlockIntoCells(globcfg, blockData)
|
let cells = splitBlockIntoCells(globcfg, blockData)
|
||||||
let leaves = collect( newSeq , (for i in 0..<cellsPerBlock(globcfg): hashCell(globcfg, cells[i]) ))
|
let leaves = collect( newSeq , (for i in 0..<cellsPerBlock(globcfg): hashCell(hashcfg, globcfg, cells[i]) ))
|
||||||
return Merkle.digest(leaves) # merkleRoot(leaves)
|
return merkleDigestBN254(leaves) # Merkle.digest(leaves) # merkleRoot(leaves)
|
||||||
|
|
||||||
func hashNetworkBlock*(globcfg: GlobalConfig, blockData: Block): Hash =
|
func hashNetworkBlock*( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: Block): Hash =
|
||||||
hashNetworkBlockOpen(globcfg, blockData)
|
hashNetworkBlockOpen(hashcfg, globcfg, blockData)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
# returns the mini Merkle tree built on the 32 cells inside a network block
|
# returns the mini Merkle tree built on the 32 cells inside a network block
|
||||||
func networkBlockTreeOpen( globcfg: GlobalConfig, blockData: openArray[byte] ): MerkleTree =
|
func networkBlockTreeOpen( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: openArray[byte] ): MerkleTree[Hash] =
|
||||||
let cells = splitBlockIntoCells( globcfg, blockData)
|
assert( hashcfg.field == BN254 )
|
||||||
let leaves = collect( newSeq , (for i in 0..<cellsPerBlock(globcfg): hashCell( globcfg, cells[i]) ))
|
let cells = splitBlockIntoCells( globcfg, blockData )
|
||||||
return merkleTree(leaves)
|
let leaves = collect( newSeq , (for i in 0..<cellsPerBlock(globcfg): hashCell( hashcfg, globcfg, cells[i]) ))
|
||||||
|
return merkleTree(hashcfg, leaves)
|
||||||
|
|
||||||
func networkBlockTree*( globcfg: GlobalConfig, blockData: Block): MerkleTree =
|
func networkBlockTree*( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: Block): MerkleTree[Hash] =
|
||||||
networkBlockTreeOpen(globcfg, blockData)
|
networkBlockTreeOpen(hashcfg, globcfg, blockData)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
76
reference/nim/proof_input/src/blocks/goldilocks.nim
Normal file
76
reference/nim/proof_input/src/blocks/goldilocks.nim
Normal file
@ -0,0 +1,76 @@
|
|||||||
|
|
||||||
|
|
||||||
|
import sugar
|
||||||
|
import std/sequtils
|
||||||
|
|
||||||
|
import goldilocks_hash/types
|
||||||
|
#import goldilocks_hash/poseidon2/sponge
|
||||||
|
#import goldilocks_hash/monolith/sponge
|
||||||
|
|
||||||
|
import ../types
|
||||||
|
import ../types/goldilocks
|
||||||
|
#import ../merkle
|
||||||
|
import ../merkle/goldilocks/poseidon2
|
||||||
|
import ../merkle/goldilocks/monolith
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func merkleTree*( hashcfg: HashConfig, what: openarray[Digest] ): MerkleTree[Digest] =
|
||||||
|
assert( hashcfg.field == Goldilocks )
|
||||||
|
case hashcfg.hashFun:
|
||||||
|
of Poseidon2: return poseidon2.merkleTreeGoldilocksPoseidon2( what )
|
||||||
|
of Monolith: return monolith.merkleTreeGoldilocksMonolith( what )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func hashCellOpen( hashcfg: HashConfig, globcfg: GlobalConfig, cellData: openArray[byte] ): Digest =
|
||||||
|
assert( hashcfg.field == Goldilocks )
|
||||||
|
assert( cellData.len == globcfg.cellSize , ("cells are expected to be exactly " & $globcfg.cellSize & " bytes") )
|
||||||
|
case hashcfg.hashFun:
|
||||||
|
of Poseidon2: return digestBytesPoseidon2( cellData )
|
||||||
|
of Monolith: return digestBytesMonolith( cellData )
|
||||||
|
|
||||||
|
func hashCell*(hashcfg: HashConfig, globcfg: GlobalConfig, cellData: Cell): Digest =
|
||||||
|
hashCellOpen(hashcfg, globcfg, cellData)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func splitBlockIntoCells( globcfg: GlobalConfig, blockData: openArray[byte] ): seq[Cell] =
|
||||||
|
assert( blockData.len == globcfg.blockSize , ("network blocks are expected to be exactly" & $globcfg.blockSize & " bytes" ) )
|
||||||
|
|
||||||
|
var cells : seq[seq[byte]] = newSeq[seq[byte]]( cellsPerBlock(globcfg) )
|
||||||
|
|
||||||
|
let start = low(blockData)
|
||||||
|
var leaves : seq[Digest] = newSeq[Digest]( cellsPerBlock(globcfg) )
|
||||||
|
for i in 0..<cellsPerBlock(globcfg):
|
||||||
|
let a = start + i * globcfg.cellSize
|
||||||
|
let b = start + (i+1) * globcfg.cellSize
|
||||||
|
cells[i] = blockData[a..<b].toSeq()
|
||||||
|
|
||||||
|
return cells
|
||||||
|
|
||||||
|
# returns the special hash of a network block (this is a Merkle root built on the
|
||||||
|
# top of the hashes of the 32 cells inside the block)
|
||||||
|
func hashNetworkBlockOpen( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: openArray[byte] ): Digest =
|
||||||
|
let cells = splitBlockIntoCells(globcfg, blockData)
|
||||||
|
let leaves = collect( newSeq , (for i in 0..<cellsPerBlock(globcfg): hashCell(hashcfg, globcfg, cells[i]) ))
|
||||||
|
case hashcfg.hashFun:
|
||||||
|
of Poseidon2: return merkleDigestPoseidon2(leaves)
|
||||||
|
of Monolith: return merkleDigestMonolith( leaves)
|
||||||
|
|
||||||
|
func hashNetworkBlock*( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: Block): Digest =
|
||||||
|
hashNetworkBlockOpen(hashcfg, globcfg, blockData)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
# returns the mini Merkle tree built on the 32 cells inside a network block
|
||||||
|
func networkBlockTreeOpen( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: openArray[byte] ): MerkleTree[Digest] =
|
||||||
|
assert( hashcfg.field == Goldilocks )
|
||||||
|
let cells = splitBlockIntoCells( globcfg, blockData )
|
||||||
|
let leaves = collect( newSeq , (for i in 0..<cellsPerBlock(globcfg): hashCell( hashcfg, globcfg, cells[i]) ))
|
||||||
|
return merkleTree(hashcfg, leaves)
|
||||||
|
|
||||||
|
func networkBlockTree*( hashcfg: HashConfig, globcfg: GlobalConfig, blockData: Block): MerkleTree[Digest] =
|
||||||
|
networkBlockTreeOpen(hashcfg, globcfg, blockData)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
@ -7,20 +7,29 @@ import std/sequtils
|
|||||||
import std/os
|
import std/os
|
||||||
import std/parseopt
|
import std/parseopt
|
||||||
|
|
||||||
import constantine/math/arithmetic
|
# import constantine/math/arithmetic
|
||||||
|
#
|
||||||
import poseidon2/types
|
# import poseidon2/types
|
||||||
import poseidon2/merkle
|
# import poseidon2/merkle
|
||||||
import poseidon2/io
|
# import poseidon2/io
|
||||||
|
|
||||||
import types
|
import types
|
||||||
import blocks
|
import types/bn254
|
||||||
import slot
|
import types/goldilocks
|
||||||
import dataset
|
#import blocks/bn254
|
||||||
import sample
|
#import blocks/goldilocks
|
||||||
import merkle
|
#import slot
|
||||||
import gen_input
|
#import dataset
|
||||||
import json
|
#import sample
|
||||||
|
#import sample/bn254
|
||||||
|
#import sample/goldilocks
|
||||||
|
#import merkle
|
||||||
|
#import merkle/bn254
|
||||||
|
#import merkle/goldilocks
|
||||||
|
import gen_input/bn254
|
||||||
|
import gen_input/goldilocks
|
||||||
|
import json/bn254
|
||||||
|
import json/goldilocks
|
||||||
import misc
|
import misc
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
@ -38,6 +47,7 @@ type FullConfig = object
|
|||||||
const defHashCfg =
|
const defHashCfg =
|
||||||
HashConfig( field: Goldilocks # BN254
|
HashConfig( field: Goldilocks # BN254
|
||||||
, hashFun: Poseidon2
|
, hashFun: Poseidon2
|
||||||
|
, combo: Goldilocks_Poseidon2
|
||||||
)
|
)
|
||||||
|
|
||||||
const defGlobCfg =
|
const defGlobCfg =
|
||||||
@ -143,6 +153,8 @@ proc parseCliOptions(): FullConfig =
|
|||||||
of cmdEnd:
|
of cmdEnd:
|
||||||
discard
|
discard
|
||||||
|
|
||||||
|
hashCfg.combo = toFieldHashCombo( hashCfg.field , hashCfg.hashFun )
|
||||||
|
|
||||||
fullCfg.hashCfg = hashCfg
|
fullCfg.hashCfg = hashCfg
|
||||||
fullCfg.globCfg = globCfg
|
fullCfg.globCfg = globCfg
|
||||||
fullCfg.dsetCfg = dsetCfg
|
fullCfg.dsetCfg = dsetCfg
|
||||||
@ -196,7 +208,7 @@ proc writeCircomMainComponent(fullCfg: FullConfig, fname: string) =
|
|||||||
when isMainModule:
|
when isMainModule:
|
||||||
|
|
||||||
let fullCfg = parseCliOptions()
|
let fullCfg = parseCliOptions()
|
||||||
# echo fullCfg
|
let hashCfg = fullCfg.hashCfg
|
||||||
|
|
||||||
if fullCfg.verbose:
|
if fullCfg.verbose:
|
||||||
printConfig(fullCfg)
|
printConfig(fullCfg)
|
||||||
@ -212,7 +224,14 @@ when isMainModule:
|
|||||||
|
|
||||||
if fullCfg.outFile != "":
|
if fullCfg.outFile != "":
|
||||||
echo "writing proof input into `" & fullCfg.outFile & "`..."
|
echo "writing proof input into `" & fullCfg.outFile & "`..."
|
||||||
let prfInput = generateProofInput( fullCfg.globCfg, fullCfg.dsetCfg, fullCfg.slotIndex, toF(fullCfg.entropy) )
|
case hashCfg.field
|
||||||
exportProofInput( fullCfg.outFile , prfInput )
|
of BN254:
|
||||||
|
let entropy = intToBN254(fullCfg.entropy)
|
||||||
|
let prfInput = generateProofInputBN254( hashCfg, fullCfg.globCfg, fullCfg.dsetCfg, fullCfg.slotIndex, entropy )
|
||||||
|
exportProofInputBN254( hashCfg, fullCfg.outFile , prfInput )
|
||||||
|
of Goldilocks:
|
||||||
|
let entropy = intToDigest(fullCfg.entropy)
|
||||||
|
let prfInput = generateProofInputGoldilocks( hashCfg, fullCfg.globCfg, fullCfg.dsetCfg, fullCfg.slotIndex, entropy )
|
||||||
|
exportProofInputGoldilocks( hashCfg, fullCfg.outFile , prfInput )
|
||||||
|
|
||||||
echo "done"
|
echo "done"
|
||||||
|
|||||||
@ -1,73 +0,0 @@
|
|||||||
|
|
||||||
#
|
|
||||||
# generate the input data for the proof
|
|
||||||
# see `json.nim` to export it in Snarkjs-compatible format
|
|
||||||
#
|
|
||||||
|
|
||||||
import sugar
|
|
||||||
import std/sequtils
|
|
||||||
|
|
||||||
import blocks
|
|
||||||
import slot
|
|
||||||
import dataset
|
|
||||||
import sample
|
|
||||||
import merkle
|
|
||||||
import types
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
proc buildSlotTreeFull( globcfg: GlobalConfig, slotCfg: SlotConfig ): (seq[MerkleTree], MerkleTree) =
|
|
||||||
let ncells = slotCfg.nCells
|
|
||||||
let nblocks = ncells div cellsPerBlock(globcfg)
|
|
||||||
assert( nblocks * cellsPerBlock(globcfg) == ncells )
|
|
||||||
let blocks : seq[Block] = collect( newSeq, (for i in 0..<nblocks: slotLoadBlockData(globcfg, slotCfg, i) ))
|
|
||||||
let miniTrees : seq[MerkleTree] = collect( newSeq, (for blk in blocks: networkBlockTree(globcfg, blk) ))
|
|
||||||
let blockHashes : seq[Root] = map( miniTrees , treeRoot )
|
|
||||||
let bigTree = merkleTree( blockHashes )
|
|
||||||
return (miniTrees, bigTree)
|
|
||||||
|
|
||||||
proc buildSlotTree( globcfg: GlobalConfig, slotCfg: SlotConfig ): MerkleTree =
|
|
||||||
return buildSlotTreeFull(globcfg, slotCfg)[1]
|
|
||||||
|
|
||||||
proc generateProofInput*( globCfg: GlobalConfig, dsetCfg: DataSetConfig, slotIdx: SlotIdx, entropy: Entropy ): SlotProofInput =
|
|
||||||
let nslots = dsetCfg.nSlots
|
|
||||||
let ncells = dsetCfg.nCells
|
|
||||||
let nblocks = ncells div cellsPerBlock(globCfg)
|
|
||||||
assert( nblocks * cellsPerBlock(globcfg) == ncells )
|
|
||||||
|
|
||||||
let slotCfgs = collect( newSeq , (for i in 0..<nslots: slotCfgFromDataSetCfg(dsetcfg, i) ))
|
|
||||||
let slotTrees = collect( newSeq , (for scfg in slotcfgs: buildSlotTree(globCfg, scfg) ))
|
|
||||||
let slotRoots = map( slotTrees, treeRoot )
|
|
||||||
|
|
||||||
let ourSlotCfg = slotCfgs[slotIdx]
|
|
||||||
let ourSlotRoot = slotRoots[slotIdx]
|
|
||||||
let ourSlotTree = slotTrees[slotIdx]
|
|
||||||
|
|
||||||
let dsetTree = merkleTree( slotRoots )
|
|
||||||
let dsetRoot = treeRoot( dsetTree )
|
|
||||||
let slotProof = merkleProof( dsetTree , slotIdx )
|
|
||||||
|
|
||||||
let indices = cellIndices(entropy, ourSlotRoot, ncells, dsetCfg.nSamples)
|
|
||||||
|
|
||||||
var inputs : seq[CellProofInput]
|
|
||||||
for cellIdx in indices:
|
|
||||||
let (miniTrees, bigTree) = buildSlotTreeFull( globCfg, ourSlotCfg )
|
|
||||||
let blockIdx = cellIdx div cellsPerBlock(globcfg)
|
|
||||||
let blockTree = miniTrees[ blockIdx ]
|
|
||||||
let cellData = slotLoadCellData( globCfg, ourSlotCfg, cellIdx )
|
|
||||||
let botProof = merkleProof( blockTree , cellIdx mod cellsPerBlock(globcfg) )
|
|
||||||
let topProof = merkleProof( bigTree , blockIdx )
|
|
||||||
let prf = padMerkleProof( mergeMerkleProofs( botProof, topProof ), globCfg.maxDepth )
|
|
||||||
inputs.add( CellProofInput(cellData: cellData, merkleProof: prf) )
|
|
||||||
|
|
||||||
return SlotProofInput( dataSetRoot: dsetRoot
|
|
||||||
, entropy: entropy
|
|
||||||
, nCells: ncells
|
|
||||||
, nSlots: nslots
|
|
||||||
, slotIndex: slotIdx
|
|
||||||
, slotRoot: ourSlotRoot
|
|
||||||
, slotProof: padMerkleProof( slotProof, globCfg.maxLog2NSlots )
|
|
||||||
, proofInputs: inputs
|
|
||||||
)
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
81
reference/nim/proof_input/src/gen_input/bn254.nim
Normal file
81
reference/nim/proof_input/src/gen_input/bn254.nim
Normal file
@ -0,0 +1,81 @@
|
|||||||
|
|
||||||
|
#
|
||||||
|
# generate the input data for the proof
|
||||||
|
# see `json.nim` to export it in Snarkjs-compatible format
|
||||||
|
#
|
||||||
|
|
||||||
|
import sugar
|
||||||
|
import std/sequtils
|
||||||
|
|
||||||
|
import ../blocks/bn254
|
||||||
|
import ../slot
|
||||||
|
import ../dataset
|
||||||
|
import ../sample/bn254
|
||||||
|
import ../merkle
|
||||||
|
import ../merkle/bn254
|
||||||
|
import ../types
|
||||||
|
import ../types/bn254
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc buildSlotTreeFull( hashcfg: HashConfig, globcfg: GlobalConfig, slotCfg: SlotConfig ): (seq[MerkleTree[Hash]], MerkleTree[Hash]) =
|
||||||
|
|
||||||
|
let ncells = slotCfg.nCells
|
||||||
|
let nblocks = ncells div cellsPerBlock(globcfg)
|
||||||
|
assert( nblocks * cellsPerBlock(globcfg) == ncells )
|
||||||
|
let blocks : seq[Block] = collect( newSeq, (for i in 0..<nblocks: slotLoadBlockData(globcfg, slotCfg, i) ))
|
||||||
|
let miniTrees : seq[MerkleTree[Hash]] = collect( newSeq, (for blk in blocks: networkBlockTree(hashcfg, globcfg, blk) ))
|
||||||
|
let blockHashes : seq[Root] = map( miniTrees , treeRoot )
|
||||||
|
let bigTree = merkleTree( hashcfg, blockHashes )
|
||||||
|
return (miniTrees, bigTree)
|
||||||
|
|
||||||
|
proc buildSlotTree( hashcfg: HashConfig, globcfg: GlobalConfig, slotCfg: SlotConfig ): MerkleTree[Hash] =
|
||||||
|
return buildSlotTreeFull(hashcfg, globcfg, slotCfg)[1]
|
||||||
|
|
||||||
|
proc generateProofInput*( hashCfg: HashConfig, globCfg: GlobalConfig, dsetCfg: DataSetConfig, slotIdx: SlotIdx, entropy: Entropy ): SlotProofInput[Hash] =
|
||||||
|
let nslots = dsetCfg.nSlots
|
||||||
|
let ncells = dsetCfg.nCells
|
||||||
|
let nblocks = ncells div cellsPerBlock(globCfg)
|
||||||
|
assert( nblocks * cellsPerBlock(globcfg) == ncells )
|
||||||
|
|
||||||
|
let slotCfgs = collect( newSeq , (for i in 0..<nslots: slotCfgFromDataSetCfg(dsetcfg, i) ))
|
||||||
|
let slotTrees = collect( newSeq , (for scfg in slotcfgs: buildSlotTree(hashCfg, globCfg, scfg) ))
|
||||||
|
let slotRoots = map( slotTrees, treeRoot )
|
||||||
|
|
||||||
|
let ourSlotCfg = slotCfgs[slotIdx]
|
||||||
|
let ourSlotRoot = slotRoots[slotIdx]
|
||||||
|
let ourSlotTree = slotTrees[slotIdx]
|
||||||
|
|
||||||
|
let dsetTree : MerkleTree[Hash] = merkleTree( hashcfg, slotRoots )
|
||||||
|
let dsetRoot : Hash = treeRoot( dsetTree )
|
||||||
|
let slotProof : MerkleProof[Hash] = merkleProof( dsetTree , slotIdx )
|
||||||
|
|
||||||
|
let indices = cellIndices(hashCfg, entropy, ourSlotRoot, ncells, dsetCfg.nSamples)
|
||||||
|
|
||||||
|
var inputs : seq[CellProofInput[Hash]]
|
||||||
|
for cellIdx in indices:
|
||||||
|
let (miniTrees, bigTree) = buildSlotTreeFull( hashCfg, globCfg, ourSlotCfg )
|
||||||
|
let blockIdx = cellIdx div cellsPerBlock(globcfg)
|
||||||
|
let blockTree = miniTrees[ blockIdx ]
|
||||||
|
let cellData = slotLoadCellData( globCfg, ourSlotCfg, cellIdx )
|
||||||
|
let botProof = merkleProof( blockTree , cellIdx mod cellsPerBlock(globcfg) )
|
||||||
|
let topProof = merkleProof( bigTree , blockIdx )
|
||||||
|
let prf = padMerkleProof( mergeMerkleProofs( compressWithKey, botProof, topProof ), globCfg.maxDepth )
|
||||||
|
inputs.add( CellProofInput[Hash](cellData: cellData, merkleProof: prf) )
|
||||||
|
|
||||||
|
return SlotProofInput[Hash]( dataSetRoot: dsetRoot
|
||||||
|
, entropy: entropy
|
||||||
|
, nCells: ncells
|
||||||
|
, nSlots: nslots
|
||||||
|
, slotIndex: slotIdx
|
||||||
|
, slotRoot: ourSlotRoot
|
||||||
|
, slotProof: padMerkleProof( slotProof, globCfg.maxLog2NSlots )
|
||||||
|
, proofInputs: inputs
|
||||||
|
)
|
||||||
|
|
||||||
|
#---------------------------------------
|
||||||
|
|
||||||
|
proc generateProofInputBN254*( hashCfg: HashConfig, globCfg: GlobalConfig, dsetCfg: DataSetConfig, slotIdx: SlotIdx, entropy: Entropy ): SlotProofInput[Hash] =
|
||||||
|
generateProofInput( hashCfg, globCfg, dsetCfg, slotIdx, entropy)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
87
reference/nim/proof_input/src/gen_input/goldilocks.nim
Normal file
87
reference/nim/proof_input/src/gen_input/goldilocks.nim
Normal file
@ -0,0 +1,87 @@
|
|||||||
|
|
||||||
|
#
|
||||||
|
# generate the input data for the proof
|
||||||
|
# see `json.nim` to export it in Snarkjs-compatible format
|
||||||
|
#
|
||||||
|
|
||||||
|
import sugar
|
||||||
|
import std/sequtils
|
||||||
|
|
||||||
|
import ../blocks/goldilocks
|
||||||
|
import ../slot
|
||||||
|
import ../dataset
|
||||||
|
import ../sample/goldilocks
|
||||||
|
import ../merkle
|
||||||
|
import ../merkle/goldilocks/poseidon2
|
||||||
|
import ../merkle/goldilocks/monolith
|
||||||
|
import ../types
|
||||||
|
import ../types/goldilocks
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc buildSlotTreeFull( hashcfg: HashConfig, globcfg: GlobalConfig, slotCfg: SlotConfig ): (seq[MerkleTree[Digest]], MerkleTree[Digest]) =
|
||||||
|
|
||||||
|
let ncells = slotCfg.nCells
|
||||||
|
let nblocks = ncells div cellsPerBlock(globcfg)
|
||||||
|
assert( nblocks * cellsPerBlock(globcfg) == ncells )
|
||||||
|
let blocks : seq[Block] = collect( newSeq, (for i in 0..<nblocks: slotLoadBlockData(globcfg, slotCfg, i) ))
|
||||||
|
let miniTrees : seq[MerkleTree[Digest]] = collect( newSeq, (for blk in blocks: networkBlockTree(hashcfg, globcfg, blk) ))
|
||||||
|
let blockHashes : seq[Root] = map( miniTrees , treeRoot )
|
||||||
|
let bigTree = merkleTree( hashcfg, blockHashes )
|
||||||
|
return (miniTrees, bigTree)
|
||||||
|
|
||||||
|
proc buildSlotTree( hashcfg: HashConfig, globcfg: GlobalConfig, slotCfg: SlotConfig ): MerkleTree[Digest] =
|
||||||
|
return buildSlotTreeFull(hashcfg, globcfg, slotCfg)[1]
|
||||||
|
|
||||||
|
proc generateProofInput*( hashCfg: HashConfig, globCfg: GlobalConfig, dsetCfg: DataSetConfig, slotIdx: SlotIdx, entropy: Entropy ): SlotProofInput[Digest] =
|
||||||
|
let nslots = dsetCfg.nSlots
|
||||||
|
let ncells = dsetCfg.nCells
|
||||||
|
let nblocks = ncells div cellsPerBlock(globCfg)
|
||||||
|
assert( nblocks * cellsPerBlock(globcfg) == ncells )
|
||||||
|
|
||||||
|
let slotCfgs = collect( newSeq , (for i in 0..<nslots: slotCfgFromDataSetCfg(dsetcfg, i) ))
|
||||||
|
let slotTrees = collect( newSeq , (for scfg in slotcfgs: buildSlotTree(hashCfg, globCfg, scfg) ))
|
||||||
|
let slotRoots = map( slotTrees, treeRoot )
|
||||||
|
|
||||||
|
let ourSlotCfg = slotCfgs[slotIdx]
|
||||||
|
let ourSlotRoot = slotRoots[slotIdx]
|
||||||
|
let ourSlotTree = slotTrees[slotIdx]
|
||||||
|
|
||||||
|
let dsetTree = merkleTree( hashcfg, slotRoots )
|
||||||
|
let dsetRoot = treeRoot( dsetTree )
|
||||||
|
let slotProof = merkleProof( dsetTree , slotIdx )
|
||||||
|
|
||||||
|
let indices = cellIndices(hashCfg, entropy, ourSlotRoot, ncells, dsetCfg.nSamples)
|
||||||
|
|
||||||
|
var inputs : seq[CellProofInput[Digest]]
|
||||||
|
for cellIdx in indices:
|
||||||
|
let (miniTrees, bigTree) = buildSlotTreeFull( hashCfg, globCfg, ourSlotCfg )
|
||||||
|
let blockIdx = cellIdx div cellsPerBlock(globcfg)
|
||||||
|
let blockTree = miniTrees[ blockIdx ]
|
||||||
|
let cellData = slotLoadCellData( globCfg, ourSlotCfg, cellIdx )
|
||||||
|
let botProof = merkleProof( blockTree , cellIdx mod cellsPerBlock(globcfg) )
|
||||||
|
let topProof = merkleProof( bigTree , blockIdx )
|
||||||
|
|
||||||
|
var prf : MerkleProof[Digest]
|
||||||
|
case hashCfg.hashFun:
|
||||||
|
of Poseidon2: prf = padMerkleProof( mergeMerkleProofs( poseidon2.compressWithKey, botProof, topProof ), globCfg.maxDepth )
|
||||||
|
of Monolith: prf = padMerkleProof( mergeMerkleProofs( monolith.compressWithKey , botProof, topProof ), globCfg.maxDepth )
|
||||||
|
|
||||||
|
inputs.add( CellProofInput[Digest](cellData: cellData, merkleProof: prf) )
|
||||||
|
|
||||||
|
return SlotProofInput[Digest]( dataSetRoot: dsetRoot
|
||||||
|
, entropy: entropy
|
||||||
|
, nCells: ncells
|
||||||
|
, nSlots: nslots
|
||||||
|
, slotIndex: slotIdx
|
||||||
|
, slotRoot: ourSlotRoot
|
||||||
|
, slotProof: padMerkleProof( slotProof, globCfg.maxLog2NSlots )
|
||||||
|
, proofInputs: inputs
|
||||||
|
)
|
||||||
|
|
||||||
|
#---------------------------------------
|
||||||
|
|
||||||
|
proc generateProofInputGoldilocks*( hashCfg: HashConfig, globCfg: GlobalConfig, dsetCfg: DataSetConfig, slotIdx: SlotIdx, entropy: Entropy ): SlotProofInput[Digest] =
|
||||||
|
generateProofInput( hashCfg, globCfg, dsetCfg, slotIdx, entropy)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
@ -1,112 +1,17 @@
|
|||||||
|
|
||||||
#
|
|
||||||
# export the proof inputs as a JSON file suitable for `snarkjs`
|
|
||||||
#
|
|
||||||
|
|
||||||
import sugar
|
|
||||||
import std/strutils
|
|
||||||
import std/sequtils
|
|
||||||
import std/streams
|
|
||||||
|
|
||||||
from poseidon2/io import elements
|
|
||||||
|
|
||||||
import types
|
import types
|
||||||
|
from types/bn254 import Hash
|
||||||
|
from types/goldilocks import Digest
|
||||||
|
|
||||||
|
import json/bn254
|
||||||
|
import json/goldilocks
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
func toQuotedDecimalF(x: F): string =
|
proc exportProofInput*(hashcfg: HashConfig, fname: string, prfInput: SlotProofInput[Digest]) =
|
||||||
let s : string = toDecimalF(x)
|
exportProofInputGoldilocks(hashcfg, fname, prfInput)
|
||||||
return ("\"" & s & "\"")
|
|
||||||
|
|
||||||
func mkIndent(foo: string): string =
|
proc exportProofInput*(hashcfg: HashConfig, fname: string, prfInput: SlotProofInput[Hash]) =
|
||||||
return spaces(foo.len)
|
exportProofInputBN254(hashcfg, fname, prfInput)
|
||||||
|
|
||||||
proc writeF(h: Stream, prefix: string, x: F) =
|
|
||||||
h.writeLine(prefix & toQuotedDecimalF(x))
|
|
||||||
|
|
||||||
#[
|
|
||||||
proc writeSeq(h: Stream, prefix: string, xs: seq[F])
|
|
||||||
let n = xs.len
|
|
||||||
let indent = mkIndent(prefix)
|
|
||||||
for i in 0..<n:
|
|
||||||
let str : string = toQuotedF( xs[i] )
|
|
||||||
if i==0:
|
|
||||||
h.writeLine(prefix & "[ " & str)
|
|
||||||
else:
|
|
||||||
h.writeLine(indent & ", " & str)
|
|
||||||
h.writeLine(indent & "] ")
|
|
||||||
]#
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
type
|
|
||||||
WriteFun[T] = proc (stream: Stream, prefix: string, what: T) {.closure.}
|
|
||||||
|
|
||||||
proc writeList[T](h: Stream, prefix: string, xs: seq[T], writeFun: WriteFun[T]) =
|
|
||||||
let n = xs.len
|
|
||||||
let indent = mkIndent(prefix)
|
|
||||||
for i in 0..<n:
|
|
||||||
if i==0:
|
|
||||||
writeFun(h, prefix & "[ ", xs[i])
|
|
||||||
else:
|
|
||||||
writeFun(h, indent & ", ", xs[i])
|
|
||||||
h.writeLine( indent & "]" )
|
|
||||||
|
|
||||||
proc writeFieldElems(h: Stream, prefix: string, xs: seq[F]) =
|
|
||||||
writeList[F]( h, prefix, xs, writeF )
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
proc writeSingleCellData(h: Stream, prefix:string , cell: Cell) =
|
|
||||||
let flds : seq[F] = cell.elements(F).toSeq()
|
|
||||||
writeFieldElems(h, prefix, flds)
|
|
||||||
|
|
||||||
proc writeAllCellData(h: Stream, cells: seq[Cell]) =
|
|
||||||
writeList(h, " ", cells, writeSingleCellData )
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
proc writeSingleMerklePath(h: Stream, prefix: string, path: MerkleProof) =
|
|
||||||
let flds = path.merklePath
|
|
||||||
writeFieldElems(h, prefix, flds)
|
|
||||||
|
|
||||||
proc writeAllMerklePaths(h: Stream, cells: seq[MerkleProof]) =
|
|
||||||
writeList(h, " ", cells, writeSingleMerklePath )
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
#[
|
|
||||||
|
|
||||||
signal input entropy; // public input
|
|
||||||
signal input dataSetRoot; // public input
|
|
||||||
signal input slotIndex; // must be public, otherwise we could prove a different slot
|
|
||||||
|
|
||||||
signal input slotRoot; // can be private input
|
|
||||||
signal input nCellsPerSlot; // can be private input (Merkle tree is safe)
|
|
||||||
signal input nSlotsPerDataSet; // can be private input (Merkle tree is safe)
|
|
||||||
|
|
||||||
signal input slotProof[maxLog2NSlots]; // path from the slot root the the dataset root (private input)
|
|
||||||
|
|
||||||
signal input cellData[nSamples][nFieldElemsPerCell]; // private input
|
|
||||||
signal input merklePaths[nSamples][maxDepth]; // private input
|
|
||||||
]#
|
|
||||||
|
|
||||||
proc exportProofInput*(fname: string, prfInput: SlotProofInput) =
|
|
||||||
let h = openFileStream(fname, fmWrite)
|
|
||||||
defer: h.close()
|
|
||||||
|
|
||||||
h.writeLine("{")
|
|
||||||
h.writeLine(" \"dataSetRoot\": " & toQuotedDecimalF(prfInput.dataSetRoot) )
|
|
||||||
h.writeLine(", \"entropy\": " & toQuotedDecimalF(prfInput.entropy ) )
|
|
||||||
h.writeLine(", \"nCellsPerSlot\": " & $(prfInput.nCells) )
|
|
||||||
h.writeLine(", \"nSlotsPerDataSet\": " & $(prfInput.nSlots) )
|
|
||||||
h.writeLine(", \"slotIndex\": " & $(prfInput.slotIndex) )
|
|
||||||
h.writeLine(", \"slotRoot\": " & toQuotedDecimalF(prfInput.slotRoot) )
|
|
||||||
h.writeLine(", \"slotProof\":")
|
|
||||||
writeSingleMerklePath(h, " ", prfInput.slotProof )
|
|
||||||
h.writeLine(", \"cellData\":")
|
|
||||||
writeAllCellData(h, collect( newSeq , (for p in prfInput.proofInputs: p.cellData) ))
|
|
||||||
h.writeLine(", \"merklePaths\":")
|
|
||||||
writeAllMerklePaths(h, collect( newSeq , (for p in prfInput.proofInputs: p.merkleProof) ))
|
|
||||||
h.writeLine("}")
|
|
||||||
|
|
||||||
|
|||||||
81
reference/nim/proof_input/src/json/bn254.nim
Normal file
81
reference/nim/proof_input/src/json/bn254.nim
Normal file
@ -0,0 +1,81 @@
|
|||||||
|
|
||||||
|
#
|
||||||
|
# export the proof inputs as a JSON file suitable for `snarkjs`
|
||||||
|
#
|
||||||
|
|
||||||
|
import sugar
|
||||||
|
#import std/strutils
|
||||||
|
import std/sequtils
|
||||||
|
import std/streams
|
||||||
|
|
||||||
|
from poseidon2/io import elements
|
||||||
|
|
||||||
|
import ../types
|
||||||
|
import ../types/bn254
|
||||||
|
import shared
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc writeFieldElems(h: Stream, prefix: string, xs: seq[F]) =
|
||||||
|
writeList[F]( h, prefix, xs, writeF )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc writeSingleCellData(h: Stream, prefix:string , cell: Cell) =
|
||||||
|
let flds : seq[F] = cell.elements(F).toSeq()
|
||||||
|
writeFieldElems(h, prefix, flds)
|
||||||
|
|
||||||
|
proc writeAllCellData(h: Stream, cells: seq[Cell]) =
|
||||||
|
writeList(h, " ", cells, writeSingleCellData )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc writeSingleMerklePath(h: Stream, prefix: string, path: MerkleProof[Hash]) =
|
||||||
|
let flds = path.merklePath
|
||||||
|
writeFieldElems(h, prefix, flds)
|
||||||
|
|
||||||
|
proc writeAllMerklePaths(h: Stream, paths: seq[MerkleProof[Hash]]) =
|
||||||
|
writeList(h, " ", paths, writeSingleMerklePath )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
#[
|
||||||
|
signal input entropy; // public input
|
||||||
|
signal input dataSetRoot; // public input
|
||||||
|
signal input slotIndex; // must be public, otherwise we could prove a different slot
|
||||||
|
|
||||||
|
signal input slotRoot; // can be private input
|
||||||
|
signal input nCellsPerSlot; // can be private input (Merkle tree is safe)
|
||||||
|
signal input nSlotsPerDataSet; // can be private input (Merkle tree is safe)
|
||||||
|
|
||||||
|
signal input slotProof[maxLog2NSlots]; // path from the slot root the the dataset root (private input)
|
||||||
|
|
||||||
|
signal input cellData[nSamples][nFieldElemsPerCell]; // private input
|
||||||
|
signal input merklePaths[nSamples][maxDepth]; // private input
|
||||||
|
]#
|
||||||
|
|
||||||
|
proc exportProofInput*(fname: string, prfInput: SlotProofInput[Hash]) =
|
||||||
|
let h = openFileStream(fname, fmWrite)
|
||||||
|
defer: h.close()
|
||||||
|
|
||||||
|
h.writeLine("{")
|
||||||
|
h.writeLine(" \"dataSetRoot\": " & toQuotedDecimalF(prfInput.dataSetRoot) )
|
||||||
|
h.writeLine(", \"entropy\": " & toQuotedDecimalF(prfInput.entropy ) )
|
||||||
|
h.writeLine(", \"nCellsPerSlot\": " & $(prfInput.nCells) )
|
||||||
|
h.writeLine(", \"nSlotsPerDataSet\": " & $(prfInput.nSlots) )
|
||||||
|
h.writeLine(", \"slotIndex\": " & $(prfInput.slotIndex) )
|
||||||
|
h.writeLine(", \"slotRoot\": " & toQuotedDecimalF(prfInput.slotRoot) )
|
||||||
|
h.writeLine(", \"slotProof\":")
|
||||||
|
writeSingleMerklePath(h, " ", prfInput.slotProof )
|
||||||
|
h.writeLine(", \"cellData\":")
|
||||||
|
writeAllCellData(h, collect( newSeq , (for p in prfInput.proofInputs: p.cellData) ))
|
||||||
|
h.writeLine(", \"merklePaths\":")
|
||||||
|
writeAllMerklePaths(h, collect( newSeq , (for p in prfInput.proofInputs: p.merkleProof) ))
|
||||||
|
h.writeLine("}")
|
||||||
|
|
||||||
|
|
||||||
|
proc exportProofInputBN254*(hashcfg: HashConfig, fname: string, prfInput: SlotProofInput[Hash]) =
|
||||||
|
assert( hashcfg.field == BN254 )
|
||||||
|
exportProofInput(fname, prfInput)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
86
reference/nim/proof_input/src/json/goldilocks.nim
Normal file
86
reference/nim/proof_input/src/json/goldilocks.nim
Normal file
@ -0,0 +1,86 @@
|
|||||||
|
|
||||||
|
#
|
||||||
|
# export the proof inputs as a JSON file suitable for `snarkjs`
|
||||||
|
#
|
||||||
|
|
||||||
|
import sugar
|
||||||
|
#import std/strutils
|
||||||
|
#import std/sequtils
|
||||||
|
import std/streams
|
||||||
|
|
||||||
|
import goldilocks_hash/marshal
|
||||||
|
|
||||||
|
import ../types
|
||||||
|
import ../types/goldilocks
|
||||||
|
import shared
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func bytesToFieldElements( bytes: openArray[byte] ): seq[F] =
|
||||||
|
let digests = padAndDecodeBytesToDigest62(bytes)
|
||||||
|
return digestSeqToFeltSeq(digests)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc writeFieldElems(h: Stream, prefix: string, xs: seq[F]) =
|
||||||
|
writeList[F]( h, prefix, xs, writeF )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc writeSingleCellData(h: Stream, prefix:string , cell: Cell) =
|
||||||
|
let flds : seq[F] = bytesToFieldElements(cell) # cell.elements(F).toSeq()
|
||||||
|
writeFieldElems(h, prefix, flds)
|
||||||
|
|
||||||
|
proc writeAllCellData(h: Stream, cells: seq[Cell]) =
|
||||||
|
writeList(h, " ", cells, writeSingleCellData )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
proc writeSingleMerklePath(h: Stream, prefix: string, path: MerkleProof[Digest]) =
|
||||||
|
let flds : seq[F] = digestSeqToFeltSeq( path.merklePath )
|
||||||
|
writeFieldElems(h, prefix, flds)
|
||||||
|
|
||||||
|
proc writeAllMerklePaths(h: Stream, paths: seq[MerkleProof[Digest]]) =
|
||||||
|
writeList(h, " ", paths, writeSingleMerklePath )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
#[
|
||||||
|
signal input entropy; // public input
|
||||||
|
signal input dataSetRoot; // public input
|
||||||
|
signal input slotIndex; // must be public, otherwise we could prove a different slot
|
||||||
|
|
||||||
|
signal input slotRoot; // can be private input
|
||||||
|
signal input nCellsPerSlot; // can be private input (Merkle tree is safe)
|
||||||
|
signal input nSlotsPerDataSet; // can be private input (Merkle tree is safe)
|
||||||
|
|
||||||
|
signal input slotProof[maxLog2NSlots]; // path from the slot root the the dataset root (private input)
|
||||||
|
|
||||||
|
signal input cellData[nSamples][nFieldElemsPerCell]; // private input
|
||||||
|
signal input merklePaths[nSamples][maxDepth]; // private input
|
||||||
|
]#
|
||||||
|
|
||||||
|
proc exportProofInput*(fname: string, prfInput: SlotProofInput[Digest]) =
|
||||||
|
let h = openFileStream(fname, fmWrite)
|
||||||
|
defer: h.close()
|
||||||
|
|
||||||
|
h.writeLine("{")
|
||||||
|
h.writeLine(" \"dataSetRoot\": " & digestToJsonString(prfInput.dataSetRoot) )
|
||||||
|
h.writeLine(", \"entropy\": " & digestToJsonString(prfInput.entropy ) )
|
||||||
|
h.writeLine(", \"nCellsPerSlot\": " & $(prfInput.nCells) )
|
||||||
|
h.writeLine(", \"nSlotsPerDataSet\": " & $(prfInput.nSlots) )
|
||||||
|
h.writeLine(", \"slotIndex\": " & $(prfInput.slotIndex) )
|
||||||
|
h.writeLine(", \"slotRoot\": " & digestToJsonString(prfInput.slotRoot) )
|
||||||
|
h.writeLine(", \"slotProof\":")
|
||||||
|
writeSingleMerklePath(h, " ", prfInput.slotProof )
|
||||||
|
h.writeLine(", \"cellData\":")
|
||||||
|
writeAllCellData(h, collect( newSeq , (for p in prfInput.proofInputs: p.cellData) ))
|
||||||
|
h.writeLine(", \"merklePaths\":")
|
||||||
|
writeAllMerklePaths(h, collect( newSeq , (for p in prfInput.proofInputs: p.merkleProof) ))
|
||||||
|
h.writeLine("}")
|
||||||
|
|
||||||
|
proc exportProofInputGoldilocks*(hashcfg: HashConfig, fname: string, prfInput: SlotProofInput[Digest]) =
|
||||||
|
assert( hashcfg.field == Goldilocks )
|
||||||
|
exportProofInput(fname, prfInput)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
27
reference/nim/proof_input/src/json/shared.nim
Normal file
27
reference/nim/proof_input/src/json/shared.nim
Normal file
@ -0,0 +1,27 @@
|
|||||||
|
|
||||||
|
#import sugar
|
||||||
|
#import std/sequtils
|
||||||
|
import std/strutils
|
||||||
|
import std/streams
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func mkIndent*(foo: string): string =
|
||||||
|
return spaces(foo.len)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
type
|
||||||
|
WriteFun*[T] = proc (stream: Stream, prefix: string, what: T) {.closure.}
|
||||||
|
|
||||||
|
proc writeList*[T](h: Stream, prefix: string, xs: seq[T], writeFun: WriteFun[T]) =
|
||||||
|
let n = xs.len
|
||||||
|
let indent = mkIndent(prefix)
|
||||||
|
for i in 0..<n:
|
||||||
|
if i==0:
|
||||||
|
writeFun(h, prefix & "[ ", xs[i])
|
||||||
|
else:
|
||||||
|
writeFun(h, indent & ", ", xs[i])
|
||||||
|
h.writeLine( indent & "]" )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
@ -1,39 +1,32 @@
|
|||||||
|
|
||||||
import std/bitops
|
import std/bitops
|
||||||
import std/sequtils
|
|
||||||
|
|
||||||
import constantine/math/arithmetic
|
|
||||||
import constantine/math/io/io_fields
|
|
||||||
|
|
||||||
import poseidon2/types
|
|
||||||
import poseidon2/compress
|
|
||||||
import poseidon2/io
|
|
||||||
|
|
||||||
import types
|
import types
|
||||||
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
func treeDepth*(tree: MerkleTree): int =
|
func treeDepth*[H](tree: MerkleTree[H]): int =
|
||||||
return tree.layers.len - 1
|
return tree.layers.len - 1
|
||||||
|
|
||||||
func treeNumberOfLeaves*(tree: MerkleTree): int =
|
func treeNumberOfLeaves*[H](tree: MerkleTree[H]): int =
|
||||||
return tree.layers[0].len
|
return tree.layers[0].len
|
||||||
|
|
||||||
func treeRoot*(tree: MerkleTree): Hash =
|
func treeRoot*[H](tree: MerkleTree[H]): H =
|
||||||
let last = tree.layers[tree.layers.len-1]
|
let last = tree.layers[tree.layers.len-1]
|
||||||
assert( last.len == 1 )
|
assert( last.len == 1 )
|
||||||
return last[0]
|
return last[0]
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
func merkleProof*(tree: MerkleTree, index: int): MerkleProof =
|
func merkleProof*[H](tree: MerkleTree[H], index: int): MerkleProof[H] =
|
||||||
let depth = treeDepth(tree)
|
let depth = treeDepth(tree)
|
||||||
let nleaves = treeNumberOfLeaves(tree)
|
let nleaves = treeNumberOfLeaves(tree)
|
||||||
|
|
||||||
|
var zero : H # hackety hack, it should be initialized with zeros
|
||||||
|
|
||||||
assert( index >= 0 and index < nleaves )
|
assert( index >= 0 and index < nleaves )
|
||||||
|
|
||||||
var path : seq[Hash] = newSeq[Hash](depth)
|
var path : seq[H] = newSeq[H](depth)
|
||||||
var k = index
|
var k = index
|
||||||
var m = nleaves
|
var m = nleaves
|
||||||
for i in 0..<depth:
|
for i in 0..<depth:
|
||||||
@ -42,21 +35,23 @@ func merkleProof*(tree: MerkleTree, index: int): MerkleProof =
|
|||||||
k = k shr 1
|
k = k shr 1
|
||||||
m = (m+1) shr 1
|
m = (m+1) shr 1
|
||||||
|
|
||||||
return MerkleProof( leafIndex: index
|
return MerkleProof[H]( leafIndex: index
|
||||||
, leafValue: tree.layers[0][index]
|
, leafValue: tree.layers[0][index]
|
||||||
, merklePath: path
|
, merklePath: path
|
||||||
, numberOfLeaves: nleaves
|
, numberOfLeaves: nleaves
|
||||||
)
|
)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
func compressWithKey(key: int, x: F, y: F): F =
|
type CompressWithKey[H] = proc (key: int, x, y: H): H {.closure.}
|
||||||
compress(x,y, key=toF(key))
|
|
||||||
|
|
||||||
func reconstructRoot*(proof: MerkleProof): Hash =
|
# func compressWithKey(key: int, x: F, y: F): F =
|
||||||
var m : int = proof.numberOfLeaves
|
# compress(x,y, key=toF(key))
|
||||||
var j : int = proof.leafIndex
|
|
||||||
var h : Hash = proof.leafValue
|
func reconstructRoot*[H](compressWithKey: CompressWithKey[H], proof: MerkleProof[H]): H =
|
||||||
|
var m : int = proof.numberOfLeaves
|
||||||
|
var j : int = proof.leafIndex
|
||||||
|
var h : H = proof.leafValue
|
||||||
var bottomFlag : int = 1
|
var bottomFlag : int = 1
|
||||||
for p in proof.merklePath:
|
for p in proof.merklePath:
|
||||||
let oddIndex : bool = (bitand(j,1) != 0)
|
let oddIndex : bool = (bitand(j,1) != 0)
|
||||||
@ -78,50 +73,8 @@ func reconstructRoot*(proof: MerkleProof): Hash =
|
|||||||
m = (m+1) shr 1
|
m = (m+1) shr 1
|
||||||
return h
|
return h
|
||||||
|
|
||||||
func checkMerkleProof*(root: Root, proof: MerkleProof): bool =
|
func checkMerkleProof*[H](compressWithKey: CompressWithKey[H], root: H, proof: MerkleProof[H]): bool =
|
||||||
return bool(root == reconstructRoot(proof))
|
return bool( root == reconstructRoot[H](compressWithKey, proof) )
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
# TODO: maybe move this (and the rest?) into poseidon2-nim
|
|
||||||
|
|
||||||
const KeyNone = F.fromHex("0x0")
|
|
||||||
const KeyBottomLayer = F.fromHex("0x1")
|
|
||||||
const KeyOdd = F.fromHex("0x2")
|
|
||||||
const KeyOddAndBottomLayer = F.fromhex("0x3")
|
|
||||||
|
|
||||||
func merkleTreeWorker(xs: openArray[F], isBottomLayer: static bool) : seq[seq[F]] =
|
|
||||||
let a = low(xs)
|
|
||||||
let b = high(xs)
|
|
||||||
let m = b-a+1
|
|
||||||
|
|
||||||
when not isBottomLayer:
|
|
||||||
if m==1:
|
|
||||||
return @[ xs.toSeq() ]
|
|
||||||
|
|
||||||
let halfn : int = m div 2
|
|
||||||
let n : int = 2*halfn
|
|
||||||
let isOdd : bool = (n != m)
|
|
||||||
|
|
||||||
var ys : seq[F]
|
|
||||||
if not isOdd:
|
|
||||||
ys = newSeq[F](halfn)
|
|
||||||
else:
|
|
||||||
ys = newSeq[F](halfn+1)
|
|
||||||
|
|
||||||
for i in 0..<halfn:
|
|
||||||
const key = when isBottomLayer: KeyBottomLayer else: KeyNone
|
|
||||||
ys[i] = compress( xs[a+2*i], xs[a+2*i+1], key = key )
|
|
||||||
if isOdd:
|
|
||||||
const key = when isBottomLayer: KeyOddAndBottomLayer else: KeyOdd
|
|
||||||
ys[halfn] = compress( xs[n], zero, key = key )
|
|
||||||
|
|
||||||
var ls : seq[seq[F]]
|
|
||||||
ls = @[ xs.toSeq() ]
|
|
||||||
ls = ls & merkleTreeWorker(ys, isBottomLayer = false)
|
|
||||||
return ls
|
|
||||||
|
|
||||||
func merkleTree*(xs: openArray[F]) : MerkleTree =
|
|
||||||
return MerkleTree(layers: merkleTreeWorker(xs, isBottomLayer = true))
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
@ -130,9 +83,9 @@ func merkleTree*(xs: openArray[F]) : MerkleTree =
|
|||||||
# we can compose proofs (though for checking these proofs we need to remember
|
# we can compose proofs (though for checking these proofs we need to remember
|
||||||
# this and use a specific custom convention, because we mark the bottom layers)
|
# this and use a specific custom convention, because we mark the bottom layers)
|
||||||
#
|
#
|
||||||
func mergeMerkleProofs*(bottomProof, topProof: MerkleProof): MerkleProof =
|
func mergeMerkleProofs*[H](compressWithKey: CompressWithKey[H], bottomProof, topProof: MerkleProof[H]): MerkleProof[H] =
|
||||||
|
|
||||||
let botRoot = reconstructRoot( bottomProof )
|
let botRoot = reconstructRoot[H]( compressWithKey, bottomProof )
|
||||||
assert( bool(botRoot == topProof.leafValue) )
|
assert( bool(botRoot == topProof.leafValue) )
|
||||||
|
|
||||||
let idx = topProof.leafIndex * bottomProof.numberOfLeaves + bottomProof.leafIndex
|
let idx = topProof.leafIndex * bottomProof.numberOfLeaves + bottomProof.leafIndex
|
||||||
@ -140,10 +93,10 @@ func mergeMerkleProofs*(bottomProof, topProof: MerkleProof): MerkleProof =
|
|||||||
let nlvs = bottomProof.numberOfLeaves * topProof.numberOfLeaves
|
let nlvs = bottomProof.numberOfLeaves * topProof.numberOfLeaves
|
||||||
let path = bottomProof.merklePath & topProof.merklePath
|
let path = bottomProof.merklePath & topProof.merklePath
|
||||||
|
|
||||||
return MerkleProof( leafIndex: idx
|
return MerkleProof[H]( leafIndex: idx
|
||||||
, leafValue: val
|
, leafValue: val
|
||||||
, merklePath: path
|
, merklePath: path
|
||||||
, numberOfLeaves: nlvs
|
, numberOfLeaves: nlvs
|
||||||
)
|
)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|||||||
65
reference/nim/proof_input/src/merkle/bn254.nim
Normal file
65
reference/nim/proof_input/src/merkle/bn254.nim
Normal file
@ -0,0 +1,65 @@
|
|||||||
|
|
||||||
|
import std/sequtils
|
||||||
|
|
||||||
|
import
|
||||||
|
constantine/math/arithmetic,
|
||||||
|
constantine/math/io/io_fields
|
||||||
|
|
||||||
|
import poseidon2/types
|
||||||
|
import poseidon2/compress
|
||||||
|
import poseidon2/merkle
|
||||||
|
import poseidon2/io
|
||||||
|
|
||||||
|
import ../types
|
||||||
|
import ../types/bn254
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func compressWithkey*(key: int, x, y: F): F = compress(x,y, key=toF(key))
|
||||||
|
|
||||||
|
func merkleDigestBN254*(xs: openArray[F]): F = Merkle.digest(xs)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
const KeyNone = F.fromHex("0x0")
|
||||||
|
const KeyBottomLayer = F.fromHex("0x1")
|
||||||
|
const KeyOdd = F.fromHex("0x2")
|
||||||
|
const KeyOddAndBottomLayer = F.fromhex("0x3")
|
||||||
|
|
||||||
|
func merkleTreeWorker(xs: openArray[F], isBottomLayer: static bool) : seq[seq[F]] =
|
||||||
|
let a = low(xs)
|
||||||
|
let b = high(xs)
|
||||||
|
let m = b-a+1
|
||||||
|
|
||||||
|
when not isBottomLayer:
|
||||||
|
if m==1:
|
||||||
|
return @[ xs.toSeq() ]
|
||||||
|
|
||||||
|
let halfn : int = m div 2
|
||||||
|
let n : int = 2*halfn
|
||||||
|
let isOdd : bool = (n != m)
|
||||||
|
|
||||||
|
var ys : seq[F]
|
||||||
|
if not isOdd:
|
||||||
|
ys = newSeq[F](halfn)
|
||||||
|
else:
|
||||||
|
ys = newSeq[F](halfn+1)
|
||||||
|
|
||||||
|
for i in 0..<halfn:
|
||||||
|
const key = when isBottomLayer: KeyBottomLayer else: KeyNone
|
||||||
|
ys[i] = compress( xs[a+2*i], xs[a+2*i+1], key = key )
|
||||||
|
if isOdd:
|
||||||
|
const key = when isBottomLayer: KeyOddAndBottomLayer else: KeyOdd
|
||||||
|
ys[halfn] = compress( xs[n], zero, key = key )
|
||||||
|
|
||||||
|
var ls : seq[seq[F]]
|
||||||
|
ls = @[ xs.toSeq() ]
|
||||||
|
ls = ls & merkleTreeWorker(ys, isBottomLayer = false)
|
||||||
|
return ls
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func merkleTreeBN254*(xs: openArray[F]) : MerkleTree[F] =
|
||||||
|
return MerkleTree[F](layers: merkleTreeWorker(xs, isBottomLayer = true))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
65
reference/nim/proof_input/src/merkle/goldilocks/monolith.nim
Normal file
65
reference/nim/proof_input/src/merkle/goldilocks/monolith.nim
Normal file
@ -0,0 +1,65 @@
|
|||||||
|
|
||||||
|
import std/sequtils
|
||||||
|
|
||||||
|
import goldilocks_hash/types
|
||||||
|
import goldilocks_hash/monolith/compress
|
||||||
|
import goldilocks_hash/monolith/merkle
|
||||||
|
import goldilocks_hash/monolith/sponge
|
||||||
|
|
||||||
|
import ../../types
|
||||||
|
import ../../types/goldilocks
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func compressWithkey*(key: int, x, y: Digest): Digest = compress(x,y, key=uint64(key))
|
||||||
|
|
||||||
|
func merkleDigestMonolith*(xs: openArray[Digest]): Digest = Merkle.digest(xs)
|
||||||
|
|
||||||
|
# TODO: move these somewhere else
|
||||||
|
func digestFeltsMonolith*(xs: openArray[F] ): Digest = digestFeltsC( rate=8, xs )
|
||||||
|
func digestBytesMonolith*(xs: openArray[byte]): Digest = digestBytesC( rate=8, xs )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
const KeyNone : uint64 = 0x00
|
||||||
|
const KeyBottomLayer : uint64 = 0x01
|
||||||
|
const KeyOdd : uint64 = 0x02
|
||||||
|
const KeyOddAndBottomLayer : uint64 = 0x03
|
||||||
|
|
||||||
|
func merkleTreeWorker(xs: openArray[Digest], isBottomLayer: static bool) : seq[seq[Digest]] =
|
||||||
|
let a = low(xs)
|
||||||
|
let b = high(xs)
|
||||||
|
let m = b-a+1
|
||||||
|
|
||||||
|
when not isBottomLayer:
|
||||||
|
if m==1:
|
||||||
|
return @[ xs.toSeq() ]
|
||||||
|
|
||||||
|
let halfn : int = m div 2
|
||||||
|
let n : int = 2*halfn
|
||||||
|
let isOdd : bool = (n != m)
|
||||||
|
|
||||||
|
var ys : seq[Digest]
|
||||||
|
if not isOdd:
|
||||||
|
ys = newSeq[Digest](halfn)
|
||||||
|
else:
|
||||||
|
ys = newSeq[Digest](halfn+1)
|
||||||
|
|
||||||
|
for i in 0..<halfn:
|
||||||
|
const key = when isBottomLayer: KeyBottomLayer else: KeyNone
|
||||||
|
ys[i] = compress( xs[a+2*i], xs[a+2*i+1], key = key )
|
||||||
|
if isOdd:
|
||||||
|
const key = when isBottomLayer: KeyOddAndBottomLayer else: KeyOdd
|
||||||
|
ys[halfn] = compress( xs[n], zeroDigest, key = key )
|
||||||
|
|
||||||
|
var ls : seq[seq[Digest]]
|
||||||
|
ls = @[ xs.toSeq() ]
|
||||||
|
ls = ls & merkleTreeWorker(ys, isBottomLayer = false)
|
||||||
|
return ls
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func merkleTreeGoldilocksMonolith*(xs: openArray[Digest]) : MerkleTree[Digest] =
|
||||||
|
return MerkleTree[Digest](layers: merkleTreeWorker(xs, isBottomLayer = true))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
@ -0,0 +1,65 @@
|
|||||||
|
|
||||||
|
import std/sequtils
|
||||||
|
|
||||||
|
import goldilocks_hash/types
|
||||||
|
import goldilocks_hash/poseidon2/compress
|
||||||
|
import goldilocks_hash/poseidon2/merkle
|
||||||
|
import goldilocks_hash/poseidon2/sponge
|
||||||
|
|
||||||
|
import ../../types
|
||||||
|
import ../../types/goldilocks
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func compressWithkey*(key: int, x, y: Digest): Digest = compress(x,y, key=uint64(key))
|
||||||
|
|
||||||
|
func merkleDigestPoseidon2*(xs: openArray[Digest]): Digest = Merkle.digest(xs)
|
||||||
|
|
||||||
|
# TODO: move these somewhere else
|
||||||
|
func digestFeltsPoseidon2*(xs: openArray[F] ): Digest = digestFeltsC( rate=8, xs )
|
||||||
|
func digestBytesPoseidon2*(xs: openArray[byte]): Digest = digestBytesC( rate=8, xs )
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
const KeyNone : uint64 = 0x00
|
||||||
|
const KeyBottomLayer : uint64 = 0x01
|
||||||
|
const KeyOdd : uint64 = 0x02
|
||||||
|
const KeyOddAndBottomLayer : uint64 = 0x03
|
||||||
|
|
||||||
|
func merkleTreeWorker(xs: openArray[Digest], isBottomLayer: static bool) : seq[seq[Digest]] =
|
||||||
|
let a = low(xs)
|
||||||
|
let b = high(xs)
|
||||||
|
let m = b-a+1
|
||||||
|
|
||||||
|
when not isBottomLayer:
|
||||||
|
if m==1:
|
||||||
|
return @[ xs.toSeq() ]
|
||||||
|
|
||||||
|
let halfn : int = m div 2
|
||||||
|
let n : int = 2*halfn
|
||||||
|
let isOdd : bool = (n != m)
|
||||||
|
|
||||||
|
var ys : seq[Digest]
|
||||||
|
if not isOdd:
|
||||||
|
ys = newSeq[Digest](halfn)
|
||||||
|
else:
|
||||||
|
ys = newSeq[Digest](halfn+1)
|
||||||
|
|
||||||
|
for i in 0..<halfn:
|
||||||
|
const key = when isBottomLayer: KeyBottomLayer else: KeyNone
|
||||||
|
ys[i] = compress( xs[a+2*i], xs[a+2*i+1], key = key )
|
||||||
|
if isOdd:
|
||||||
|
const key = when isBottomLayer: KeyOddAndBottomLayer else: KeyOdd
|
||||||
|
ys[halfn] = compress( xs[n], zeroDigest, key = key )
|
||||||
|
|
||||||
|
var ls : seq[seq[Digest]]
|
||||||
|
ls = @[ xs.toSeq() ]
|
||||||
|
ls = ls & merkleTreeWorker(ys, isBottomLayer = false)
|
||||||
|
return ls
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func merkleTreeGoldilocksPoseidon2*(xs: openArray[Digest]) : MerkleTree[Digest] =
|
||||||
|
return MerkleTree[Digest](layers: merkleTreeWorker(xs, isBottomLayer = true))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
@ -1,43 +0,0 @@
|
|||||||
|
|
||||||
import sugar
|
|
||||||
import std/bitops
|
|
||||||
|
|
||||||
import constantine/math/arithmetic
|
|
||||||
|
|
||||||
import poseidon2/types
|
|
||||||
import poseidon2/sponge
|
|
||||||
import poseidon2/io
|
|
||||||
|
|
||||||
import types
|
|
||||||
import misc
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
func extractLowBits[n: static int]( A: BigInt[n], k: int): uint64 =
|
|
||||||
assert( k>0 and k<=64 )
|
|
||||||
var r : uint64 = 0
|
|
||||||
for i in 0..<k:
|
|
||||||
let b = bit[n](A, i) # NOTE: the docunmentation seems to lie about the conventions here....
|
|
||||||
let y = uint64(b)
|
|
||||||
if (y != 0):
|
|
||||||
r = bitor( r, 1'u64 shl i )
|
|
||||||
return r
|
|
||||||
|
|
||||||
func extractLowBits(fld: F, k: int): uint64 =
|
|
||||||
let A : BigInt[254] = fld.toBig()
|
|
||||||
return extractLowBits(A, k);
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
func cellIndex*(entropy: Entropy, slotRoot: Root, numberOfCells: int, counter: int): int =
|
|
||||||
let log2 = ceilingLog2(numberOfCells)
|
|
||||||
assert( 1 shl log2 == numberOfCells , "for this version, `numberOfCells` is assumed to be a power of two")
|
|
||||||
|
|
||||||
let input : seq[F] = @[ entropy, slotRoot, toF(counter) ]
|
|
||||||
let H : Hash = Sponge.digest( input, rate = 2 )
|
|
||||||
return int(extractLowBits(H,log2))
|
|
||||||
|
|
||||||
func cellIndices*(entropy: Entropy, slotRoot: Root, numberOfCells: int, nSamples: int): seq[int] =
|
|
||||||
return collect( newSeq, (for i in 1..nSamples: cellIndex(entropy, slotRoot, numberOfCells, i) ))
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
29
reference/nim/proof_input/src/sample/bn254.nim
Normal file
29
reference/nim/proof_input/src/sample/bn254.nim
Normal file
@ -0,0 +1,29 @@
|
|||||||
|
|
||||||
|
import sugar
|
||||||
|
#import std/bitops
|
||||||
|
|
||||||
|
# import constantine/math/arithmetic
|
||||||
|
import poseidon2/types
|
||||||
|
import poseidon2/sponge
|
||||||
|
import poseidon2/io
|
||||||
|
|
||||||
|
import ../types
|
||||||
|
import ../types/bn254
|
||||||
|
import ../misc
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func cellIndex*(hashcfg: HashConfig, entropy: Entropy, slotRoot: Root, numberOfCells: int, counter: int): int =
|
||||||
|
assert( hashcfg.field == BN254 )
|
||||||
|
|
||||||
|
let log2 = ceilingLog2(numberOfCells)
|
||||||
|
assert( 1 shl log2 == numberOfCells , "for this version, `numberOfCells` is assumed to be a power of two")
|
||||||
|
|
||||||
|
let input : seq[F] = @[ entropy, slotRoot, toF(counter) ]
|
||||||
|
let H : Hash = Sponge.digest( input, rate = 2 )
|
||||||
|
return int(extractLowBits(H,log2))
|
||||||
|
|
||||||
|
func cellIndices*(hashcfg: HashConfig, entropy: Entropy, slotRoot: Root, numberOfCells: int, nSamples: int): seq[int] =
|
||||||
|
return collect( newSeq, (for i in 1..nSamples: cellIndex(hashcfg, entropy, slotRoot, numberOfCells, i) ))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
40
reference/nim/proof_input/src/sample/goldilocks.nim
Normal file
40
reference/nim/proof_input/src/sample/goldilocks.nim
Normal file
@ -0,0 +1,40 @@
|
|||||||
|
|
||||||
|
import sugar
|
||||||
|
#import std/bitops
|
||||||
|
#import std/sequtils
|
||||||
|
|
||||||
|
import goldilocks_hash/types
|
||||||
|
#import goldilocks_hash/poseidon2/sponge
|
||||||
|
#import goldilocks_hash/monolith/sponge
|
||||||
|
|
||||||
|
import ../types
|
||||||
|
import ../types/goldilocks
|
||||||
|
import ../merkle/goldilocks/poseidon2
|
||||||
|
import ../merkle/goldilocks/monolith
|
||||||
|
import ../misc
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func cellIndex*(hashcfg: HashConfig, entropy: Entropy, slotRoot: Digest, numberOfCells: int, counter: int): int =
|
||||||
|
assert( hashcfg.field == Goldilocks )
|
||||||
|
|
||||||
|
let log2 = ceilingLog2(numberOfCells)
|
||||||
|
assert( 1 shl log2 == numberOfCells , "for this version, `numberOfCells` is assumed to be a power of two")
|
||||||
|
|
||||||
|
let inputD : seq[Digest] = @[ entropy, slotRoot, intToDigest( counter ) ]
|
||||||
|
let input : seq[F] = digestSeqToFeltSeq(inputD)
|
||||||
|
|
||||||
|
var hash : Digest
|
||||||
|
case hashcfg.hashFun:
|
||||||
|
of Poseidon2: hash = digestFeltsPoseidon2( input )
|
||||||
|
of Monolith: hash = digestFeltsMonolith( input )
|
||||||
|
|
||||||
|
let hash4 : F4 = fromDigest(hash)
|
||||||
|
let hash0 : F = hash4[0]
|
||||||
|
|
||||||
|
return int(extractLowBits(hash0,log2))
|
||||||
|
|
||||||
|
func cellIndices*(hashcfg: HashConfig, entropy: Entropy, slotRoot: Digest, numberOfCells: int, nSamples: int): seq[int] =
|
||||||
|
return collect( newSeq, (for i in 1..nSamples: cellIndex(hashcfg, entropy, slotRoot, numberOfCells, i) ))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
@ -32,6 +32,8 @@ proc genFakeCell(globcfg: GlobalConfig, cfg: SlotConfig, seed: Seed, idx: CellId
|
|||||||
return cell
|
return cell
|
||||||
|
|
||||||
#[
|
#[
|
||||||
|
-- original Haskell version:
|
||||||
|
--
|
||||||
genFakeCell :: SlotConfig -> Seed -> CellIdx -> CellData
|
genFakeCell :: SlotConfig -> Seed -> CellIdx -> CellData
|
||||||
genFakeCell cfg (Seed seed) (CellIdx idx) = (mkCellData cfg $ B.pack list) where
|
genFakeCell cfg (Seed seed) (CellIdx idx) = (mkCellData cfg $ B.pack list) where
|
||||||
list = go (fromIntegral $ _cellSize cfg) 1
|
list = go (fromIntegral $ _cellSize cfg) 1
|
||||||
|
|||||||
@ -2,25 +2,6 @@
|
|||||||
import std/strutils
|
import std/strutils
|
||||||
import std/sequtils
|
import std/sequtils
|
||||||
|
|
||||||
from constantine/math/io/io_fields import toDecimal
|
|
||||||
|
|
||||||
import poseidon2/types
|
|
||||||
export types
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
type Entropy* = F
|
|
||||||
type Hash* = F
|
|
||||||
type Root* = Hash
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
func toDecimalF*(a : F): string =
|
|
||||||
var s : string = toDecimal(a)
|
|
||||||
s = s.strip( leading=true, trailing=false, chars={'0'} )
|
|
||||||
if s.len == 0: s="0"
|
|
||||||
return s
|
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
type Cell* = seq[byte]
|
type Cell* = seq[byte]
|
||||||
@ -30,27 +11,30 @@ type Block* = seq[byte]
|
|||||||
|
|
||||||
type
|
type
|
||||||
|
|
||||||
MerkleProof* = object
|
MerkleProof*[H] = object
|
||||||
leafIndex* : int # linear index of the leaf, starting from 0
|
leafIndex* : int # linear index of the leaf, starting from 0
|
||||||
leafValue* : Hash # value of the leaf
|
leafValue* : H # value of the leaf
|
||||||
merklePath* : seq[Hash] # order: from the bottom to the top
|
merklePath* : seq[H] # order: from the bottom to the top
|
||||||
numberOfLeaves* : int # number of leaves in the tree (=size of input)
|
numberOfLeaves* : int # number of leaves in the tree (=size of input)
|
||||||
|
|
||||||
MerkleTree* = object
|
MerkleTree*[H] = object
|
||||||
layers*: seq[seq[Hash]]
|
layers*: seq[seq[H]]
|
||||||
# ^^^ note: the first layer is the bottom layer, and the last layer is the root
|
# ^^^ note: the first layer is the bottom layer, and the last layer is the root
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
# the circuit expect merkle path of statically known length, so we need to pad them
|
# the circuit expect merkle path of statically known length, so we need to pad them
|
||||||
func padMerkleProof*( old: MerkleProof, newlen: int ): MerkleProof =
|
func padMerkleProof*[H]( old: MerkleProof[H], newlen: int ): MerkleProof[H] =
|
||||||
let pad = newlen - old.merklePath.len
|
let pad = newlen - old.merklePath.len
|
||||||
assert( pad >= 0 )
|
assert( pad >= 0 )
|
||||||
return MerkleProof( leafIndex: old.leafIndex
|
|
||||||
, leafValue: old.leafValue
|
var zero : H # hackety hack hack, it should be initialized to zero
|
||||||
, merklePath: old.merklePath & repeat(zero,pad)
|
|
||||||
, numberOfLeaves: old.numberOfLeaves
|
return MerkleProof[H]( leafIndex: old.leafIndex
|
||||||
)
|
, leafValue: old.leafValue
|
||||||
|
, merklePath: old.merklePath & repeat(zero,pad)
|
||||||
|
, numberOfLeaves: old.numberOfLeaves
|
||||||
|
)
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
@ -61,19 +45,19 @@ type
|
|||||||
BlockIdx* = int
|
BlockIdx* = int
|
||||||
SlotIdx* = int
|
SlotIdx* = int
|
||||||
|
|
||||||
CellProofInput* = object
|
CellProofInput*[H] = object
|
||||||
cellData*: Cell
|
cellData*: Cell
|
||||||
merkleProof*: MerkleProof
|
merkleProof*: MerkleProof[H]
|
||||||
|
|
||||||
SlotProofInput* = object
|
SlotProofInput*[H] = object
|
||||||
dataSetRoot*: Root
|
dataSetRoot*: H # Root
|
||||||
entropy*: Entropy
|
entropy*: H # Entropy
|
||||||
nSlots*: int
|
nSlots*: int
|
||||||
nCells*: int
|
nCells*: int
|
||||||
slotRoot*: Root
|
slotRoot*: H # Root
|
||||||
slotIndex*: SlotIdx
|
slotIndex*: SlotIdx
|
||||||
slotProof*: MerkleProof
|
slotProof*: MerkleProof[H]
|
||||||
proofInputs*: seq[CellProofInput]
|
proofInputs*: seq[CellProofInput[H]]
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
@ -109,6 +93,7 @@ type
|
|||||||
HashConfig* = object
|
HashConfig* = object
|
||||||
field* : FieldSelect
|
field* : FieldSelect
|
||||||
hashFun* : HashSelect
|
hashFun* : HashSelect
|
||||||
|
combo* : FieldHashCombo
|
||||||
|
|
||||||
FieldSelect* = enum
|
FieldSelect* = enum
|
||||||
BN254,
|
BN254,
|
||||||
@ -118,6 +103,11 @@ type
|
|||||||
Poseidon2,
|
Poseidon2,
|
||||||
Monolith
|
Monolith
|
||||||
|
|
||||||
|
FieldHashCombo* = enum
|
||||||
|
BN254_Poseidon2,
|
||||||
|
Goldilocks_Poseidon2,
|
||||||
|
Goldilocks_Monolith
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
func cellsPerBlock*(glob: GlobalConfig): int =
|
func cellsPerBlock*(glob: GlobalConfig): int =
|
||||||
@ -142,3 +132,19 @@ func parseHashFun*(str0: string): HashSelect =
|
|||||||
else: raiseAssert("parsefield: unrecognized hash function `" & str0 & "`")
|
else: raiseAssert("parsefield: unrecognized hash function `" & str0 & "`")
|
||||||
|
|
||||||
#-------------------------------------------------------------------------------
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
{. warning[UnreachableElse]:off .}
|
||||||
|
func toFieldHashCombo*( field: FieldSelect, hash: HashSelect ): FieldHashCombo =
|
||||||
|
let msg = "invalid hash function `" & ($hash) & "` choice for field `" & ($field) & "`"
|
||||||
|
case field:
|
||||||
|
of BN254:
|
||||||
|
case hash:
|
||||||
|
of Poseidon2: return BN254_Poseidon2
|
||||||
|
else: raiseAssert(msg)
|
||||||
|
of Goldilocks:
|
||||||
|
case hash:
|
||||||
|
of Poseidon2: return Goldilocks_Poseidon2
|
||||||
|
of Monolith: return Goldilocks_Monolith
|
||||||
|
else: raiseAssert(msg)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|||||||
59
reference/nim/proof_input/src/types/bn254.nim
Normal file
59
reference/nim/proof_input/src/types/bn254.nim
Normal file
@ -0,0 +1,59 @@
|
|||||||
|
|
||||||
|
import std/strutils
|
||||||
|
import std/bitops
|
||||||
|
import std/streams
|
||||||
|
|
||||||
|
import
|
||||||
|
constantine/math/arithmetic,
|
||||||
|
constantine/math/io/io_fields,
|
||||||
|
constantine/math/io/io_bigints,
|
||||||
|
constantine/math/config/curves
|
||||||
|
|
||||||
|
#from constantine/math/io/io_fields import toDecimal
|
||||||
|
|
||||||
|
import poseidon2/types
|
||||||
|
import poseidon2/io
|
||||||
|
export types
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
type BN254_T* = F
|
||||||
|
type Entropy* = F
|
||||||
|
type Hash* = F
|
||||||
|
type Root* = Hash
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func intToBN254*(x: int): F = toF(x)
|
||||||
|
|
||||||
|
func toDecimalF*(a : F): string =
|
||||||
|
var s : string = toDecimal(a)
|
||||||
|
s = s.strip( leading=true, trailing=false, chars={'0'} )
|
||||||
|
if s.len == 0: s="0"
|
||||||
|
return s
|
||||||
|
|
||||||
|
func toQuotedDecimalF*(x: F): string =
|
||||||
|
let s : string = toDecimalF(x)
|
||||||
|
return ("\"" & s & "\"")
|
||||||
|
|
||||||
|
proc writeF*(h: Stream, prefix: string, x: F) =
|
||||||
|
h.writeLine(prefix & toQuotedDecimalF(x))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func extractLowBits[n: static int]( A: BigInt[n], k: int): uint64 =
|
||||||
|
assert( k>0 and k<=64 )
|
||||||
|
var r : uint64 = 0
|
||||||
|
for i in 0..<k:
|
||||||
|
let b = bit[n](A, i) # NOTE: the docunmentation seems to lie about the conventions here....
|
||||||
|
let y = uint64(b)
|
||||||
|
if (y != 0):
|
||||||
|
r = bitor( r, 1'u64 shl i )
|
||||||
|
return r
|
||||||
|
|
||||||
|
func extractLowBits*(fld: F, k: int): uint64 =
|
||||||
|
let A : BigInt[254] = fld.toBig()
|
||||||
|
return extractLowBits(A, k);
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
44
reference/nim/proof_input/src/types/goldilocks.nim
Normal file
44
reference/nim/proof_input/src/types/goldilocks.nim
Normal file
@ -0,0 +1,44 @@
|
|||||||
|
|
||||||
|
#import std/strutils
|
||||||
|
import std/bitops
|
||||||
|
import std/streams
|
||||||
|
|
||||||
|
import goldilocks_hash/types
|
||||||
|
export types
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
type Goldi_T* = F
|
||||||
|
type Entropy* = Digest
|
||||||
|
type Hash* = Digest
|
||||||
|
type Root* = Hash
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func toDecimalF*(a : F): string =
|
||||||
|
var s : string = $fromF(a)
|
||||||
|
return s
|
||||||
|
|
||||||
|
func toQuotedDecimalF*(x: F): string =
|
||||||
|
let s : string = toDecimalF(x)
|
||||||
|
return ("\"" & s & "\"")
|
||||||
|
|
||||||
|
proc writeF*(h: Stream, prefix: string, x: F) =
|
||||||
|
h.writeLine(prefix & toQuotedDecimalF(x))
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func extractLowBits*(fld: F, k: int): uint64 =
|
||||||
|
assert( k>0 and k<=56 )
|
||||||
|
let val : uint64 = fromF(fld)
|
||||||
|
let mask : uint64 = (1'u64 shl k) - 1
|
||||||
|
return bitand(val, mask)
|
||||||
|
|
||||||
|
#-------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
func digestToJsonString*( d: Digest ): string =
|
||||||
|
let xs: F4 = fromDigest(d)
|
||||||
|
return "[ " & toDecimalF(xs[0]) & ", " &
|
||||||
|
toDecimalF(xs[1]) & ", " &
|
||||||
|
toDecimalF(xs[2]) & ", " &
|
||||||
|
toDecimalF(xs[3]) & " ]"
|
||||||
Loading…
x
Reference in New Issue
Block a user