use std::collections::{HashMap, HashSet}; use std::marker::PhantomData; use itertools::Itertools; use plonky2::field::extension::{Extendable, FieldExtension}; use plonky2::field::packed::PackedField; use plonky2::field::polynomial::PolynomialValues; use plonky2::field::types::Field; use plonky2::hash::hash_types::RichField; use plonky2::timed; use plonky2::util::timing::TimingTree; use rand::Rng; use crate::constraint_consumer::{ConstraintConsumer, RecursiveConstraintConsumer}; use crate::cross_table_lookup::Column; use crate::lookup::{eval_lookups, eval_lookups_circuit, permuted_cols}; use crate::memory::columns::{ is_channel, value_limb, ADDR_CONTEXT, ADDR_SEGMENT, ADDR_VIRTUAL, CONTEXT_FIRST_CHANGE, COUNTER, COUNTER_PERMUTED, IS_READ, NUM_COLUMNS, RANGE_CHECK, RANGE_CHECK_PERMUTED, SEGMENT_FIRST_CHANGE, TIMESTAMP, VIRTUAL_FIRST_CHANGE, }; use crate::memory::{NUM_CHANNELS, VALUE_LIMBS}; use crate::permutation::PermutationPair; use crate::stark::Stark; use crate::util::trace_rows_to_poly_values; use crate::vars::{StarkEvaluationTargets, StarkEvaluationVars}; pub(crate) const NUM_PUBLIC_INPUTS: usize = 0; pub fn ctl_data() -> Vec> { let mut res = Column::singles([IS_READ, ADDR_CONTEXT, ADDR_SEGMENT, ADDR_VIRTUAL]).collect_vec(); res.extend(Column::singles((0..8).map(value_limb))); res.push(Column::single(TIMESTAMP)); res } pub fn ctl_filter(channel: usize) -> Column { Column::single(is_channel(channel)) } #[derive(Copy, Clone, Default)] pub struct MemoryStark { pub(crate) f: PhantomData, } #[derive(Clone, Debug)] pub struct MemoryOp { /// The channel this operation came from, or `None` if it's a dummy operation for padding. pub channel_index: Option, pub timestamp: usize, pub is_read: bool, pub context: usize, pub segment: usize, pub virt: usize, pub value: [F; 8], } pub fn generate_random_memory_ops( num_ops: usize, rng: &mut R, ) -> Vec> { let mut memory_ops = Vec::new(); let mut current_memory_values: HashMap<(usize, usize, usize), [F; 8]> = HashMap::new(); let num_cycles = num_ops / 2; for clock in 0..num_cycles { let mut used_indices = HashSet::new(); let mut new_writes_this_cycle = HashMap::new(); let mut has_read = false; for _ in 0..2 { let mut channel_index = rng.gen_range(0..NUM_CHANNELS); while used_indices.contains(&channel_index) { channel_index = rng.gen_range(0..NUM_CHANNELS); } used_indices.insert(channel_index); let is_read = if clock == 0 { false } else { !has_read && rng.gen() }; has_read = has_read || is_read; let (context, segment, virt, vals) = if is_read { let written: Vec<_> = current_memory_values.keys().collect(); let &(context, segment, virt) = written[rng.gen_range(0..written.len())]; let &vals = current_memory_values .get(&(context, segment, virt)) .unwrap(); (context, segment, virt, vals) } else { // TODO: with taller memory table or more padding (to enable range-checking bigger diffs), // test larger address values. let mut context = rng.gen_range(0..40); let mut segment = rng.gen_range(0..8); let mut virt = rng.gen_range(0..20); while new_writes_this_cycle.contains_key(&(context, segment, virt)) { context = rng.gen_range(0..40); segment = rng.gen_range(0..8); virt = rng.gen_range(0..20); } let val: [u32; 8] = rng.gen(); let vals: [F; 8] = val.map(F::from_canonical_u32); new_writes_this_cycle.insert((context, segment, virt), vals); (context, segment, virt, vals) }; let timestamp = clock * NUM_CHANNELS + channel_index; memory_ops.push(MemoryOp { channel_index: Some(channel_index), timestamp, is_read, context, segment, virt, value: vals, }); } for (k, v) in new_writes_this_cycle { current_memory_values.insert(k, v); } } memory_ops } fn get_max_range_check(memory_ops: &[MemoryOp]) -> usize { memory_ops .iter() .tuple_windows() .map(|(curr, next)| { if curr.context != next.context { next.context - curr.context - 1 } else if curr.segment != next.segment { next.segment - curr.segment - 1 } else if curr.virt != next.virt { next.virt - curr.virt - 1 } else { next.timestamp - curr.timestamp - 1 } }) .max() .unwrap_or(0) } pub fn generate_first_change_flags( context: &[F], segment: &[F], virtuals: &[F], ) -> (Vec, Vec, Vec) { let num_ops = context.len(); let mut context_first_change = Vec::with_capacity(num_ops); let mut segment_first_change = Vec::with_capacity(num_ops); let mut virtual_first_change = Vec::with_capacity(num_ops); for idx in 0..num_ops - 1 { let this_context_first_change = context[idx] != context[idx + 1]; let this_segment_first_change = segment[idx] != segment[idx + 1] && !this_context_first_change; let this_virtual_first_change = virtuals[idx] != virtuals[idx + 1] && !this_segment_first_change && !this_context_first_change; context_first_change.push(F::from_bool(this_context_first_change)); segment_first_change.push(F::from_bool(this_segment_first_change)); virtual_first_change.push(F::from_bool(this_virtual_first_change)); } context_first_change.push(F::ZERO); segment_first_change.push(F::ZERO); virtual_first_change.push(F::ZERO); ( context_first_change, segment_first_change, virtual_first_change, ) } pub fn generate_range_check_value( context: &[F], segment: &[F], virtuals: &[F], timestamp: &[F], context_first_change: &[F], segment_first_change: &[F], virtual_first_change: &[F], ) -> Vec { let num_ops = context.len(); let mut range_check = Vec::new(); for idx in 0..num_ops - 1 { let this_address_unchanged = F::ONE - context_first_change[idx] - segment_first_change[idx] - virtual_first_change[idx]; range_check.push( context_first_change[idx] * (context[idx + 1] - context[idx] - F::ONE) + segment_first_change[idx] * (segment[idx + 1] - segment[idx] - F::ONE) + virtual_first_change[idx] * (virtuals[idx + 1] - virtuals[idx] - F::ONE) + this_address_unchanged * (timestamp[idx + 1] - timestamp[idx] - F::ONE), ); } range_check.push(F::ZERO); range_check } impl, const D: usize> MemoryStark { pub(crate) fn generate_trace_rows( &self, mut memory_ops: Vec>, ) -> Vec<[F; NUM_COLUMNS]> { memory_ops.sort_by_key(|op| (op.context, op.segment, op.virt, op.timestamp)); Self::pad_memory_ops(&mut memory_ops); let num_ops = memory_ops.len(); let mut trace_cols = [(); NUM_COLUMNS].map(|_| vec![F::ZERO; num_ops]); for i in 0..num_ops { let MemoryOp { channel_index, timestamp, is_read, context, segment, virt, value, } = memory_ops[i]; if let Some(channel) = channel_index { trace_cols[is_channel(channel)][i] = F::ONE; } trace_cols[TIMESTAMP][i] = F::from_canonical_usize(timestamp); trace_cols[IS_READ][i] = F::from_bool(is_read); trace_cols[ADDR_CONTEXT][i] = F::from_canonical_usize(context); trace_cols[ADDR_SEGMENT][i] = F::from_canonical_usize(segment); trace_cols[ADDR_VIRTUAL][i] = F::from_canonical_usize(virt); for j in 0..VALUE_LIMBS { trace_cols[value_limb(j)][i] = value[j]; } } self.generate_memory(&mut trace_cols); let mut trace_rows = vec![[F::ZERO; NUM_COLUMNS]; num_ops]; for (i, col) in trace_cols.iter().enumerate() { for (j, &val) in col.iter().enumerate() { trace_rows[j][i] = val; } } trace_rows } fn pad_memory_ops(memory_ops: &mut Vec>) { let num_ops = memory_ops.len(); let max_range_check = get_max_range_check(&memory_ops); let num_ops_padded = num_ops.max(max_range_check + 1).next_power_of_two(); let to_pad = num_ops_padded - num_ops; let last_op = memory_ops.last().expect("No memory ops?").clone(); // We essentially repeat the last operation until our operation list has the desired size, // with a few changes: // - We change its channel to `None` to indicate that this is a dummy operation. // - We increment its timestamp in order to pass the ordering check. // - We make sure it's a read, sine dummy operations must be reads. for i in 0..to_pad { memory_ops.push(MemoryOp { channel_index: None, timestamp: last_op.timestamp + i + 1, is_read: true, ..last_op }); } } fn generate_memory(&self, trace_cols: &mut [Vec]) { let num_trace_rows = trace_cols[0].len(); let timestamp = &trace_cols[TIMESTAMP]; let context = &trace_cols[ADDR_CONTEXT]; let segment = &trace_cols[ADDR_SEGMENT]; let virtuals = &trace_cols[ADDR_VIRTUAL]; let (context_first_change, segment_first_change, virtual_first_change) = generate_first_change_flags(context, segment, virtuals); trace_cols[RANGE_CHECK] = generate_range_check_value( context, segment, virtuals, timestamp, &context_first_change, &segment_first_change, &virtual_first_change, ); trace_cols[CONTEXT_FIRST_CHANGE] = context_first_change; trace_cols[SEGMENT_FIRST_CHANGE] = segment_first_change; trace_cols[VIRTUAL_FIRST_CHANGE] = virtual_first_change; trace_cols[COUNTER] = (0..num_trace_rows) .map(|i| F::from_canonical_usize(i)) .collect(); let (permuted_inputs, permuted_table) = permuted_cols(&trace_cols[RANGE_CHECK], &trace_cols[COUNTER]); trace_cols[RANGE_CHECK_PERMUTED] = permuted_inputs; trace_cols[COUNTER_PERMUTED] = permuted_table; } pub fn generate_trace(&self, memory_ops: Vec>) -> Vec> { let mut timing = TimingTree::new("generate trace", log::Level::Debug); // Generate the witness. let trace_rows = timed!( &mut timing, "generate trace rows", self.generate_trace_rows(memory_ops) ); let trace_polys = timed!( &mut timing, "convert to PolynomialValues", trace_rows_to_poly_values(trace_rows) ); timing.print(); trace_polys } } impl, const D: usize> Stark for MemoryStark { const COLUMNS: usize = NUM_COLUMNS; const PUBLIC_INPUTS: usize = NUM_PUBLIC_INPUTS; fn eval_packed_generic( &self, vars: StarkEvaluationVars, yield_constr: &mut ConstraintConsumer

, ) where FE: FieldExtension, P: PackedField, { let one = P::from(FE::ONE); let timestamp = vars.local_values[TIMESTAMP]; let addr_context = vars.local_values[ADDR_CONTEXT]; let addr_segment = vars.local_values[ADDR_SEGMENT]; let addr_virtual = vars.local_values[ADDR_VIRTUAL]; let values: Vec<_> = (0..8).map(|i| vars.local_values[value_limb(i)]).collect(); let next_timestamp = vars.next_values[TIMESTAMP]; let next_is_read = vars.next_values[IS_READ]; let next_addr_context = vars.next_values[ADDR_CONTEXT]; let next_addr_segment = vars.next_values[ADDR_SEGMENT]; let next_addr_virtual = vars.next_values[ADDR_VIRTUAL]; let next_values: Vec<_> = (0..8).map(|i| vars.next_values[value_limb(i)]).collect(); // Each `is_channel` value must be 0 or 1. for c in 0..NUM_CHANNELS { let is_channel = vars.local_values[is_channel(c)]; yield_constr.constraint(is_channel * (is_channel - P::ONES)); } // The sum of `is_channel` flags, `has_channel`, must also be 0 or 1. let has_channel: P = (0..NUM_CHANNELS) .map(|c| vars.local_values[is_channel(c)]) .sum(); yield_constr.constraint(has_channel * (has_channel - P::ONES)); // If this is a dummy row (with no channel), it must be a read. This means the prover can // insert reads which never appear in the CPU trace (which are harmless), but not writes. let is_dummy = P::ONES - has_channel; let is_write = P::ONES - vars.local_values[IS_READ]; yield_constr.constraint(is_dummy * is_write); let context_first_change = vars.local_values[CONTEXT_FIRST_CHANGE]; let segment_first_change = vars.local_values[SEGMENT_FIRST_CHANGE]; let virtual_first_change = vars.local_values[VIRTUAL_FIRST_CHANGE]; let address_unchanged = one - context_first_change - segment_first_change - virtual_first_change; let range_check = vars.local_values[RANGE_CHECK]; let not_context_first_change = one - context_first_change; let not_segment_first_change = one - segment_first_change; let not_virtual_first_change = one - virtual_first_change; let not_address_unchanged = one - address_unchanged; // First set of ordering constraint: first_change flags are boolean. yield_constr.constraint(context_first_change * not_context_first_change); yield_constr.constraint(segment_first_change * not_segment_first_change); yield_constr.constraint(virtual_first_change * not_virtual_first_change); yield_constr.constraint(address_unchanged * not_address_unchanged); // Second set of ordering constraints: no change before the column corresponding to the nonzero first_change flag. yield_constr .constraint_transition(segment_first_change * (next_addr_context - addr_context)); yield_constr .constraint_transition(virtual_first_change * (next_addr_context - addr_context)); yield_constr .constraint_transition(virtual_first_change * (next_addr_segment - addr_segment)); yield_constr.constraint_transition(address_unchanged * (next_addr_context - addr_context)); yield_constr.constraint_transition(address_unchanged * (next_addr_segment - addr_segment)); yield_constr.constraint_transition(address_unchanged * (next_addr_virtual - addr_virtual)); // Third set of ordering constraints: range-check difference in the column that should be increasing. let computed_range_check = context_first_change * (next_addr_context - addr_context - one) + segment_first_change * (next_addr_segment - addr_segment - one) + virtual_first_change * (next_addr_virtual - addr_virtual - one) + address_unchanged * (next_timestamp - timestamp - one); yield_constr.constraint_transition(range_check - computed_range_check); // Enumerate purportedly-ordered log. for i in 0..8 { yield_constr .constraint(next_is_read * address_unchanged * (next_values[i] - values[i])); } eval_lookups(vars, yield_constr, RANGE_CHECK_PERMUTED, COUNTER_PERMUTED) } fn eval_ext_circuit( &self, builder: &mut plonky2::plonk::circuit_builder::CircuitBuilder, vars: StarkEvaluationTargets, yield_constr: &mut RecursiveConstraintConsumer, ) { let one = builder.one_extension(); let addr_context = vars.local_values[ADDR_CONTEXT]; let addr_segment = vars.local_values[ADDR_SEGMENT]; let addr_virtual = vars.local_values[ADDR_VIRTUAL]; let values: Vec<_> = (0..8).map(|i| vars.local_values[value_limb(i)]).collect(); let timestamp = vars.local_values[TIMESTAMP]; let next_addr_context = vars.next_values[ADDR_CONTEXT]; let next_addr_segment = vars.next_values[ADDR_SEGMENT]; let next_addr_virtual = vars.next_values[ADDR_VIRTUAL]; let next_values: Vec<_> = (0..8).map(|i| vars.next_values[value_limb(i)]).collect(); let next_is_read = vars.next_values[IS_READ]; let next_timestamp = vars.next_values[TIMESTAMP]; // Each `is_channel` value must be 0 or 1. for c in 0..NUM_CHANNELS { let is_channel = vars.local_values[is_channel(c)]; let constraint = builder.mul_sub_extension(is_channel, is_channel, is_channel); yield_constr.constraint(builder, constraint); } // The sum of `is_channel` flags, `has_channel`, must also be 0 or 1. let has_channel = builder.add_many_extension((0..NUM_CHANNELS).map(|c| vars.local_values[is_channel(c)])); let has_channel_bool = builder.mul_sub_extension(has_channel, has_channel, has_channel); yield_constr.constraint(builder, has_channel_bool); // If this is a dummy row (with no channel), it must be a read. This means the prover can // insert reads which never appear in the CPU trace (which are harmless), but not writes. let is_dummy = builder.sub_extension(one, has_channel); let is_write = builder.sub_extension(one, vars.local_values[IS_READ]); let is_dummy_write = builder.mul_extension(is_dummy, is_write); yield_constr.constraint(builder, is_dummy_write); let context_first_change = vars.local_values[CONTEXT_FIRST_CHANGE]; let segment_first_change = vars.local_values[SEGMENT_FIRST_CHANGE]; let virtual_first_change = vars.local_values[VIRTUAL_FIRST_CHANGE]; let address_unchanged = { let mut cur = builder.sub_extension(one, context_first_change); cur = builder.sub_extension(cur, segment_first_change); builder.sub_extension(cur, virtual_first_change) }; let range_check = vars.local_values[RANGE_CHECK]; let not_context_first_change = builder.sub_extension(one, context_first_change); let not_segment_first_change = builder.sub_extension(one, segment_first_change); let not_virtual_first_change = builder.sub_extension(one, virtual_first_change); let not_address_unchanged = builder.sub_extension(one, address_unchanged); let addr_context_diff = builder.sub_extension(next_addr_context, addr_context); let addr_segment_diff = builder.sub_extension(next_addr_segment, addr_segment); let addr_virtual_diff = builder.sub_extension(next_addr_virtual, addr_virtual); // First set of ordering constraint: traces are boolean. let context_first_change_bool = builder.mul_extension(context_first_change, not_context_first_change); yield_constr.constraint(builder, context_first_change_bool); let segment_first_change_bool = builder.mul_extension(segment_first_change, not_segment_first_change); yield_constr.constraint(builder, segment_first_change_bool); let virtual_first_change_bool = builder.mul_extension(virtual_first_change, not_virtual_first_change); yield_constr.constraint(builder, virtual_first_change_bool); let address_unchanged_bool = builder.mul_extension(address_unchanged, not_address_unchanged); yield_constr.constraint(builder, address_unchanged_bool); // Second set of ordering constraints: no change before the column corresponding to the nonzero first_change flag. let segment_first_change_check = builder.mul_extension(segment_first_change, addr_context_diff); yield_constr.constraint_transition(builder, segment_first_change_check); let virtual_first_change_check_1 = builder.mul_extension(virtual_first_change, addr_context_diff); yield_constr.constraint_transition(builder, virtual_first_change_check_1); let virtual_first_change_check_2 = builder.mul_extension(virtual_first_change, addr_segment_diff); yield_constr.constraint_transition(builder, virtual_first_change_check_2); let address_unchanged_check_1 = builder.mul_extension(address_unchanged, addr_context_diff); yield_constr.constraint_transition(builder, address_unchanged_check_1); let address_unchanged_check_2 = builder.mul_extension(address_unchanged, addr_segment_diff); yield_constr.constraint_transition(builder, address_unchanged_check_2); let address_unchanged_check_3 = builder.mul_extension(address_unchanged, addr_virtual_diff); yield_constr.constraint_transition(builder, address_unchanged_check_3); // Third set of ordering constraints: range-check difference in the column that should be increasing. let context_diff = { let diff = builder.sub_extension(next_addr_context, addr_context); builder.sub_extension(diff, one) }; let context_range_check = builder.mul_extension(context_first_change, context_diff); let segment_diff = { let diff = builder.sub_extension(next_addr_segment, addr_segment); builder.sub_extension(diff, one) }; let segment_range_check = builder.mul_extension(segment_first_change, segment_diff); let virtual_diff = { let diff = builder.sub_extension(next_addr_virtual, addr_virtual); builder.sub_extension(diff, one) }; let virtual_range_check = builder.mul_extension(virtual_first_change, virtual_diff); let timestamp_diff = { let diff = builder.sub_extension(next_timestamp, timestamp); builder.sub_extension(diff, one) }; let timestamp_range_check = builder.mul_extension(address_unchanged, timestamp_diff); let computed_range_check = { let mut sum = builder.add_extension(context_range_check, segment_range_check); sum = builder.add_extension(sum, virtual_range_check); builder.add_extension(sum, timestamp_range_check) }; let range_check_diff = builder.sub_extension(range_check, computed_range_check); yield_constr.constraint_transition(builder, range_check_diff); // Enumerate purportedly-ordered log. for i in 0..8 { let value_diff = builder.sub_extension(next_values[i], values[i]); let zero_if_read = builder.mul_extension(address_unchanged, value_diff); let read_constraint = builder.mul_extension(next_is_read, zero_if_read); yield_constr.constraint(builder, read_constraint); } eval_lookups_circuit( builder, vars, yield_constr, RANGE_CHECK_PERMUTED, COUNTER_PERMUTED, ) } fn constraint_degree(&self) -> usize { 3 } fn permutation_pairs(&self) -> Vec { vec![ PermutationPair::singletons(RANGE_CHECK, RANGE_CHECK_PERMUTED), PermutationPair::singletons(COUNTER, COUNTER_PERMUTED), ] } } #[cfg(test)] mod tests { use anyhow::Result; use plonky2::plonk::config::{GenericConfig, PoseidonGoldilocksConfig}; use crate::memory::memory_stark::MemoryStark; use crate::stark_testing::{test_stark_circuit_constraints, test_stark_low_degree}; #[test] fn test_stark_degree() -> Result<()> { const D: usize = 2; type C = PoseidonGoldilocksConfig; type F = >::F; type S = MemoryStark; let stark = S { f: Default::default(), }; test_stark_low_degree(stark) } #[test] fn test_stark_circuit() -> Result<()> { const D: usize = 2; type C = PoseidonGoldilocksConfig; type F = >::F; type S = MemoryStark; let stark = S { f: Default::default(), }; test_stark_circuit_constraints::(stark) } }