sp1_recursion_core/chips/
alu_base.rs

1use crate::{builder::SP1RecursionAirBuilder, *};
2use core::borrow::Borrow;
3use p3_air::{Air, AirBuilder, BaseAir, PairBuilder};
4use p3_baby_bear::BabyBear;
5use p3_field::{AbstractField, Field, PrimeField32};
6use p3_matrix::{dense::RowMajorMatrix, Matrix};
7use p3_maybe_rayon::prelude::*;
8use sp1_core_machine::utils::next_power_of_two;
9use sp1_derive::AlignedBorrow;
10use sp1_stark::air::MachineAir;
11use std::{borrow::BorrowMut, iter::zip};
12
13pub const NUM_BASE_ALU_ENTRIES_PER_ROW: usize = 4;
14
15#[derive(Default)]
16pub struct BaseAluChip;
17
18pub const NUM_BASE_ALU_COLS: usize = core::mem::size_of::<BaseAluCols<u8>>();
19
20#[derive(AlignedBorrow, Debug, Clone, Copy)]
21#[repr(C)]
22pub struct BaseAluCols<F: Copy> {
23    pub values: [BaseAluValueCols<F>; NUM_BASE_ALU_ENTRIES_PER_ROW],
24}
25
26pub const NUM_BASE_ALU_VALUE_COLS: usize = core::mem::size_of::<BaseAluValueCols<u8>>();
27
28#[derive(AlignedBorrow, Debug, Clone, Copy)]
29#[repr(C)]
30pub struct BaseAluValueCols<F: Copy> {
31    pub vals: BaseAluIo<F>,
32}
33
34pub const NUM_BASE_ALU_PREPROCESSED_COLS: usize =
35    core::mem::size_of::<BaseAluPreprocessedCols<u8>>();
36
37#[derive(AlignedBorrow, Debug, Clone, Copy)]
38#[repr(C)]
39pub struct BaseAluPreprocessedCols<F: Copy> {
40    pub accesses: [BaseAluAccessCols<F>; NUM_BASE_ALU_ENTRIES_PER_ROW],
41}
42
43pub const NUM_BASE_ALU_ACCESS_COLS: usize = core::mem::size_of::<BaseAluAccessCols<u8>>();
44
45#[derive(AlignedBorrow, Debug, Clone, Copy)]
46#[repr(C)]
47pub struct BaseAluAccessCols<F: Copy> {
48    pub addrs: BaseAluIo<Address<F>>,
49    pub is_add: F,
50    pub is_sub: F,
51    pub is_mul: F,
52    pub is_div: F,
53    pub mult: F,
54}
55
56impl<F: Field> BaseAir<F> for BaseAluChip {
57    fn width(&self) -> usize {
58        NUM_BASE_ALU_COLS
59    }
60}
61
62impl<F: PrimeField32> MachineAir<F> for BaseAluChip {
63    type Record = ExecutionRecord<F>;
64
65    type Program = crate::RecursionProgram<F>;
66
67    fn name(&self) -> String {
68        "BaseAlu".to_string()
69    }
70
71    fn preprocessed_width(&self) -> usize {
72        NUM_BASE_ALU_PREPROCESSED_COLS
73    }
74
75    fn preprocessed_num_rows(&self, program: &Self::Program, instrs_len: usize) -> Option<usize> {
76        let nb_rows = instrs_len.div_ceil(NUM_BASE_ALU_ENTRIES_PER_ROW);
77        let fixed_log2_rows = program.fixed_log2_rows(self);
78        Some(match fixed_log2_rows {
79            Some(log2_rows) => 1 << log2_rows,
80            None => next_power_of_two(nb_rows, None),
81        })
82    }
83
84    fn generate_preprocessed_trace(&self, program: &Self::Program) -> Option<RowMajorMatrix<F>> {
85        assert_eq!(
86            std::any::TypeId::of::<F>(),
87            std::any::TypeId::of::<BabyBear>(),
88            "generate_preprocessed_trace only supports BabyBear field"
89        );
90
91        let instrs = unsafe {
92            std::mem::transmute::<Vec<&BaseAluInstr<F>>, Vec<&BaseAluInstr<BabyBear>>>(
93                program
94                    .inner
95                    .iter()
96                    .filter_map(|instruction| match instruction {
97                        Instruction::BaseAlu(x) => Some(x),
98                        _ => None,
99                    })
100                    .collect::<Vec<_>>(),
101            )
102        };
103        let padded_nb_rows = self.preprocessed_num_rows(program, instrs.len()).unwrap();
104        let mut values = vec![BabyBear::zero(); padded_nb_rows * NUM_BASE_ALU_PREPROCESSED_COLS];
105
106        // Generate the trace rows & corresponding records for each chunk of events in parallel.
107        let populate_len = instrs.len() * NUM_BASE_ALU_ACCESS_COLS;
108        values[..populate_len].par_chunks_mut(NUM_BASE_ALU_ACCESS_COLS).zip_eq(instrs).for_each(
109            |(row, instr)| {
110                let access: &mut BaseAluAccessCols<_> = row.borrow_mut();
111                unsafe {
112                    crate::sys::alu_base_instr_to_row_babybear(instr, access);
113                }
114            },
115        );
116
117        // Convert the trace to a row major matrix.
118        Some(RowMajorMatrix::new(
119            unsafe { std::mem::transmute::<Vec<BabyBear>, Vec<F>>(values) },
120            NUM_BASE_ALU_PREPROCESSED_COLS,
121        ))
122    }
123
124    fn generate_dependencies(&self, _: &Self::Record, _: &mut Self::Record) {
125        // This is a no-op.
126    }
127
128    fn num_rows(&self, input: &Self::Record) -> Option<usize> {
129        let nb_rows = input.base_alu_events.len().div_ceil(NUM_BASE_ALU_ENTRIES_PER_ROW);
130        let fixed_log2_rows = input.fixed_log2_rows(self);
131        Some(match fixed_log2_rows {
132            Some(log2_rows) => 1 << log2_rows,
133            None => next_power_of_two(nb_rows, None),
134        })
135    }
136
137    fn generate_trace(&self, input: &Self::Record, _: &mut Self::Record) -> RowMajorMatrix<F> {
138        assert_eq!(
139            std::any::TypeId::of::<F>(),
140            std::any::TypeId::of::<BabyBear>(),
141            "generate_trace only supports BabyBear field"
142        );
143
144        let events = unsafe {
145            std::mem::transmute::<&Vec<BaseAluIo<F>>, &Vec<BaseAluIo<BabyBear>>>(
146                &input.base_alu_events,
147            )
148        };
149        let padded_nb_rows = self.num_rows(input).unwrap();
150        let mut values = vec![BabyBear::zero(); padded_nb_rows * NUM_BASE_ALU_COLS];
151
152        // Generate the trace rows & corresponding records for each chunk of events in parallel.
153        let populate_len = events.len() * NUM_BASE_ALU_VALUE_COLS;
154        values[..populate_len].par_chunks_mut(NUM_BASE_ALU_VALUE_COLS).zip_eq(events).for_each(
155            |(row, &vals)| {
156                let cols: &mut BaseAluValueCols<_> = row.borrow_mut();
157                unsafe {
158                    crate::sys::alu_base_event_to_row_babybear(&vals, cols);
159                }
160            },
161        );
162
163        // Convert the trace to a row major matrix.
164        RowMajorMatrix::new(
165            unsafe { std::mem::transmute::<Vec<BabyBear>, Vec<F>>(values) },
166            NUM_BASE_ALU_COLS,
167        )
168    }
169
170    fn included(&self, _record: &Self::Record) -> bool {
171        true
172    }
173
174    fn local_only(&self) -> bool {
175        true
176    }
177}
178
179impl<AB> Air<AB> for BaseAluChip
180where
181    AB: SP1RecursionAirBuilder + PairBuilder,
182{
183    fn eval(&self, builder: &mut AB) {
184        let main = builder.main();
185        let local = main.row_slice(0);
186        let local: &BaseAluCols<AB::Var> = (*local).borrow();
187        let prep = builder.preprocessed();
188        let prep_local = prep.row_slice(0);
189        let prep_local: &BaseAluPreprocessedCols<AB::Var> = (*prep_local).borrow();
190
191        for (
192            BaseAluValueCols { vals: BaseAluIo { out, in1, in2 } },
193            BaseAluAccessCols { addrs, is_add, is_sub, is_mul, is_div, mult },
194        ) in zip(local.values, prep_local.accesses)
195        {
196            // Check exactly one flag is enabled.
197            let is_real = is_add + is_sub + is_mul + is_div;
198            builder.assert_bool(is_real.clone());
199
200            builder.when(is_add).assert_eq(in1 + in2, out);
201            builder.when(is_sub).assert_eq(in1, in2 + out);
202            builder.when(is_mul).assert_eq(out, in1 * in2);
203            builder.when(is_div).assert_eq(in2 * out, in1);
204
205            builder.receive_single(addrs.in1, in1, is_real.clone());
206
207            builder.receive_single(addrs.in2, in2, is_real);
208
209            builder.send_single(addrs.out, out, mult);
210        }
211    }
212}
213
214#[cfg(test)]
215mod tests {
216    use crate::{chips::test_fixtures, runtime::instruction as instr};
217    use machine::tests::test_recursion_linear_program;
218    use p3_baby_bear::BabyBear;
219    use p3_field::AbstractField;
220    use p3_matrix::dense::RowMajorMatrix;
221    use rand::{rngs::StdRng, Rng, SeedableRng};
222    use sp1_stark::{baby_bear_poseidon2::BabyBearPoseidon2, StarkGenericConfig};
223
224    use super::*;
225
226    fn generate_trace_reference(
227        input: &ExecutionRecord<BabyBear>,
228        _: &mut ExecutionRecord<BabyBear>,
229    ) -> RowMajorMatrix<BabyBear> {
230        let events = &input.base_alu_events;
231        let padded_nb_rows = BaseAluChip.num_rows(input).unwrap();
232        let mut values = vec![BabyBear::zero(); padded_nb_rows * NUM_BASE_ALU_COLS];
233
234        let populate_len = events.len() * NUM_BASE_ALU_VALUE_COLS;
235        values[..populate_len].par_chunks_mut(NUM_BASE_ALU_VALUE_COLS).zip_eq(events).for_each(
236            |(row, &vals)| {
237                let cols: &mut BaseAluValueCols<_> = row.borrow_mut();
238                *cols = BaseAluValueCols { vals };
239            },
240        );
241
242        RowMajorMatrix::new(values, NUM_BASE_ALU_COLS)
243    }
244
245    #[test]
246    fn generate_trace() {
247        let shard = test_fixtures::shard();
248        let mut execution_record = test_fixtures::default_execution_record();
249        let trace = BaseAluChip.generate_trace(&shard, &mut execution_record);
250        assert!(trace.height() >= test_fixtures::MIN_TEST_CASES);
251
252        assert_eq!(trace, generate_trace_reference(&shard, &mut execution_record));
253    }
254
255    fn generate_preprocessed_trace_reference(
256        program: &RecursionProgram<BabyBear>,
257    ) -> RowMajorMatrix<BabyBear> {
258        type F = BabyBear;
259
260        let instrs = program
261            .inner
262            .iter()
263            .filter_map(|instruction| match instruction {
264                Instruction::BaseAlu(x) => Some(x),
265                _ => None,
266            })
267            .collect::<Vec<_>>();
268        let padded_nb_rows = BaseAluChip.preprocessed_num_rows(program, instrs.len()).unwrap();
269        let mut values = vec![F::zero(); padded_nb_rows * NUM_BASE_ALU_PREPROCESSED_COLS];
270
271        let populate_len = instrs.len() * NUM_BASE_ALU_ACCESS_COLS;
272        values[..populate_len].par_chunks_mut(NUM_BASE_ALU_ACCESS_COLS).zip_eq(instrs).for_each(
273            |(row, instr)| {
274                let BaseAluInstr { opcode, mult, addrs } = instr;
275                let access: &mut BaseAluAccessCols<_> = row.borrow_mut();
276                *access = BaseAluAccessCols {
277                    addrs: addrs.to_owned(),
278                    is_add: F::from_bool(false),
279                    is_sub: F::from_bool(false),
280                    is_mul: F::from_bool(false),
281                    is_div: F::from_bool(false),
282                    mult: mult.to_owned(),
283                };
284                let target_flag = match opcode {
285                    BaseAluOpcode::AddF => &mut access.is_add,
286                    BaseAluOpcode::SubF => &mut access.is_sub,
287                    BaseAluOpcode::MulF => &mut access.is_mul,
288                    BaseAluOpcode::DivF => &mut access.is_div,
289                };
290                *target_flag = F::from_bool(true);
291            },
292        );
293
294        RowMajorMatrix::new(values, NUM_BASE_ALU_PREPROCESSED_COLS)
295    }
296
297    #[test]
298    #[ignore = "Failing due to merge conflicts. Will be fixed shortly."]
299    fn generate_preprocessed_trace() {
300        let program = test_fixtures::program();
301        let trace = BaseAluChip.generate_preprocessed_trace(&program).unwrap();
302        assert!(trace.height() >= test_fixtures::MIN_TEST_CASES);
303
304        assert_eq!(trace, generate_preprocessed_trace_reference(&program));
305    }
306
307    #[test]
308    pub fn four_ops() {
309        type SC = BabyBearPoseidon2;
310        type F = <SC as StarkGenericConfig>::Val;
311
312        let mut rng = StdRng::seed_from_u64(0xDEADBEEF);
313        let mut random_felt = move || -> F { rng.sample(rand::distributions::Standard) };
314        let mut addr = 0;
315
316        let instructions = (0..1000)
317            .flat_map(|_| {
318                let quot = random_felt();
319                let in2 = random_felt();
320                let in1 = in2 * quot;
321                let alloc_size = 6;
322                let a = (0..alloc_size).map(|x| x + addr).collect::<Vec<_>>();
323                addr += alloc_size;
324                [
325                    instr::mem_single(MemAccessKind::Write, 4, a[0], in1),
326                    instr::mem_single(MemAccessKind::Write, 4, a[1], in2),
327                    instr::base_alu(BaseAluOpcode::AddF, 1, a[2], a[0], a[1]),
328                    instr::mem_single(MemAccessKind::Read, 1, a[2], in1 + in2),
329                    instr::base_alu(BaseAluOpcode::SubF, 1, a[3], a[0], a[1]),
330                    instr::mem_single(MemAccessKind::Read, 1, a[3], in1 - in2),
331                    instr::base_alu(BaseAluOpcode::MulF, 1, a[4], a[0], a[1]),
332                    instr::mem_single(MemAccessKind::Read, 1, a[4], in1 * in2),
333                    instr::base_alu(BaseAluOpcode::DivF, 1, a[5], a[0], a[1]),
334                    instr::mem_single(MemAccessKind::Read, 1, a[5], quot),
335                ]
336            })
337            .collect::<Vec<Instruction<F>>>();
338
339        test_recursion_linear_program(instructions);
340    }
341}