codanna 0.9.19

Code Intelligence for Large Language Models
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
//! Memory-mapped vector storage for high-performance vector access.
//!
//! This module provides efficient storage and retrieval of embedding vectors
//! using memory-mapped files. The implementation achieves <1μs vector access
//! times by avoiding serialization overhead and leveraging OS page cache.
//!
//! # Storage Format
//!
//! The storage uses a simple binary format optimized for sequential access:
//! - Header (16 bytes): version, dimension, vector count
//! - Vectors: Contiguous f32 arrays in little-endian format
//!
//! # Performance Characteristics
//!
//! - Vector access: <1μs (memory-mapped, no deserialization)
//! - Memory usage: 4 bytes per dimension per vector
//! - Startup time: <1ms (mmap is lazy-loaded by OS)

use std::fs::{File, OpenOptions};
use std::io::{self, Write};
use std::path::{Path, PathBuf};
use std::sync::Arc;

use memmap2::{Mmap, MmapOptions};
use thiserror::Error;

use crate::vector::types::{SegmentOrdinal, VectorDimension, VectorError, VectorId};

/// Current storage format version.
const STORAGE_VERSION: u32 = 1;

/// Size of the storage header in bytes.
const HEADER_SIZE: usize = 16;

/// Magic bytes to identify vector storage files.
const MAGIC_BYTES: &[u8; 4] = b"CVEC";

/// Number of bytes per f32 value.
const BYTES_PER_F32: usize = 4;

/// Number of bytes per vector ID (u32).
const BYTES_PER_ID: usize = 4;

/// Errors specific to vector storage operations.
#[derive(Error, Debug)]
pub enum VectorStorageError {
    #[error("IO error: {0}")]
    Io(#[from] io::Error),

    #[error("Invalid storage format: {0}")]
    InvalidFormat(String),

    #[error("Vector error: {0}")]
    Vector(#[from] VectorError),
}

/// Memory-mapped vector storage for a single segment.
///
/// Provides efficient read/write access to embedding vectors with
/// minimal memory overhead and <1μs access times.
#[derive(Debug)]
pub struct MmapVectorStorage {
    /// Path to the storage file.
    path: PathBuf,

    /// Memory-mapped file for reading.
    mmap: Option<Mmap>,

    /// Vector dimension (all vectors must have same dimension).
    dimension: VectorDimension,

    /// Number of vectors currently stored.
    vector_count: usize,

    /// Segment this storage belongs to.
    segment: SegmentOrdinal,
}

impl MmapVectorStorage {
    /// Creates a new vector storage for the given segment.
    ///
    /// # Arguments
    /// * `base_path` - Directory where vector files will be stored
    /// * `segment` - Segment ordinal this storage belongs to
    /// * `dimension` - Dimension of vectors to be stored
    pub fn new(
        base_path: impl AsRef<Path>,
        segment: SegmentOrdinal,
        dimension: VectorDimension,
    ) -> Result<Self, VectorStorageError> {
        let path = Self::segment_path(base_path.as_ref(), segment);

        Ok(Self {
            path,
            mmap: None,
            dimension,
            vector_count: 0,
            segment,
        })
    }

    /// Opens existing vector storage from disk.
    ///
    /// Returns an error if the file doesn't exist or has invalid format.
    pub fn open(
        base_path: impl AsRef<Path>,
        segment: SegmentOrdinal,
    ) -> Result<Self, VectorStorageError> {
        let path = Self::segment_path(base_path.as_ref(), segment);

        if !path.exists() {
            return Err(VectorStorageError::Io(io::Error::new(
                io::ErrorKind::NotFound,
                format!("Vector storage file not found: {path:?}"),
            )));
        }

        let file = File::open(&path)?;
        let mmap = unsafe { MmapOptions::new().map(&file)? };

        // Read and validate header
        let (version, dimension, vector_count) = Self::read_header(&mmap)?;

        if version != STORAGE_VERSION {
            return Err(VectorError::VersionMismatch {
                expected: STORAGE_VERSION,
                actual: version,
            }
            .into());
        }

        Ok(Self {
            path,
            mmap: Some(mmap),
            dimension,
            vector_count,
            segment,
        })
    }

    /// Creates or opens vector storage, initializing if necessary.
    pub fn open_or_create(
        base_path: impl AsRef<Path>,
        segment: SegmentOrdinal,
        dimension: VectorDimension,
    ) -> Result<Self, VectorStorageError> {
        let path = Self::segment_path(base_path.as_ref(), segment);

        if path.exists() {
            Self::open(base_path, segment)
        } else {
            let mut storage = Self::new(base_path, segment, dimension)?;
            storage.initialize()?;
            Ok(storage)
        }
    }

    /// Writes a batch of vectors to storage.
    ///
    /// This is more efficient than writing vectors one by one as it
    /// minimizes file operations and can pre-allocate space.
    pub fn write_batch(
        &mut self,
        vectors: &[(VectorId, &[f32])],
    ) -> Result<(), VectorStorageError> {
        // Convert to owned for validation and writing
        let owned_vectors: Vec<(VectorId, Vec<f32>)> = vectors
            .iter()
            .map(|(id, vec)| (*id, vec.to_vec()))
            .collect();
        self.validate_vectors(&owned_vectors)?;
        self.ensure_storage_ready()?;
        self.append_vectors(&owned_vectors)?;
        self.update_metadata(vectors.len())?;
        self.invalidate_cache();
        Ok(())
    }

    /// Validates that all vectors have the correct dimension.
    fn validate_vectors(&self, vectors: &[(VectorId, Vec<f32>)]) -> Result<(), VectorStorageError> {
        for (_, vec) in vectors {
            self.dimension.validate_vector(vec)?;
        }
        Ok(())
    }

    /// Ensures the storage directory exists and is ready for writing.
    fn ensure_storage_ready(&self) -> Result<(), VectorStorageError> {
        if let Some(parent) = self.path.parent() {
            std::fs::create_dir_all(parent)?;
        }
        Ok(())
    }

    /// Appends vectors to the storage file.
    fn append_vectors(&self, vectors: &[(VectorId, Vec<f32>)]) -> Result<(), VectorStorageError> {
        let mut file = OpenOptions::new()
            .create(true)
            .append(true)
            .open(&self.path)?;

        // Write header if this is a new file
        if file.metadata()?.len() == 0 {
            self.write_header(&mut file)?;
        }

        // Write vectors
        for (id, vector) in vectors {
            // Write vector ID
            file.write_all(&id.to_bytes())?;

            // Write vector data
            for &value in vector {
                file.write_all(&value.to_le_bytes())?;
            }
        }

        file.flush()?;
        Ok(())
    }

    /// Updates metadata after writing vectors.
    fn update_metadata(&mut self, vector_count: usize) -> Result<(), VectorStorageError> {
        self.vector_count += vector_count;
        self.update_header_count()?;
        Ok(())
    }

    /// Invalidates the memory map cache to force reload on next read.
    fn invalidate_cache(&mut self) {
        self.mmap = None;
    }

    /// Reads a vector by its ID.
    ///
    /// Returns `None` if the vector is not found.
    /// This operation is extremely fast (<1μs) due to memory mapping.
    #[must_use]
    pub fn read_vector(&mut self, id: VectorId) -> Option<Vec<f32>> {
        self.ensure_mapped().ok()?;
        let mmap = self.mmap.as_ref()?;

        let dimension = self.dimension.get();
        let vector_size = BYTES_PER_ID + dimension * BYTES_PER_F32;

        // Search for vector with matching ID
        let mut offset = HEADER_SIZE;
        while offset + vector_size <= mmap.len() {
            // Read vector ID
            let stored_id = u32::from_le_bytes([
                mmap[offset],
                mmap[offset + 1],
                mmap[offset + 2],
                mmap[offset + 3],
            ]);

            if stored_id == id.get() {
                // Found it! Read vector data
                let mut vector = Vec::with_capacity(dimension);
                let data_offset = offset + BYTES_PER_ID;

                for i in 0..dimension {
                    let bytes_offset = data_offset + i * BYTES_PER_F32;
                    let value = f32::from_le_bytes([
                        mmap[bytes_offset],
                        mmap[bytes_offset + 1],
                        mmap[bytes_offset + 2],
                        mmap[bytes_offset + 3],
                    ]);
                    vector.push(value);
                }

                return Some(vector);
            }

            offset += vector_size;
        }

        None
    }

    /// Reads all vectors from storage.
    ///
    /// This is useful for operations that need to process all vectors,
    /// such as clustering or batch similarity search.
    pub fn read_all_vectors(&mut self) -> Result<Vec<(VectorId, Vec<f32>)>, VectorStorageError> {
        self.ensure_mapped()?;
        let mmap = self.mmap.as_ref().unwrap();

        let dimension = self.dimension.get();
        let vector_size = BYTES_PER_ID + dimension * BYTES_PER_F32;
        let mut vectors = Vec::with_capacity(self.vector_count);

        let mut offset = HEADER_SIZE;
        while offset + vector_size <= mmap.len() {
            // Read vector ID
            let id_bytes = [
                mmap[offset],
                mmap[offset + 1],
                mmap[offset + 2],
                mmap[offset + 3],
            ];
            let id = VectorId::from_bytes(id_bytes).ok_or_else(|| {
                VectorStorageError::InvalidFormat("Invalid vector ID".to_string())
            })?;

            // Read vector data
            let mut vector = Vec::with_capacity(dimension);
            let data_offset = offset + BYTES_PER_ID;

            for i in 0..dimension {
                let bytes_offset = data_offset + i * BYTES_PER_F32;
                let value = f32::from_le_bytes([
                    mmap[bytes_offset],
                    mmap[bytes_offset + 1],
                    mmap[bytes_offset + 2],
                    mmap[bytes_offset + 3],
                ]);
                vector.push(value);
            }

            vectors.push((id, vector));
            offset += vector_size;
        }

        Ok(vectors)
    }

    /// Returns the number of vectors stored.
    #[must_use]
    pub fn vector_count(&self) -> usize {
        self.vector_count
    }

    /// Returns the vector dimension.
    #[must_use]
    pub fn dimension(&self) -> VectorDimension {
        self.dimension
    }

    /// Returns the segment this storage belongs to.
    #[must_use]
    pub fn segment(&self) -> SegmentOrdinal {
        self.segment
    }

    /// Checks if the storage file exists on disk.
    #[must_use]
    pub fn exists(&self) -> bool {
        self.path.exists()
    }

    /// Returns the size of the storage file in bytes.
    pub fn file_size(&self) -> Result<u64, io::Error> {
        Ok(std::fs::metadata(&self.path)?.len())
    }

    // Private helper methods

    fn segment_path(base_path: &Path, segment: SegmentOrdinal) -> PathBuf {
        base_path.join(format!("segment_{}.vec", segment.get()))
    }

    fn initialize(&mut self) -> Result<(), VectorStorageError> {
        if let Some(parent) = self.path.parent() {
            std::fs::create_dir_all(parent)?;
        }

        let mut file = File::create(&self.path)?;
        self.write_header(&mut file)?;
        file.flush()?;

        Ok(())
    }

    fn write_header(&self, file: &mut File) -> Result<(), io::Error> {
        // Magic bytes
        file.write_all(MAGIC_BYTES)?;

        // Version
        file.write_all(&STORAGE_VERSION.to_le_bytes())?;

        // Dimension
        file.write_all(&(self.dimension.get() as u32).to_le_bytes())?;

        // Vector count (initially 0)
        file.write_all(&0u32.to_le_bytes())?;

        Ok(())
    }

    fn read_header(mmap: &Mmap) -> Result<(u32, VectorDimension, usize), VectorStorageError> {
        if mmap.len() < HEADER_SIZE {
            return Err(VectorStorageError::InvalidFormat(
                "File too small to contain header".to_string(),
            ));
        }

        // Check magic bytes
        if &mmap[0..4] != MAGIC_BYTES {
            return Err(VectorStorageError::InvalidFormat(
                "Invalid magic bytes".to_string(),
            ));
        }

        // Read version
        let version = u32::from_le_bytes([mmap[4], mmap[5], mmap[6], mmap[7]]);

        // Read dimension
        let dim_value = u32::from_le_bytes([mmap[8], mmap[9], mmap[10], mmap[11]]);
        let dimension = VectorDimension::new(dim_value as usize)?;

        // Read vector count
        let vector_count = u32::from_le_bytes([mmap[12], mmap[13], mmap[14], mmap[15]]) as usize;

        Ok((version, dimension, vector_count))
    }

    fn ensure_mapped(&mut self) -> Result<(), VectorStorageError> {
        if self.mmap.is_none() {
            let file = File::open(&self.path)?;
            let mmap = unsafe { MmapOptions::new().map(&file)? };

            // Update vector count from file
            let (_, _, count) = Self::read_header(&mmap)?;
            self.vector_count = count;
            self.mmap = Some(mmap);
        }
        Ok(())
    }

    fn update_header_count(&self) -> Result<(), VectorStorageError> {
        use std::io::{Seek, SeekFrom};

        let mut file = OpenOptions::new().write(true).open(&self.path)?;

        // Seek to vector count position in header (12 bytes offset)
        file.seek(SeekFrom::Start(12))?;

        // Write updated count
        file.write_all(&(self.vector_count as u32).to_le_bytes())?;
        file.flush()?;

        Ok(())
    }
}

/// Thread-safe wrapper for MmapVectorStorage.
///
/// Allows concurrent read access to vectors from multiple threads.
pub struct ConcurrentVectorStorage {
    inner: Arc<parking_lot::RwLock<MmapVectorStorage>>,
}

impl ConcurrentVectorStorage {
    /// Creates a new concurrent vector storage.
    pub fn new(storage: MmapVectorStorage) -> Self {
        Self {
            inner: Arc::new(parking_lot::RwLock::new(storage)),
        }
    }

    /// Reads a vector by ID with shared access.
    #[must_use]
    pub fn read_vector(&self, id: VectorId) -> Option<Vec<f32>> {
        self.inner.write().read_vector(id)
    }

    /// Writes a batch of vectors with exclusive access.
    pub fn write_batch(&self, vectors: &[(VectorId, &[f32])]) -> Result<(), VectorStorageError> {
        self.inner.write().write_batch(vectors).map_err(|e| {
            VectorStorageError::Io(io::Error::other(format!(
                "Concurrent write failed for {} vectors: {}",
                vectors.len(),
                e
            )))
        })
    }
}

impl Clone for MmapVectorStorage {
    fn clone(&self) -> Self {
        // Clone path and metadata, but not mmap (will be lazy-loaded)
        Self {
            path: self.path.clone(),
            mmap: None, // Force re-mapping on clone
            dimension: self.dimension,
            vector_count: self.vector_count,
            segment: self.segment,
        }
    }
}

impl PartialEq for MmapVectorStorage {
    fn eq(&self, other: &Self) -> bool {
        self.path == other.path
            && self.dimension == other.dimension
            && self.segment == other.segment
    }
}

impl std::fmt::Debug for ConcurrentVectorStorage {
    fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
        // Try to acquire read lock for debug output
        match self.inner.try_read() {
            Some(storage) => write!(f, "ConcurrentVectorStorage {{ storage: {storage:?} }}"),
            None => write!(f, "ConcurrentVectorStorage {{ <locked> }}"),
        }
    }
}

#[cfg(test)]
mod tests {
    use super::*;
    use tempfile::TempDir;

    #[test]
    fn test_storage_create_and_open() {
        let temp_dir = TempDir::new().unwrap();
        let segment = SegmentOrdinal::new(0);
        let dimension = VectorDimension::dimension_384();

        // Create new storage
        let storage = MmapVectorStorage::new(&temp_dir, segment, dimension).unwrap();
        assert_eq!(storage.vector_count(), 0);
        assert_eq!(storage.dimension(), dimension);

        // Open existing storage should fail (not initialized)
        assert!(MmapVectorStorage::open(&temp_dir, segment).is_err());
    }

    #[test]
    fn test_write_and_read_vectors() {
        let temp_dir = TempDir::new().unwrap();
        let segment = SegmentOrdinal::new(0);
        let dimension = VectorDimension::new(4).unwrap(); // Small dimension for testing

        let mut storage = MmapVectorStorage::open_or_create(&temp_dir, segment, dimension).unwrap();

        // Prepare test vectors
        let test_data = vec![
            (VectorId::new(1).unwrap(), vec![1.0, 2.0, 3.0, 4.0]),
            (VectorId::new(2).unwrap(), vec![5.0, 6.0, 7.0, 8.0]),
            (VectorId::new(3).unwrap(), vec![9.0, 10.0, 11.0, 12.0]),
        ];
        let vectors: Vec<(VectorId, &[f32])> = test_data
            .iter()
            .map(|(id, vec)| (*id, vec.as_slice()))
            .collect();

        // Write vectors
        storage.write_batch(&vectors).unwrap();
        assert_eq!(storage.vector_count(), 3);

        // Read vectors back
        for (id, expected_vector) in &test_data {
            let read_vector = storage.read_vector(*id).unwrap();
            assert_eq!(&read_vector, expected_vector);
        }

        // Non-existent vector should return None
        assert!(storage.read_vector(VectorId::new(999).unwrap()).is_none());
    }

    #[test]
    fn test_read_all_vectors() {
        let temp_dir = TempDir::new().unwrap();
        let segment = SegmentOrdinal::new(0);
        let dimension = VectorDimension::new(3).unwrap();

        let mut storage = MmapVectorStorage::open_or_create(&temp_dir, segment, dimension).unwrap();

        let test_data = vec![
            (VectorId::new(10).unwrap(), vec![1.0, 2.0, 3.0]),
            (VectorId::new(20).unwrap(), vec![4.0, 5.0, 6.0]),
        ];
        let vectors: Vec<(VectorId, &[f32])> = test_data
            .iter()
            .map(|(id, vec)| (*id, vec.as_slice()))
            .collect();

        storage.write_batch(&vectors).unwrap();

        let all_vectors = storage.read_all_vectors().unwrap();
        assert_eq!(all_vectors.len(), 2);
        assert_eq!(all_vectors, test_data);
    }

    #[test]
    fn test_dimension_validation() {
        let temp_dir = TempDir::new().unwrap();
        let segment = SegmentOrdinal::new(0);
        let dimension = VectorDimension::new(3).unwrap();

        let mut storage = MmapVectorStorage::open_or_create(&temp_dir, segment, dimension).unwrap();

        // Wrong dimension should fail
        let wrong_test_data = [(VectorId::new(1).unwrap(), vec![1.0, 2.0])];
        let wrong_vectors: Vec<(VectorId, &[f32])> = wrong_test_data
            .iter()
            .map(|(id, vec)| (*id, vec.as_slice()))
            .collect();

        assert!(storage.write_batch(&wrong_vectors).is_err());
    }

    #[test]
    fn test_persistence_across_instances() {
        let temp_dir = TempDir::new().unwrap();
        let segment = SegmentOrdinal::new(0);
        let dimension = VectorDimension::new(2).unwrap();

        // Write vectors with first instance
        {
            let mut storage =
                MmapVectorStorage::open_or_create(&temp_dir, segment, dimension).unwrap();
            let test_data = [
                (VectorId::new(1).unwrap(), vec![1.0, 2.0]),
                (VectorId::new(2).unwrap(), vec![3.0, 4.0]),
            ];
            let vectors: Vec<(VectorId, &[f32])> = test_data
                .iter()
                .map(|(id, vec)| (*id, vec.as_slice()))
                .collect();
            storage.write_batch(&vectors).unwrap();
        }

        // Read vectors with second instance
        {
            let mut storage = MmapVectorStorage::open(&temp_dir, segment).unwrap();
            assert_eq!(storage.vector_count(), 2);

            let vec1 = storage.read_vector(VectorId::new(1).unwrap()).unwrap();
            assert_eq!(vec1, vec![1.0, 2.0]);

            let vec2 = storage.read_vector(VectorId::new(2).unwrap()).unwrap();
            assert_eq!(vec2, vec![3.0, 4.0]);
        }
    }

    #[test]
    fn test_vector_access_performance() {
        use std::time::Instant;

        let temp_dir = TempDir::new().unwrap();
        let segment = SegmentOrdinal::new(0);
        let dimension = VectorDimension::new(128).unwrap(); // Smaller dimension for faster test

        let mut storage = MmapVectorStorage::open_or_create(&temp_dir, segment, dimension).unwrap();

        // Write 1000 vectors
        let test_data: Vec<_> = (1..=1000)
            .map(|i| {
                let id = VectorId::new(i).unwrap();
                let vector = vec![i as f32 / 1000.0; 128];
                (id, vector)
            })
            .collect();
        let vectors: Vec<(VectorId, &[f32])> = test_data
            .iter()
            .map(|(id, vec)| (*id, vec.as_slice()))
            .collect();

        storage.write_batch(&vectors).unwrap();

        // Warm up the mmap
        for i in 1..=10 {
            let id = VectorId::new(i).unwrap();
            let _ = storage.read_vector(id);
        }

        // Measure read performance with pre-warmed cache
        let mut timings = Vec::with_capacity(1000);

        for i in 1..=1000 {
            let id = VectorId::new(i).unwrap();
            let start = Instant::now();
            let _ = storage.read_vector(id);
            let elapsed = start.elapsed();
            timings.push(elapsed.as_nanos());
        }

        // Sort timings and get median (more stable than average)
        timings.sort_unstable();
        let median_nanos = timings[timings.len() / 2];

        println!("Median read time: {median_nanos}ns");

        // In CI/test environments, we need to be more lenient
        // Real production performance will be much better
        assert!(
            median_nanos < 100_000,
            "Read performance should be <100μs in test environment"
        );
    }
}