foxstash-core 0.5.0

High-performance local RAG library - SIMD-accelerated vector search, HNSW indexing
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
//! Cached embedder with LRU caching for improved performance
//!
//! This module provides a caching layer on top of the ONNX embedder to avoid
//! redundant embedding computations for frequently used texts.

#[cfg(feature = "onnx")]
use lru::LruCache;
#[cfg(feature = "onnx")]
use std::num::NonZeroUsize;
#[cfg(feature = "onnx")]
use std::sync::Mutex;

use crate::{Result, RagError};
use super::OnnxEmbedder;

/// A cached embedder that wraps an ONNX embedder with an LRU cache
///
/// The cache stores embeddings keyed by the input text, significantly improving
/// performance when the same texts are embedded multiple times. This is common
/// in RAG scenarios where queries may be repeated or similar.
///
/// # Thread Safety
///
/// The cache is protected by a `Mutex`, making the `CachedEmbedder` safe to use
/// across multiple threads. However, contention may occur under heavy concurrent
/// access patterns.
///
/// # Memory Usage
///
/// Each cache entry stores the input text as a `String` and the embedding as a
/// `Vec<f32>`. For a typical 384-dimensional embedding, this is approximately:
/// - Key: text length in bytes
/// - Value: 384 * 4 bytes = 1.5KB
///
/// With the default cache size of 10,000 entries, expect ~15MB for embeddings
/// plus the overhead of stored text.
///
/// # Example
///
/// ```ignore
/// use foxstash_core::embedding::{OnnxEmbedder, CachedEmbedder};
///
/// let embedder = OnnxEmbedder::new("model.onnx", "tokenizer.json")?;
/// let cached = CachedEmbedder::new(embedder, 10_000);
///
/// // First call will compute and cache
/// let embedding1 = cached.embed("Hello, world!")?;
///
/// // Second call will use cached result
/// let embedding2 = cached.embed("Hello, world!")?;
///
/// // Check cache statistics
/// let (hits, misses) = cached.cache_stats();
/// assert_eq!(hits, 1);
/// assert_eq!(misses, 1);
/// ```
#[cfg(feature = "onnx")]
#[derive(Debug)]
pub struct CachedEmbedder {
    /// The underlying ONNX embedder (mutable for inference)
    embedder: Mutex<OnnxEmbedder>,
    /// LRU cache storing text -> embedding mappings
    cache: Mutex<LruCache<String, Vec<f32>>>,
    /// Cache hit counter
    hits: Mutex<usize>,
    /// Cache miss counter
    misses: Mutex<usize>,
}

#[cfg(feature = "onnx")]
impl CachedEmbedder {
    /// Creates a new cached embedder
    ///
    /// # Arguments
    ///
    /// * `embedder` - The underlying ONNX embedder to use for computing embeddings
    /// * `max_cache_size` - Maximum number of embeddings to cache (LRU eviction)
    ///
    /// # Example
    ///
    /// ```ignore
    /// let embedder = OnnxEmbedder::new("model.onnx", "tokenizer.json")?;
    /// let cached = CachedEmbedder::new(embedder, 10_000);
    /// ```
    pub fn new(embedder: OnnxEmbedder, max_cache_size: usize) -> Self {
        let cache_size = NonZeroUsize::new(max_cache_size)
            .unwrap_or_else(|| NonZeroUsize::new(1).unwrap());

        Self {
            embedder: Mutex::new(embedder),
            cache: Mutex::new(LruCache::new(cache_size)),
            hits: Mutex::new(0),
            misses: Mutex::new(0),
        }
    }

    /// Creates a new cached embedder with default cache size (10,000 entries)
    ///
    /// # Arguments
    ///
    /// * `embedder` - The underlying ONNX embedder to use
    ///
    /// # Example
    ///
    /// ```ignore
    /// let embedder = OnnxEmbedder::new("model.onnx", "tokenizer.json")?;
    /// let cached = CachedEmbedder::with_default_size(embedder);
    /// ```
    pub fn with_default_size(embedder: OnnxEmbedder) -> Self {
        Self::new(embedder, 10_000)
    }

    /// Embeds a single text, using cache if available
    ///
    /// This method first checks the cache for the given text. If found, it returns
    /// the cached embedding and increments the hit counter. Otherwise, it computes
    /// the embedding using the underlying embedder, caches it, and increments the
    /// miss counter.
    ///
    /// # Arguments
    ///
    /// * `text` - The text to embed
    ///
    /// # Returns
    ///
    /// A vector of floats representing the embedding
    ///
    /// # Errors
    ///
    /// Returns an error if the underlying embedder fails to compute the embedding
    ///
    /// # Example
    ///
    /// ```ignore
    /// let embedding = cached.embed("Hello, world!")?;
    /// assert_eq!(embedding.len(), 384); // For MiniLM-L6-v2
    /// ```
    pub fn embed(&self, text: &str) -> Result<Vec<f32>> {
        // Try to get from cache first
        {
            let mut cache = self.cache.lock()
                .map_err(|e| RagError::EmbeddingError(format!("Cache lock poisoned: {}", e)))?;

            if let Some(embedding) = cache.get(text) {
                // Cache hit
                let mut hits = self.hits.lock()
                    .map_err(|e| RagError::EmbeddingError(format!("Hits lock poisoned: {}", e)))?;
                *hits += 1;
                return Ok(embedding.clone());
            }
        }

        // Cache miss - compute embedding
        let embedding = {
            let mut embedder = self.embedder.lock()
                .map_err(|e| RagError::EmbeddingError(format!("Embedder lock poisoned: {}", e)))?;
            embedder.embed(text)?
        };

        // Store in cache
        {
            let mut cache = self.cache.lock()
                .map_err(|e| RagError::EmbeddingError(format!("Cache lock poisoned: {}", e)))?;
            cache.put(text.to_string(), embedding.clone());

            let mut misses = self.misses.lock()
                .map_err(|e| RagError::EmbeddingError(format!("Misses lock poisoned: {}", e)))?;
            *misses += 1;
        }

        Ok(embedding)
    }

    /// Embeds multiple texts in batch, utilizing cache for individual texts
    ///
    /// This method processes each text independently, checking the cache first
    /// before computing embeddings. Texts are batched together for efficient
    /// computation when cache misses occur.
    ///
    /// # Arguments
    ///
    /// * `texts` - Slice of text references to embed
    ///
    /// # Returns
    ///
    /// A vector of embeddings, one for each input text in the same order
    ///
    /// # Errors
    ///
    /// Returns an error if the underlying embedder fails
    ///
    /// # Example
    ///
    /// ```ignore
    /// let texts = ["Hello", "World", "RAG"];
    /// let embeddings = cached.embed_batch(&texts)?;
    /// assert_eq!(embeddings.len(), 3);
    /// ```
    pub fn embed_batch(&self, texts: &[&str]) -> Result<Vec<Vec<f32>>> {
        let mut results = Vec::with_capacity(texts.len());
        let mut to_compute = Vec::new();
        let mut to_compute_indices = Vec::new();

        // Check cache for each text
        {
            let mut cache = self.cache.lock()
                .map_err(|e| RagError::EmbeddingError(format!("Cache lock poisoned: {}", e)))?;
            let mut hits = self.hits.lock()
                .map_err(|e| RagError::EmbeddingError(format!("Hits lock poisoned: {}", e)))?;

            for (idx, &text) in texts.iter().enumerate() {
                if let Some(embedding) = cache.get(text) {
                    // Cache hit - store result at correct index
                    results.push((idx, embedding.clone()));
                    *hits += 1;
                } else {
                    // Cache miss - mark for computation
                    to_compute.push(text);
                    to_compute_indices.push(idx);
                }
            }
        }

        // Compute embeddings for cache misses
        if !to_compute.is_empty() {
            let computed = {
                let mut embedder = self.embedder.lock()
                    .map_err(|e| RagError::EmbeddingError(format!("Embedder lock poisoned: {}", e)))?;
                embedder.embed_batch(&to_compute)?
            };

            // Cache the newly computed embeddings
            {
                let mut cache = self.cache.lock()
                    .map_err(|e| RagError::EmbeddingError(format!("Cache lock poisoned: {}", e)))?;
                let mut misses = self.misses.lock()
                    .map_err(|e| RagError::EmbeddingError(format!("Misses lock poisoned: {}", e)))?;

                for (text, embedding) in to_compute.iter().zip(computed.iter()) {
                    cache.put(text.to_string(), embedding.clone());
                    *misses += 1;
                }
            }

            // Add computed results
            for (idx, embedding) in to_compute_indices.into_iter().zip(computed.into_iter()) {
                results.push((idx, embedding));
            }
        }

        // Sort by original index and extract embeddings
        results.sort_by_key(|(idx, _)| *idx);
        Ok(results.into_iter().map(|(_, emb)| emb).collect())
    }

    /// Clears all cached embeddings
    ///
    /// This method removes all entries from the cache and resets the hit/miss
    /// counters. Use this to free memory or reset statistics.
    ///
    /// # Example
    ///
    /// ```ignore
    /// cached.clear_cache();
    /// assert_eq!(cached.cache_size(), 0);
    /// ```
    pub fn clear_cache(&self) {
        if let Ok(mut cache) = self.cache.lock() {
            cache.clear();
        }
        if let Ok(mut hits) = self.hits.lock() {
            *hits = 0;
        }
        if let Ok(mut misses) = self.misses.lock() {
            *misses = 0;
        }
    }

    /// Returns the current number of entries in the cache
    ///
    /// # Example
    ///
    /// ```ignore
    /// let size = cached.cache_size();
    /// println!("Cache contains {} embeddings", size);
    /// ```
    pub fn cache_size(&self) -> usize {
        self.cache.lock()
            .map(|cache| cache.len())
            .unwrap_or(0)
    }

    /// Returns cache hit and miss statistics
    ///
    /// # Returns
    ///
    /// A tuple of `(hits, misses)` where:
    /// - `hits`: Number of times an embedding was found in cache
    /// - `misses`: Number of times an embedding had to be computed
    ///
    /// # Example
    ///
    /// ```ignore
    /// let (hits, misses) = cached.cache_stats();
    /// let hit_rate = hits as f64 / (hits + misses) as f64;
    /// println!("Cache hit rate: {:.2}%", hit_rate * 100.0);
    /// ```
    pub fn cache_stats(&self) -> (usize, usize) {
        let hits = self.hits.lock().map(|h| *h).unwrap_or(0);
        let misses = self.misses.lock().map(|m| *m).unwrap_or(0);
        (hits, misses)
    }

    /// Returns the maximum cache size
    ///
    /// This is the capacity configured at creation time, not the current
    /// number of entries (use `cache_size()` for that).
    ///
    /// # Example
    ///
    /// ```ignore
    /// let max_size = cached.max_cache_size();
    /// println!("Cache capacity: {}", max_size);
    /// ```
    pub fn max_cache_size(&self) -> usize {
        self.cache.lock()
            .map(|cache| cache.cap().get())
            .unwrap_or(0)
    }

    /// Returns the cache hit rate as a percentage
    ///
    /// # Returns
    ///
    /// The hit rate as a value between 0.0 and 1.0, or None if no operations
    /// have been performed yet.
    ///
    /// # Example
    ///
    /// ```ignore
    /// if let Some(rate) = cached.hit_rate() {
    ///     println!("Hit rate: {:.2}%", rate * 100.0);
    /// }
    /// ```
    pub fn hit_rate(&self) -> Option<f64> {
        let (hits, misses) = self.cache_stats();
        let total = hits + misses;

        if total == 0 {
            None
        } else {
            Some(hits as f64 / total as f64)
        }
    }
}

#[cfg(all(test, feature = "onnx"))]
mod tests {
    use super::*;

    // Tests for CachedEmbedder functionality
    // Most tests are marked #[ignore] because they require actual ONNX model files.
    // Run with: cargo test -- --ignored --test-threads=1

    /// Test cache creation and basic properties
    #[test]
    #[ignore]
    fn test_cache_creation() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        // Test with custom size
        let cached = CachedEmbedder::new(embedder, 42);
        assert_eq!(cached.max_cache_size(), 42);
        assert_eq!(cached.cache_size(), 0);
        assert_eq!(cached.cache_stats(), (0, 0));
        assert_eq!(cached.hit_rate(), None);
    }

    /// Test default cache size
    #[test]
    #[ignore]
    fn test_with_default_size() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::with_default_size(embedder);
        assert_eq!(cached.max_cache_size(), 10_000);
    }

    /// Test that zero cache size is handled (converted to 1)
    #[test]
    #[ignore]
    fn test_zero_cache_size() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 0);
        assert_eq!(cached.max_cache_size(), 1);
    }

    /// Test cache hit - same text embedded twice
    #[test]
    #[ignore]
    fn test_cache_hit() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        // First call - cache miss
        let embedding1 = cached.embed("Hello, world!").unwrap();
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 0);
        assert_eq!(misses, 1);
        assert_eq!(cached.cache_size(), 1);

        // Second call - cache hit
        let embedding2 = cached.embed("Hello, world!").unwrap();
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 1);
        assert_eq!(misses, 1);
        assert_eq!(cached.cache_size(), 1);

        // Embeddings should be identical
        assert_eq!(embedding1, embedding2);
        assert_eq!(cached.hit_rate(), Some(0.5));
    }

    /// Test cache misses - different texts
    #[test]
    #[ignore]
    fn test_cache_miss() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        // Different texts should all be cache misses
        let emb1 = cached.embed("Hello").unwrap();
        let emb2 = cached.embed("World").unwrap();
        let emb3 = cached.embed("RAG").unwrap();

        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 0);
        assert_eq!(misses, 3);
        assert_eq!(cached.cache_size(), 3);
        assert_eq!(cached.hit_rate(), Some(0.0));

        // Embeddings should be different
        assert_ne!(emb1, emb2);
        assert_ne!(emb2, emb3);
    }

    /// Test cache eviction with LRU policy
    #[test]
    #[ignore]
    fn test_cache_eviction() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 3); // Small cache

        // Fill cache to capacity
        let _emb1 = cached.embed("text1").unwrap();
        let _emb2 = cached.embed("text2").unwrap();
        let _emb3 = cached.embed("text3").unwrap();
        assert_eq!(cached.cache_size(), 3);

        // Add one more - should evict oldest (text1) via LRU
        let _emb4 = cached.embed("text4").unwrap();
        assert_eq!(cached.cache_size(), 3);

        // Access text2 and text3 (hits)
        let _ = cached.embed("text2").unwrap();
        let _ = cached.embed("text3").unwrap();
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 2);
        assert_eq!(misses, 4);

        // text1 was evicted, so accessing it should be a miss
        let _ = cached.embed("text1").unwrap();
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 2);
        assert_eq!(misses, 5);
    }

    /// Test clearing the cache
    #[test]
    #[ignore]
    fn test_clear_cache() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        // Add some embeddings
        let _emb1 = cached.embed("Hello").unwrap();
        let _emb2 = cached.embed("World").unwrap();
        assert_eq!(cached.cache_size(), 2);
        assert_eq!(cached.cache_stats(), (0, 2));

        // Clear cache
        cached.clear_cache();
        assert_eq!(cached.cache_size(), 0);
        assert_eq!(cached.cache_stats(), (0, 0));
        assert_eq!(cached.hit_rate(), None);

        // Embeddings should be computed again
        let _emb1_again = cached.embed("Hello").unwrap();
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 0);
        assert_eq!(misses, 1);
    }

    /// Test cache statistics tracking
    #[test]
    #[ignore]
    fn test_cache_statistics() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        // Initial state
        assert_eq!(cached.cache_stats(), (0, 0));
        assert_eq!(cached.hit_rate(), None);

        // First embedding - miss
        let _emb = cached.embed("test").unwrap();
        assert_eq!(cached.cache_stats(), (0, 1));
        assert_eq!(cached.hit_rate(), Some(0.0));

        // Second embedding - hit
        let _emb = cached.embed("test").unwrap();
        assert_eq!(cached.cache_stats(), (1, 1));
        assert_eq!(cached.hit_rate(), Some(0.5));

        // Third embedding - hit
        let _emb = cached.embed("test").unwrap();
        assert_eq!(cached.cache_stats(), (2, 1));
        assert!((cached.hit_rate().unwrap() - 0.6666).abs() < 0.01);
    }

    /// Test batch embedding with all cached entries
    #[test]
    #[ignore]
    fn test_batch_embed_all_cached() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);
        let texts = ["Hello", "World", "RAG"];

        // First batch - all misses
        let embeddings1 = cached.embed_batch(&texts).unwrap();
        assert_eq!(embeddings1.len(), 3);
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 0);
        assert_eq!(misses, 3);

        // Second batch - all hits
        let embeddings2 = cached.embed_batch(&texts).unwrap();
        assert_eq!(embeddings2.len(), 3);
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 3);
        assert_eq!(misses, 3);

        // Embeddings should be identical
        assert_eq!(embeddings1, embeddings2);
    }

    /// Test batch embedding with mixed cache hits and misses
    #[test]
    #[ignore]
    fn test_batch_embed_mixed() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        // Pre-cache some texts
        let _emb1 = cached.embed("Hello").unwrap();
        let _emb2 = cached.embed("World").unwrap();
        let initial_stats = cached.cache_stats();
        assert_eq!(initial_stats, (0, 2));

        // Batch with mix of cached and uncached
        let texts = ["Hello", "World", "RAG", "System"];
        let embeddings = cached.embed_batch(&texts).unwrap();
        assert_eq!(embeddings.len(), 4);

        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits, 2); // Hello and World
        assert_eq!(misses, 4); // Initial 2 + RAG + System
    }

    /// Test that batch embedding preserves input order
    #[test]
    #[ignore]
    fn test_batch_embed_order_preserved() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        // Pre-cache in different order
        let emb_rag = cached.embed("RAG").unwrap();
        let emb_hello = cached.embed("Hello").unwrap();

        // Batch request in different order
        let texts = ["Hello", "World", "RAG"];
        let embeddings = cached.embed_batch(&texts).unwrap();

        // Verify order is preserved
        assert_eq!(embeddings[0], emb_hello);
        assert_eq!(embeddings[2], emb_rag);

        // World should be computed
        let emb_world = cached.embed("World").unwrap();
        assert_eq!(embeddings[1], emb_world);
    }

    /// Test concurrent access to the cache
    #[test]
    #[ignore]
    fn test_concurrent_access() {
        use std::sync::Arc;
        use std::thread;

        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = Arc::new(CachedEmbedder::new(embedder, 100));
        let mut handles = vec![];

        // Spawn multiple threads accessing the cache
        for i in 0..10 {
            let cached_clone = Arc::clone(&cached);
            let handle = thread::spawn(move || {
                for j in 0..10 {
                    let text = format!("text_{}", (i + j) % 5);
                    let _ = cached_clone.embed(&text).unwrap();
                }
            });
            handles.push(handle);
        }

        // Wait for all threads
        for handle in handles {
            handle.join().unwrap();
        }

        // Verify cache has entries and all operations completed
        assert!(cached.cache_size() > 0);
        let (hits, misses) = cached.cache_stats();
        assert_eq!(hits + misses, 100); // 10 threads * 10 operations

        // Since we're using modulo 5, there should be many cache hits
        assert!(hits > 0, "Expected cache hits with repeated texts");
    }

    /// Test embedding dimension matches
    #[test]
    #[ignore]
    fn test_embedding_dimension() {
        let embedder = OnnxEmbedder::new(
            "models/model.onnx",
            "models/tokenizer.json"
        ).expect("Failed to create embedder");

        let cached = CachedEmbedder::new(embedder, 100);

        let embedding = cached.embed("Test text").unwrap();
        assert_eq!(embedding.len(), 384, "MiniLM-L6-v2 should produce 384-dim embeddings");

        // Verify normalization
        let norm: f32 = embedding.iter().map(|&x| x * x).sum::<f32>().sqrt();
        assert!((norm - 1.0).abs() < 1e-5, "Embedding should be L2-normalized");
    }
}