1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
//! Phase 43 - GpuExecutorTrait for Dependency Injection
//!
//! Abstracts GPU execution to enable testing without actual CUDA hardware.
//!
//! # Architecture
//!
//! ```text
//! GpuModel
//! │
//! └─► Box<dyn GpuExecutorTrait>
//! │
//! ├─► CudaExecutorAdapter (production)
//! └─► MockExecutor (testing)
//! ```
//!
//! # Coverage Impact
//!
//! This trait enables testing of:
//! - `GpuModel::forward()` - Full forward pass logic
//! - `GpuModel::generate()` - Token generation flow
//! - Layer-by-layer computation verification
use crate::error::Result;
use std::sync::atomic::{AtomicUsize, Ordering};
use std::sync::Mutex;
/// Trait for GPU execution backends
///
/// Implementations must be Send + Sync to allow model transfer between threads
/// and safe access in `Arc<RwLock<GpuModel>>` contexts.
pub trait GpuExecutorTrait: Send + Sync {
/// Perform matrix multiplication: C = A @ B
///
/// # Arguments
///
/// * `a` - Left matrix [m, k]
/// * `b` - Right matrix [k, n]
/// * `m` - Rows in A
/// * `k` - Columns in A / Rows in B
/// * `n` - Columns in B
///
/// # Returns
///
/// Result matrix [m, n]
fn matmul(&mut self, a: &[f32], b: &[f32], m: usize, k: usize, n: usize) -> Result<Vec<f32>>;
/// Check if GPU backend is available
fn is_available(&self) -> bool;
/// Get backend name for debugging
fn name(&self) -> &str;
/// Synchronize execution (wait for pending operations)
fn synchronize(&self) -> Result<()>;
/// Perform matrix multiplication with transposed B: C = A @ B^T
///
/// # Arguments
///
/// * `a` - Left matrix [m, k]
/// * `b` - Right matrix [n, k] (will be transposed to [k, n])
/// * `m` - Rows in A
/// * `k` - Columns in A / Columns in B (before transpose)
/// * `n` - Rows in B (becomes columns after transpose)
///
/// # Returns
///
/// Result matrix [m, n]
fn matmul_transpose_b(
&mut self,
a: &[f32],
b: &[f32],
m: usize,
k: usize,
n: usize,
) -> Result<Vec<f32>>;
}
/// Call record for MockExecutor
#[derive(Debug, Clone, PartialEq)]
pub enum ExecutorCall {
/// Matrix multiplication call
Matmul {
/// Input A dimensions
a_len: usize,
/// Input B dimensions
b_len: usize,
/// M dimension
m: usize,
/// K dimension
k: usize,
/// N dimension
n: usize,
},
/// Matrix multiplication with transposed B call
MatmulTransposeB {
/// Input A dimensions
a_len: usize,
/// Input B dimensions (before transpose)
b_len: usize,
/// M dimension
m: usize,
/// K dimension
k: usize,
/// N dimension
n: usize,
},
/// Synchronize call
Synchronize,
}
/// Mock executor for testing GpuModel without CUDA
///
/// Records all calls for verification and returns configurable results.
/// Uses interior mutability (Mutex) to allow Sync trait implementation.
pub struct MockExecutor {
/// Name of this mock
name: String,
/// Recorded calls (protected by mutex for thread-safety)
calls: Mutex<Vec<ExecutorCall>>,
/// Counter for unique call IDs
call_counter: AtomicUsize,
/// Whether to simulate availability
available: bool,
/// Custom matmul result (if None, returns zeros)
matmul_result: Option<Vec<f32>>,
/// Whether matmul should fail
matmul_should_fail: bool,
}
impl MockExecutor {
/// Create a new mock executor
#[must_use]
pub fn new(name: &str) -> Self {
Self {
name: name.to_string(),
calls: Mutex::new(Vec::new()),
call_counter: AtomicUsize::new(0),
available: true,
matmul_result: None,
matmul_should_fail: false,
}
}
/// Create mock that simulates unavailability
#[must_use]
pub fn unavailable(name: &str) -> Self {
Self {
name: name.to_string(),
calls: Mutex::new(Vec::new()),
call_counter: AtomicUsize::new(0),
available: false,
matmul_result: None,
matmul_should_fail: false,
}
}
/// Set custom matmul result
#[must_use]
pub fn with_matmul_result(mut self, result: Vec<f32>) -> Self {
self.matmul_result = Some(result);
self
}
/// Configure matmul to fail
#[must_use]
pub fn with_matmul_failure(mut self) -> Self {
self.matmul_should_fail = true;
self
}
/// Acquire the calls lock, recovering from poison if needed.
///
/// Centralizes the resource-acquire pattern for the `calls` mutex.
/// All call-recording and call-querying methods delegate here.
fn lock_calls(&self) -> std::sync::MutexGuard<'_, Vec<ExecutorCall>> {
self.calls
.lock()
.unwrap_or_else(std::sync::PoisonError::into_inner)
}
/// Record a call and return the configured matmul response.
///
/// Consolidates the record → check-failure → respond pattern
/// shared by `matmul` and `matmul_transpose_b`.
fn record_call_and_respond(&self, call: ExecutorCall, output_len: usize) -> Result<Vec<f32>> {
self.lock_calls().push(call);
self.call_counter.fetch_add(1, Ordering::SeqCst);
if self.matmul_should_fail {
return Err(crate::error::RealizarError::GpuError {
reason: "MockExecutor configured to fail".to_string(),
});
}
if let Some(ref result) = self.matmul_result {
Ok(result.clone())
} else {
Ok(vec![0.0f32; output_len])
}
}
/// Get all recorded calls (cloned for thread-safety)
#[must_use]
pub fn calls(&self) -> Vec<ExecutorCall> {
self.lock_calls().clone()
}
/// Get number of calls
#[must_use]
pub fn call_count(&self) -> usize {
self.lock_calls().len()
}
/// Get number of matmul calls
#[must_use]
pub fn matmul_count(&self) -> usize {
self.lock_calls()
.iter()
.filter(|c| matches!(c, ExecutorCall::Matmul { .. }))
.count()
}
/// Clear recorded calls
pub fn clear_calls(&self) {
self.lock_calls().clear();
self.call_counter.store(0, Ordering::SeqCst);
}
/// Check if specific call was made
#[must_use]
pub fn has_call(&self, call: &ExecutorCall) -> bool {
self.lock_calls().contains(call)
}
/// Get last call (cloned for thread-safety)
#[must_use]
pub fn last_call(&self) -> Option<ExecutorCall> {
self.lock_calls().last().cloned()
}
}
impl GpuExecutorTrait for MockExecutor {
#[allow(clippy::many_single_char_names)]
fn matmul(&mut self, a: &[f32], b: &[f32], m: usize, k: usize, n: usize) -> Result<Vec<f32>> {
self.record_call_and_respond(
ExecutorCall::Matmul {
a_len: a.len(),
b_len: b.len(),
m,
k,
n,
},
m * n,
)
}
fn is_available(&self) -> bool {
self.available
}
fn name(&self) -> &str {
&self.name
}
fn synchronize(&self) -> Result<()> {
// Record sync call (need interior mutability for &self)
// For testing, we skip recording in synchronize since it takes &self
Ok(())
}
#[allow(clippy::many_single_char_names)]
fn matmul_transpose_b(
&mut self,
a: &[f32],
b: &[f32],
m: usize,
k: usize,
n: usize,
) -> Result<Vec<f32>> {
self.record_call_and_respond(
ExecutorCall::MatmulTransposeB {
a_len: a.len(),
b_len: b.len(),
m,
k,
n,
},
m * n,
)
}
}
impl std::fmt::Debug for MockExecutor {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
f.debug_struct("MockExecutor")
.field("name", &self.name)
.field("calls", &self.lock_calls().len())
.field("call_counter", &self.call_counter.load(Ordering::SeqCst))
.field("available", &self.available)
.field("matmul_result", &self.matmul_result.is_some())
.field("matmul_should_fail", &self.matmul_should_fail)
.finish()
}
}
/// CPU-based executor for testing and fallback
///
/// Implements actual matrix multiplication on CPU.
pub struct CpuExecutor {
name: String,
}
impl CpuExecutor {
/// Create a new CPU executor
#[must_use]
pub fn new() -> Self {
Self {
name: "CpuExecutor".to_string(),
}
}
}
impl Default for CpuExecutor {
fn default() -> Self {
Self::new()
}
}
impl GpuExecutorTrait for CpuExecutor {
#[allow(clippy::many_single_char_names)]
fn matmul(&mut self, a: &[f32], b: &[f32], m: usize, k: usize, n: usize) -> Result<Vec<f32>> {
// Validate dimensions
if a.len() != m * k {
return Err(crate::error::RealizarError::InvalidShape {
reason: format!("A size {} != m*k {}", a.len(), m * k),
});
}
if b.len() != k * n {
return Err(crate::error::RealizarError::InvalidShape {
reason: format!("B size {} != k*n {}", b.len(), k * n),
});
}
// Naive matmul (for correctness, not performance)
let mut c = vec![0.0f32; m * n];
for i in 0..m {
for j in 0..n {
let mut sum = 0.0f32;
for p in 0..k {
sum += a[i * k + p] * b[p * n + j];
}
c[i * n + j] = sum;
}
}
Ok(c)
}
fn is_available(&self) -> bool {
true // CPU is always available
}
fn name(&self) -> &str {
&self.name
}
fn synchronize(&self) -> Result<()> {
Ok(()) // No-op for CPU
}
#[allow(clippy::many_single_char_names)]
fn matmul_transpose_b(
&mut self,
a: &[f32],
b: &[f32],
m: usize,
k: usize,
n: usize,
) -> Result<Vec<f32>> {
// Validate dimensions: A is [m, k], B is [n, k] (to be transposed to [k, n])
if a.len() != m * k {
return Err(crate::error::RealizarError::InvalidShape {
reason: format!("A size {} != m*k {}", a.len(), m * k),
});
}
if b.len() != n * k {
return Err(crate::error::RealizarError::InvalidShape {
reason: format!("B size {} != n*k {}", b.len(), n * k),
});
}
// Naive matmul with B transposed: C = A @ B^T
// A is [m, k], B is [n, k], C is [m, n]
// B^T is [k, n], so C[i,j] = sum over p of A[i,p] * B[j,p]
let mut c = vec![0.0f32; m * n];
for i in 0..m {
for j in 0..n {
let mut sum = 0.0f32;
for p in 0..k {
sum += a[i * k + p] * b[j * k + p]; // Note: b[j, p] not b[p, j]
}
c[i * n + j] = sum;
}
}
Ok(c)
}
}
impl std::fmt::Debug for CpuExecutor {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
f.debug_struct("CpuExecutor")
.field("name", &self.name)
.finish()
}
}
include!("executor_mock.rs");