llama-gguf 0.14.0

A high-performance Rust implementation of llama.cpp - LLM inference engine with full GGUF support
Documentation
//! Debug forward pass layer by layer

use llama_gguf::backend::Backend;
use llama_gguf::backend::cpu::CpuBackend;
use llama_gguf::gguf::GgufFile;
use llama_gguf::tensor::{DType, Tensor};
use std::path::Path;

fn load_tensor(gguf: &GgufFile, name: &str) -> Tensor {
    let info = gguf.data.get_tensor(name).unwrap();
    let data = gguf.tensor_data(name).unwrap();
    let shape: Vec<usize> = info.dims.iter().map(|&d| d as usize).collect();
    Tensor::new(data.to_vec(), shape, DType::from(info.dtype)).unwrap()
}

fn try_load_tensor(gguf: &GgufFile, name: &str) -> Option<Tensor> {
    let info = gguf.data.get_tensor(name)?;
    let data = gguf.tensor_data(name)?;
    let shape: Vec<usize> = info.dims.iter().map(|&d| d as usize).collect();
    Tensor::new(data.to_vec(), shape, DType::from(info.dtype)).ok()
}

fn dequant(backend: &CpuBackend, t: &Tensor) -> Vec<f32> {
    if t.dtype() == DType::F32 {
        t.as_f32().unwrap().to_vec()
    } else {
        let mut out = Tensor::zeros(vec![t.numel()], DType::F32);
        backend.dequantize(t, &mut out).unwrap();
        out.as_f32().unwrap().to_vec()
    }
}

fn rms_norm(x: &[f32], w: &[f32], eps: f32) -> Vec<f32> {
    let sum_sq: f32 = x.iter().map(|v| v * v).sum();
    let rms = (sum_sq / x.len() as f32 + eps).sqrt();
    let inv_rms = 1.0 / rms;
    x.iter()
        .zip(w.iter())
        .map(|(v, wt)| v * inv_rms * wt)
        .collect()
}

fn vec_mat(x: &[f32], w: &[f32], k: usize, n: usize) -> Vec<f32> {
    let mut out = vec![0.0f32; n];
    for j in 0..n {
        for i in 0..k {
            out[j] += x[i] * w[i + j * k];
        }
    }
    out
}

fn apply_rope_neox(data: &mut [f32], pos: usize, head_dim: usize, freq_base: f32) {
    let half_dim = head_dim / 2;
    for i in 0..half_dim {
        let freq = 1.0 / freq_base.powf((2 * i) as f32 / head_dim as f32);
        let theta = pos as f32 * freq;
        let (sin_t, cos_t) = theta.sin_cos();
        let x0 = data[i];
        let x1 = data[i + half_dim];
        data[i] = x0 * cos_t - x1 * sin_t;
        data[i + half_dim] = x0 * sin_t + x1 * cos_t;
    }
}

fn softmax(scores: &mut [f32]) {
    let max_score = scores.iter().cloned().fold(f32::NEG_INFINITY, f32::max);
    let mut sum = 0.0f32;
    for s in scores.iter_mut() {
        *s = (*s - max_score).exp();
        sum += *s;
    }
    for s in scores.iter_mut() {
        *s /= sum;
    }
}

fn silu(x: f32) -> f32 {
    x / (1.0 + (-x).exp())
}

fn stats(data: &[f32]) -> (f32, f32, f32, f32) {
    let min = data.iter().cloned().fold(f32::INFINITY, f32::min);
    let max = data.iter().cloned().fold(f32::NEG_INFINITY, f32::max);
    let mean = data.iter().sum::<f32>() / data.len() as f32;
    let std = (data.iter().map(|x| (x - mean).powi(2)).sum::<f32>() / data.len() as f32).sqrt();
    (min, max, mean, std)
}

fn main() {
    let model_path = "/home/joseph/Models/qwen2.5-0.5b-instruct-q4_k_m.gguf";

    eprintln!("Loading model...");
    let gguf = GgufFile::open(Path::new(model_path)).expect("Failed to open GGUF");
    let backend = CpuBackend::new();

    let emb = dequant(&backend, &load_tensor(&gguf, "token_embd.weight"));
    let output_norm_w = dequant(&backend, &load_tensor(&gguf, "output_norm.weight"));
    let output_w = dequant(&backend, &load_tensor(&gguf, "output.weight"));

    let hidden_size = 896;
    let num_heads = 14;
    let head_dim = 64;
    let num_kv_heads = 2;
    let intermediate_size = 4864;
    let scale = 1.0 / (head_dim as f32).sqrt();
    let queries_per_kv = num_heads / num_kv_heads;
    let max_seq_len = 512;
    let n_layers = 24;
    let eps = 1e-6f32;
    let freq_base = 1000000.0f32;

    // Just process single token "=" (28) first
    let token = 28u32;
    println!("Processing single token: {} ('=')", token);

    // Get embedding
    let mut hidden = emb[token as usize * hidden_size..(token as usize + 1) * hidden_size].to_vec();
    let (min, max, mean, std) = stats(&hidden);
    println!(
        "Initial embedding: min={:.4}, max={:.4}, mean={:.4}, std={:.4}",
        min, max, mean, std
    );

    // KV caches - for single token, pos=0
    let mut k_caches: Vec<Vec<Vec<f32>>> = vec![vec![vec![0.0; head_dim]; num_kv_heads]; n_layers];
    let mut v_caches: Vec<Vec<Vec<f32>>> = vec![vec![vec![0.0; head_dim]; num_kv_heads]; n_layers];

    for layer in 0..n_layers {
        let prefix = format!("blk.{}", layer);

        let attn_norm_w = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.attn_norm.weight", prefix)),
        );
        let wq = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.attn_q.weight", prefix)),
        );
        let wk = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.attn_k.weight", prefix)),
        );
        let wv = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.attn_v.weight", prefix)),
        );
        let wo = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.attn_output.weight", prefix)),
        );

        let q_bias = try_load_tensor(&gguf, &format!("{}.attn_q.bias", prefix))
            .map(|t| dequant(&backend, &t));
        let k_bias = try_load_tensor(&gguf, &format!("{}.attn_k.bias", prefix))
            .map(|t| dequant(&backend, &t));
        let v_bias = try_load_tensor(&gguf, &format!("{}.attn_v.bias", prefix))
            .map(|t| dequant(&backend, &t));

        let ffn_norm_w = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.ffn_norm.weight", prefix)),
        );
        let w_gate = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.ffn_gate.weight", prefix)),
        );
        let w_up = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.ffn_up.weight", prefix)),
        );
        let w_down = dequant(
            &backend,
            &load_tensor(&gguf, &format!("{}.ffn_down.weight", prefix)),
        );

        // Attention
        let normed = rms_norm(&hidden, &attn_norm_w, eps);
        let mut q = vec_mat(&normed, &wq, hidden_size, num_heads * head_dim);
        let mut k = vec_mat(&normed, &wk, hidden_size, num_kv_heads * head_dim);
        let mut v = vec_mat(&normed, &wv, hidden_size, num_kv_heads * head_dim);

        // Biases BEFORE RoPE
        if let Some(ref bias) = q_bias {
            for (qi, bi) in q.iter_mut().zip(bias.iter()) {
                *qi += *bi;
            }
        }
        if let Some(ref bias) = k_bias {
            for (ki, bi) in k.iter_mut().zip(bias.iter()) {
                *ki += *bi;
            }
        }
        if let Some(ref bias) = v_bias {
            for (vi, bi) in v.iter_mut().zip(bias.iter()) {
                *vi += *bi;
            }
        }

        // Apply RoPE (position 0)
        for head in 0..num_heads {
            apply_rope_neox(
                &mut q[head * head_dim..(head + 1) * head_dim],
                0,
                head_dim,
                freq_base,
            );
        }
        for kv_head in 0..num_kv_heads {
            apply_rope_neox(
                &mut k[kv_head * head_dim..(kv_head + 1) * head_dim],
                0,
                head_dim,
                freq_base,
            );
        }

        // Store K/V
        for kv_head in 0..num_kv_heads {
            k_caches[layer][kv_head] = k[kv_head * head_dim..(kv_head + 1) * head_dim].to_vec();
            v_caches[layer][kv_head] = v[kv_head * head_dim..(kv_head + 1) * head_dim].to_vec();
        }

        // Self-attention with single position (attention is just copying)
        let mut attn_out = vec![0.0f32; num_heads * head_dim];
        for head in 0..num_heads {
            let kv_head = head / queries_per_kv;
            let q_vec = &q[head * head_dim..(head + 1) * head_dim];
            let k_vec = &k_caches[layer][kv_head];

            // Score (single position so softmax is 1.0)
            let dot: f32 = q_vec.iter().zip(k_vec.iter()).map(|(a, b)| a * b).sum();
            let score = dot * scale; // For single position, softmax(score) = 1.0

            // Copy value
            let v_vec = &v_caches[layer][kv_head];
            for d in 0..head_dim {
                attn_out[head * head_dim + d] = v_vec[d]; // weight is 1.0
            }
        }

        let attn_proj = vec_mat(&attn_out, &wo, num_heads * head_dim, hidden_size);
        let h: Vec<f32> = hidden
            .iter()
            .zip(attn_proj.iter())
            .map(|(a, b)| a + b)
            .collect();

        // FFN
        let ffn_normed = rms_norm(&h, &ffn_norm_w, eps);
        let gate = vec_mat(&ffn_normed, &w_gate, hidden_size, intermediate_size);
        let up = vec_mat(&ffn_normed, &w_up, hidden_size, intermediate_size);
        let intermediate: Vec<f32> = gate
            .iter()
            .zip(up.iter())
            .map(|(g, u)| silu(*g) * u)
            .collect();
        let ffn_out = vec_mat(&intermediate, &w_down, intermediate_size, hidden_size);

        hidden = h.iter().zip(ffn_out.iter()).map(|(a, b)| a + b).collect();

        if layer == 0 || layer == 1 || layer == 11 || layer == 23 {
            let (min, max, mean, std) = stats(&hidden);
            println!(
                "After layer {:2}: min={:.4}, max={:.4}, mean={:.4}, std={:.4}",
                layer, min, max, mean, std
            );
        }
    }

    // Final norm and logits
    let normed_final = rms_norm(&hidden, &output_norm_w, eps);
    let (min, max, mean, std) = stats(&normed_final);
    println!(
        "After final norm: min={:.4}, max={:.4}, mean={:.4}, std={:.4}",
        min, max, mean, std
    );

    let logits = vec_mat(&normed_final, &output_w, hidden_size, 151936);
    let (min, max, mean, std) = stats(&logits);
    println!(
        "Logits: min={:.4}, max={:.4}, mean={:.4}, std={:.4}",
        min, max, mean, std
    );

    // Show top predictions
    let mut indexed: Vec<(usize, f32)> = logits.iter().cloned().enumerate().collect();
    indexed.sort_by(|a, b| b.1.partial_cmp(&a.1).unwrap());

    println!("\nTop 5 predictions:");
    for (idx, logit) in indexed.iter().take(5) {
        println!("  Token {}: {:.4}", idx, logit);
    }

    println!("\nNumeric tokens [16-20]:");
    for i in 16..21 {
        let rank = indexed.iter().position(|(idx, _)| *idx == i).unwrap() + 1;
        println!(
            "  Token {} ('{}'), rank {}: {:.4}",
            i,
            i - 15,
            rank,
            logits[i]
        );
    }
}