trustformers-optim 0.1.1

Optimizers for TrustformeRS
Documentation
#![allow(clippy::result_large_err)]
use std::collections::HashMap;
use std::time::Instant;
use trustformers_core::TrustformersError;
use trustformers_core::{traits::Optimizer, Tensor};
use trustformers_optim::*;

fn main() -> Result<(), TrustformersError> {
    println!("🚀 TrustformeRS Memory Usage Analysis");
    println!("====================================");
    println!("🔬 Detailed memory usage comparison: 32-bit vs 8-bit optimizers");

    // Test different parameter sizes to validate memory efficiency
    let param_sizes = vec![1000, 10000, 50000];

    for param_size in param_sizes {
        println!("\n🎯 Analyzing {} parameters", param_size);
        println!("{}", "".repeat(60));

        // Calculate theoretical memory usage
        let f32_bytes = 4; // 32-bit float
        let i8_bytes = 1; // 8-bit quantized
        let param_memory = param_size * f32_bytes;

        println!(
            "📊 Parameter memory: {} bytes ({:.2} KB)",
            param_memory,
            param_memory as f64 / 1024.0
        );

        // Create test data
        let gradients = Tensor::randn(&[param_size])?;

        // Test regular Adam optimizer memory usage
        println!("\n📊 Regular Adam (32-bit) Memory Analysis:");
        let mut adam32 = Adam::new(0.001, (0.9, 0.999), 1e-8, 0.0);
        let mut params_adam32 = Tensor::randn(&[param_size])?;

        // Run some steps to initialize optimizer state
        let start = Instant::now();
        for _ in 0..10 {
            adam32.update(&mut params_adam32, &gradients)?;
            adam32.step();
        }
        let adam32_time = start.elapsed();

        // Estimate memory usage for Adam (momentum + variance + parameters)
        let adam32_state_memory = param_size * f32_bytes * 2; // momentum + variance
        let adam32_total_memory = param_memory + adam32_state_memory;

        println!("   📦 Parameters: {} bytes", param_memory);
        println!("   🧠 Momentum state: {} bytes", param_size * f32_bytes);
        println!("   📈 Variance state: {} bytes", param_size * f32_bytes);
        println!(
            "   💾 Total memory: {} bytes ({:.2} KB)",
            adam32_total_memory,
            adam32_total_memory as f64 / 1024.0
        );
        println!("   ⏱️  10 steps: {:.2?}", adam32_time);

        // Test 8-bit Adam optimizer memory usage
        println!("\n📊 8-bit Adam Memory Analysis:");
        let mut adam8 = Adam8bit::with_config(0.001, 0.9, 0.999, 1e-8, 0.0);

        let mut params8_map = HashMap::new();
        let mut gradients_map = HashMap::new();
        params8_map.insert("param".to_string(), Tensor::randn(&[param_size])?);
        gradients_map.insert("param".to_string(), gradients.clone());

        let start = Instant::now();
        for _ in 0..10 {
            adam8.step(&mut params8_map, &gradients_map)?;
        }
        let adam8_time = start.elapsed();

        // Calculate memory usage for 8-bit Adam
        let quantization_overhead = 8; // scale + zero_point + min/max values per state
        let adam8_momentum_memory = param_size * i8_bytes + quantization_overhead;
        let adam8_variance_memory = param_size * i8_bytes + quantization_overhead;
        let adam8_total_memory = param_memory + adam8_momentum_memory + adam8_variance_memory;

        println!("   📦 Parameters: {} bytes", param_memory);
        println!(
            "   🧠 Momentum state (8-bit): {} bytes + {} overhead",
            param_size * i8_bytes,
            quantization_overhead
        );
        println!(
            "   📈 Variance state (8-bit): {} bytes + {} overhead",
            param_size * i8_bytes,
            quantization_overhead
        );
        println!(
            "   💾 Total memory: {} bytes ({:.2} KB)",
            adam8_total_memory,
            adam8_total_memory as f64 / 1024.0
        );
        println!("   ⏱️  10 steps: {:.2?}", adam8_time);

        // Calculate memory reduction
        let state_memory_32bit = adam32_state_memory;
        let state_memory_8bit = adam8_momentum_memory + adam8_variance_memory;
        let state_memory_reduction =
            (1.0 - (state_memory_8bit as f64 / state_memory_32bit as f64)) * 100.0;
        let total_memory_reduction =
            (1.0 - (adam8_total_memory as f64 / adam32_total_memory as f64)) * 100.0;

        println!("\n📈 Memory Efficiency Summary:");
        println!(
            "   🎯 State memory reduction: {:.1}% ({}{} bytes)",
            state_memory_reduction, state_memory_32bit, state_memory_8bit
        );
        println!(
            "   🎯 Total memory reduction: {:.1}% ({}{} bytes)",
            total_memory_reduction, adam32_total_memory, adam8_total_memory
        );

        if state_memory_reduction >= 70.0 {
            println!("   ✅ 8-bit Adam meets >70% state memory reduction target!");
        } else {
            println!(
                "   ⚠️  8-bit Adam: {:.1}% reduction (target: >70%)",
                state_memory_reduction
            );
        }

        // Performance impact
        let performance_ratio = if adam32_time < adam8_time {
            let slowdown = adam8_time.as_nanos() as f64 / adam32_time.as_nanos() as f64;
            println!(
                "   ⚡ Performance: 8-bit is {:.2}x slower ({:.2?} vs {:.2?})",
                slowdown, adam8_time, adam32_time
            );
            slowdown
        } else {
            let speedup = adam32_time.as_nanos() as f64 / adam8_time.as_nanos() as f64;
            println!("   🚀 Performance: 8-bit is {:.2}x faster", speedup);
            1.0 / speedup
        };

        // Calculate memory efficiency ratio
        let efficiency_ratio =
            state_memory_reduction / (performance_ratio.max(1.0) - 1.0).max(0.01) * 100.0;
        println!(
            "   🎯 Memory/Performance ratio: {:.1}% memory saved per 1% performance cost",
            efficiency_ratio
        );

        // Show expected memory savings for larger models
        println!("\n💡 Projected savings for larger models:");
        for scale in &[10, 100, 1000] {
            let scaled_size = param_size * scale;
            let scaled_32bit_memory = scaled_size * f32_bytes * 3; // params + momentum + variance
            let scaled_8bit_memory = scaled_size * f32_bytes + scaled_size * i8_bytes * 2 + 16; // params + 8bit states + overhead
            let scaled_reduction =
                (1.0 - (scaled_8bit_memory as f64 / scaled_32bit_memory as f64)) * 100.0;

            println!(
                "   📊 {}× size ({} params): {:.1}% memory reduction ({:.2} MB → {:.2} MB)",
                scale,
                scaled_size,
                scaled_reduction,
                scaled_32bit_memory as f64 / 1024.0 / 1024.0,
                scaled_8bit_memory as f64 / 1024.0 / 1024.0
            );
        }
    }

    println!("\n🎉 Memory Analysis Completed!");
    println!("   ✅ 8-bit optimizers provide significant memory savings");
    println!("   📊 State memory reduced by ~75% (8-bit vs 32-bit storage)");
    println!("   ⚡ Performance cost is acceptable for memory-constrained scenarios");
    println!("   💡 Larger models benefit proportionally more from 8-bit quantization");
    println!("   🎯 Memory efficiency claims validated through theoretical analysis");

    Ok(())
}