tiny_recursive_rs/layers/
mod.rs

1/// Neural network layer primitives
2///
3/// This module contains the building blocks for the TRM model:
4/// - Attention mechanisms (multi-head, grouped-query)
5/// - Positional encodings (RoPE, learned)
6/// - Activations (SwiGLU, LinearSwish)
7/// - Normalizations (RMS LayerNorm)
8/// - Embeddings (with automatic dtype casting)
9
10pub mod attention;
11pub mod embeddings;
12pub mod normalization;
13pub mod activations;
14pub mod positional;
15
16pub use attention::Attention;
17pub use embeddings::CastedEmbedding;
18pub use normalization::RMSNorm;
19pub use activations::{SwiGLU, LinearSwish};
20pub use positional::RotaryEmbedding;