Expand description
§llama-models
Foundational model blocks for llama.rs:
- RMSNorm
- RoPE
- Attention (scaled dot-product, single-step decode form)
- MLP (SwiGLU)
- Safetensors weight loading
Structs§
- Llama
Block - Minimal Llama block composition.
- Model
Weights - Named weight storage loaded from safetensors.
- Qwen
Block - Minimal Qwen block composition (same block primitives at this stage).
- Tensor
- Lightweight tensor holder for loaded model weights.
Enums§
- Model
Error - Errors for model operations and weight loading.
Functions§
- apply_
rope - Apply rotary positional embeddings in-place to query and key vectors.
- attention_
decode - Single-step decode attention:
- mlp_
swiglu - SwiGLU MLP:
- rms_
norm - Root mean square normalization.