Expand description
axonml-llm - Large Language Model Architectures
§File
crates/axonml-llm/src/lib.rs
§Author
Andrew Jewell Sr - AutomataNexus
§Updated
March 8, 2026
§Disclaimer
Use at own risk. This software is provided “as is”, without warranty of any kind, express or implied. The author and AutomataNexus shall not be held liable for any damages arising from the use of this software.
Re-exports§
pub use attention::CausalSelfAttention;pub use attention::FlashAttention;pub use attention::FlashAttentionConfig;pub use attention::KVCache;pub use attention::LayerKVCache;pub use attention::MultiHeadSelfAttention;pub use attention::scaled_dot_product_attention;pub use bert::Bert;pub use bert::BertForMaskedLM;pub use bert::BertForSequenceClassification;pub use chimera::ChimeraConfig;pub use chimera::ChimeraModel;pub use config::BertConfig;pub use config::GPT2Config;pub use config::TransformerConfig;pub use embedding::BertEmbedding;pub use embedding::GPT2Embedding;pub use embedding::PositionalEmbedding;pub use embedding::TokenEmbedding;pub use error::LLMError;pub use error::LLMResult;pub use generation::GenerationConfig;pub use generation::TextGenerator;pub use gpt2::GPT2;pub use gpt2::GPT2LMHead;pub use hf_loader::HFLoader;pub use hf_loader::load_llama_from_hf;pub use hf_loader::load_mistral_from_hf;pub use hub::PretrainedLLM;pub use hub::download_weights as download_llm_weights;pub use hub::llm_registry;pub use hydra::HydraConfig;pub use hydra::HydraModel;pub use llama::LLaMA;pub use llama::LLaMAConfig;pub use llama::LLaMAForCausalLM;pub use mistral::Mistral;pub use mistral::MistralConfig;pub use mistral::MistralForCausalLM;pub use phi::Phi;pub use phi::PhiConfig;pub use phi::PhiForCausalLM;pub use ssm::SSMBlock;pub use ssm::SSMConfig;pub use state_dict::LoadResult;pub use state_dict::LoadStateDict;pub use tokenizer::HFTokenizer;pub use tokenizer::SpecialTokens;pub use transformer::TransformerBlock;pub use transformer::TransformerDecoder;pub use transformer::TransformerEncoder;pub use trident::TridentConfig;pub use trident::TridentModel;
Modules§
- attention
- Attention Mechanisms Module
- bert
- BERT Model Implementation
- chimera
- Chimera - Mixture of Experts + Differential Attention Small Language Model
- config
- Model Configuration Module
- embedding
- Embedding Module
- error
- Error types for the LLM module.
- generation
- Text Generation Utilities
- gpt2
- GPT-2 Model Implementation
- hf_
loader - HuggingFace Model Loader
- hub
- LLM Model Hub - Pretrained Language Model Weights
- hydra
- Hydra - Hybrid SSM + Sparse Attention Small Language Model
- llama
- LLaMA - Large Language Model Meta AI
- mistral
- Mistral - Efficient LLM Architecture
- phi
- Phi - Microsoft’s Small Language Models
- ssm
- State Space Model (SSM) - Mamba-style Selective Scan
- state_
dict - State Dictionary Loading
- tokenizer
- HuggingFace Tokenizer Support
- transformer
- Transformer Building Blocks
- trident
- Trident - 1.58-bit Ternary Weight Small Language Model