Skip to main content

Crate axonml_llm

Crate axonml_llm 

Source
Expand description

axonml-llm - Large Language Model Architectures

§File

crates/axonml-llm/src/lib.rs

§Author

Andrew Jewell Sr - AutomataNexus

§Updated

March 8, 2026

§Disclaimer

Use at own risk. This software is provided “as is”, without warranty of any kind, express or implied. The author and AutomataNexus shall not be held liable for any damages arising from the use of this software.

Re-exports§

pub use attention::CausalSelfAttention;
pub use attention::FlashAttention;
pub use attention::FlashAttentionConfig;
pub use attention::KVCache;
pub use attention::LayerKVCache;
pub use attention::MultiHeadSelfAttention;
pub use attention::scaled_dot_product_attention;
pub use bert::Bert;
pub use bert::BertForMaskedLM;
pub use bert::BertForSequenceClassification;
pub use chimera::ChimeraConfig;
pub use chimera::ChimeraModel;
pub use config::BertConfig;
pub use config::GPT2Config;
pub use config::TransformerConfig;
pub use embedding::BertEmbedding;
pub use embedding::GPT2Embedding;
pub use embedding::PositionalEmbedding;
pub use embedding::TokenEmbedding;
pub use error::LLMError;
pub use error::LLMResult;
pub use generation::GenerationConfig;
pub use generation::TextGenerator;
pub use gpt2::GPT2;
pub use gpt2::GPT2LMHead;
pub use hf_loader::HFLoader;
pub use hf_loader::load_llama_from_hf;
pub use hf_loader::load_mistral_from_hf;
pub use hub::PretrainedLLM;
pub use hub::download_weights as download_llm_weights;
pub use hub::llm_registry;
pub use hydra::HydraConfig;
pub use hydra::HydraModel;
pub use llama::LLaMA;
pub use llama::LLaMAConfig;
pub use llama::LLaMAForCausalLM;
pub use mistral::Mistral;
pub use mistral::MistralConfig;
pub use mistral::MistralForCausalLM;
pub use phi::Phi;
pub use phi::PhiConfig;
pub use phi::PhiForCausalLM;
pub use ssm::SSMBlock;
pub use ssm::SSMConfig;
pub use state_dict::LoadResult;
pub use state_dict::LoadStateDict;
pub use tokenizer::HFTokenizer;
pub use tokenizer::SpecialTokens;
pub use transformer::TransformerBlock;
pub use transformer::TransformerDecoder;
pub use transformer::TransformerEncoder;
pub use trident::TridentConfig;
pub use trident::TridentModel;

Modules§

attention
Attention Mechanisms Module
bert
BERT Model Implementation
chimera
Chimera - Mixture of Experts + Differential Attention Small Language Model
config
Model Configuration Module
embedding
Embedding Module
error
Error types for the LLM module.
generation
Text Generation Utilities
gpt2
GPT-2 Model Implementation
hf_loader
HuggingFace Model Loader
hub
LLM Model Hub - Pretrained Language Model Weights
hydra
Hydra - Hybrid SSM + Sparse Attention Small Language Model
llama
LLaMA - Large Language Model Meta AI
mistral
Mistral - Efficient LLM Architecture
phi
Phi - Microsoft’s Small Language Models
ssm
State Space Model (SSM) - Mamba-style Selective Scan
state_dict
State Dictionary Loading
tokenizer
HuggingFace Tokenizer Support
transformer
Transformer Building Blocks
trident
Trident - 1.58-bit Ternary Weight Small Language Model