1#[cfg(feature = "llama-cpp")]
8pub mod llm;
9pub mod openai;
10pub mod claude;
11pub mod gemini;
12pub mod xai;
13pub mod groq;
14pub mod mistral;
15
16#[cfg(feature = "candle-llm")]
17pub mod candle_phi;
18
19#[cfg(feature = "llama-cpp")]
20pub use llm::LlmEngine;
21pub use openai::OpenAiEngine;
22pub use claude::ClaudeEngine;
23pub use gemini::GeminiEngine;
24pub use xai::XaiEngine;
25pub use groq::GroqEngine;
26pub use mistral::MistralEngine;
27
28#[cfg(feature = "candle-llm")]
29pub use candle_phi::CandlePhiEngine;