skill_runtime/generation/
mod.rs1mod streaming;
7mod llm_provider;
8mod validator;
9mod example_generator;
10mod evaluation;
11mod fixtures;
12
13#[cfg(test)]
14mod integration_tests;
15
16pub use streaming::*;
17pub use evaluation::*;
18pub use fixtures::*;
19
20pub use streaming::{GenerationEvent, GeneratedExample, AgentStep};
22pub use llm_provider::{
23 LlmProvider, LlmResponse, LlmChunk, TokenUsage,
24 ChatMessage, CompletionRequest, create_llm_provider,
25};
26pub use validator::{ExampleValidator, ValidationResult, ParsedCommand};
27pub use example_generator::{ExampleGenerator, GeneratorConfig};
28
29#[cfg(feature = "ollama")]
30pub use llm_provider::ollama::OllamaProvider;
31
32#[cfg(feature = "openai")]
33pub use llm_provider::openai::OpenAIProvider;