use aether_core::core::{Prompt, agent};
use aether_core::events::{AgentMessage, UserMessage};
use llm::providers::openrouter::OpenRouterProvider;
use std::io::{self, Write};
#[allow(clippy::too_many_lines)]
#[tokio::main]
async fn main() -> Result<(), Box<dyn std::error::Error>> {
tracing_subscriber::fmt::init();
let llm = OpenRouterProvider::default("z-ai/glm-4.5-air")?;
let (tx, mut rx, _handle) = agent(llm).system_prompt(Prompt::text("You are a helpful assistant.")).spawn().await?;
tx.send(UserMessage::text("Write one paragraph about a unicorn")).await?;
loop {
use AgentMessage::{
AutoContinue, Cancelled, ContextCleared, ContextCompactionResult, ContextCompactionStarted,
ContextUsageUpdate, Done, Error, ModelSwitched, Text, Thought, ToolCall, ToolCallUpdate, ToolError,
ToolProgress, ToolResult,
};
match rx.recv().await {
Some(Text { chunk, is_complete, .. }) => {
if is_complete {
println!("\n\nMessage complete");
} else {
print!("{chunk}");
io::stdout().flush().unwrap();
}
}
Some(ToolCall { request, .. }) => {
println!("Tool '{}' in progress", request.name);
}
Some(ToolCallUpdate { .. }) => {}
Some(ToolResult { result, .. }) => {
println!("Tool '{}' completed", result.name);
println!(" Result: {}", result.result);
}
Some(ToolError { error, .. }) => {
eprintln!("Tool '{}' failed: {}", error.name, error.error);
}
Some(ToolProgress { request, progress, total, message }) => {
let msg = message.as_ref().map(|m| format!("{m} ")).unwrap_or_default();
let total_str = total.map(|t| format!("/{t}")).unwrap_or_default();
println!("Tool '{}' progress: {}{}{}", request.name, msg, progress, total_str);
}
Some(Done) => {
println!("Agent finished processing");
break;
}
Some(Error { message }) => {
eprintln!("Error: {message}");
break;
}
Some(Cancelled { .. }) => {
println!("Processing cancelled");
break;
}
Some(ContextCompactionStarted { message_count }) => {
println!("Context compaction started: {message_count} messages");
}
Some(ContextCompactionResult { messages_removed, .. }) => {
println!("Context compacted: {messages_removed} messages removed");
}
Some(ContextUsageUpdate { usage_ratio, input_tokens, context_limit, .. }) => {
match (usage_ratio, context_limit) {
(Some(usage_ratio), Some(context_limit)) => {
println!(
"Context usage: {:.1}% ({}/{} tokens)",
usage_ratio * 100.0,
input_tokens,
context_limit
);
}
_ => {
println!("Context usage: unknown limit ({input_tokens} tokens used)");
}
}
}
Some(AutoContinue { attempt, max_attempts }) => {
println!("Auto-continuing: attempt {attempt}/{max_attempts} (LLM stopped due to length)");
}
Some(ModelSwitched { previous, new }) => {
println!("Model switched: {previous} -> {new}");
}
Some(ContextCleared) => {
println!("Context cleared");
}
Some(Thought { chunk, .. }) => {
print!("{chunk}");
io::stdout().flush().unwrap();
}
None => {
println!("Channel closed");
break;
}
}
}
Ok(())
}