use rig::client::{CompletionClient, ProviderClient};
use rig::providers::openai;
use rig::streaming::StreamingPrompt;
use crate::support::{
STREAMING_PREAMBLE, STREAMING_PROMPT, assert_nonempty_response, collect_stream_final_response,
};
#[tokio::test]
#[ignore = "requires OPENAI_API_KEY"]
async fn streaming_smoke() {
let client = openai::Client::from_env();
let agent = client
.agent(openai::GPT_4O)
.preamble(STREAMING_PREAMBLE)
.build();
let mut stream = agent.stream_prompt(STREAMING_PROMPT).await;
let response = collect_stream_final_response(&mut stream)
.await
.expect("streaming prompt should succeed");
assert_nonempty_response(&response);
}
#[tokio::test]
#[ignore = "requires OPENAI_API_KEY"]
async fn example_streaming_prompt() {
let client = openai::Client::from_env();
let agent = client
.agent(openai::GPT_4O)
.preamble("Be precise and concise.")
.temperature(0.5)
.build();
let mut stream = agent
.stream_prompt("When and where and what type is the next solar eclipse?")
.await;
let response = collect_stream_final_response(&mut stream)
.await
.expect("streaming prompt should succeed");
assert_nonempty_response(&response);
}