harn-vm 0.7.58

Async bytecode virtual machine for the Harn programming language
Documentation
//! Built-in LLM provider implementations.
//!
//! Each provider implements the `LlmProvider` + `LlmProviderOps` traits from
//! `super::provider`. The three main families are:
//!
//! - **Anthropic** — Claude models via the Anthropic Messages API
//! - **OpenAI-compatible** — OpenAI, OpenRouter, Together, Groq, DeepSeek,
//!   Fireworks, HuggingFace, local vLLM/SGLang servers, etc.
//! - **Ollama** — local Ollama server with NDJSON streaming
//! - **Bedrock / Azure OpenAI / Vertex AI** — enterprise cloud shims
//! - **Mock** — deterministic test responses without any network I/O

pub(crate) mod anthropic;
pub(crate) mod azure_openai;
pub(crate) mod bedrock;
mod common;
mod gemini;
mod mock;
mod ollama;
pub(crate) mod openai_compat;
pub(crate) mod vertex;

pub(crate) use anthropic::AnthropicProvider;
pub(crate) use azure_openai::AzureOpenAiProvider;
pub(crate) use bedrock::BedrockProvider;
pub(crate) use gemini::GeminiProvider;
pub(crate) use mock::MockProvider;
pub(crate) use ollama::OllamaProvider;
pub(crate) use openai_compat::OpenAiCompatibleProvider;
pub(crate) use vertex::VertexProvider;