//! Model crate — LLM provider implementations via crabtalk, enum dispatch,
//! configuration, construction, and runtime management.
//!
//! Uses `crabtalk-provider` for the actual LLM backends (OpenAI, Anthropic,
//! Google, Bedrock, Azure). Wraps them behind wcore's `Model` trait with
//! type conversion and retry logic.
/// Default model name when none is configured.
pub use ;
pub use ProviderRegistry;
pub use ;
pub use Client;