//! Baseten Provider
//!
//! Baseten provides serverless inference for machine learning models.
//! This implementation provides access to Baseten's OpenAI-compatible API
//! for both Model API and dedicated deployments.
pub use BasetenConfig;
pub use BasetenError;
pub use BasetenProvider;