//! Petals Distributed LLM Provider
//!
//! Petals allows running large language models collaboratively through distributed inference.
//! This implementation provides access to chat completions through Petals' OpenAI-compatible API.
pub use PetalsConfig;
pub use ;
pub use PetalsProvider;