1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
//! # Infernum Server
//!
//! HTTP API server with OpenAI-compatible endpoints.
//!
//! ## Features
//!
//! - **OpenAI API Compatibility**: Drop-in replacement for OpenAI's API
//! - **Chat Completions**: `/v1/chat/completions` with streaming support
//! - **Text Completions**: `/v1/completions` for raw text generation
//! - **Embeddings**: `/v1/embeddings` for vector generation
//! - **Model Management**: Load/unload models at runtime
//! - **Health Checks**: `/health` and `/ready` endpoints
//!
//! ## Example
//!
//! ```ignore
//! use infernum_server::{Server, ServerConfig};
//!
//! #[tokio::main]
//! async fn main() -> Result<(), Box<dyn std::error::Error>> {
//! let config = ServerConfig::builder()
//! .addr("0.0.0.0:8080".parse()?)
//! .model("meta-llama/Llama-3.2-3B-Instruct")
//! .build();
//!
//! let server = Server::new(config);
//! server.run().await?;
//!
//! Ok(())
//! }
//! ```
pub use ;
pub use ;