List of all items
Structs
- audit::AuditChain
- audit::AuditEntry
- audit::IntegrityFields
- budget::TokenBudget
- budget::TokenPool
- cache::CacheConfig
- cache::CacheStats
- cache::ResponseCache
- cache::semantic::SemanticCache
- cache::semantic::SemanticCacheConfig
- cache::warming::WarmingMessage
- cache::warming::WarmingPrompt
- client::HooshClient
- config::AuditSection
- config::AuthConfig
- config::BudgetPoolSection
- config::CacheSection
- config::ContextSection
- config::HooshConfig
- config::ProviderSection
- config::ServerSection
- config::TelemetrySection
- config::TtsSection
- config::WhisperSection
- context::compactor::CompactionResult
- context::compactor::ContextCompactor
- context::tokens::ProviderTokenCounter
- context::tokens::SimpleTokenCounter
- cost::CostTracker
- cost::ModelPricing
- cost::ProviderCostRecord
- cost::optimizer::CostOptimizer
- cost::optimizer::ModelRecommendation
- cost::optimizer::RequestProfile
- hardware::HardwareManager
- hardware::PlacementRecommendation
- hardware::ShardInfo
- hardware::ShardingSummary
- health::ProviderHealthState
- inference::EmbeddingData
- inference::EmbeddingsRequest
- inference::EmbeddingsResponse
- inference::EmbeddingsUsage
- inference::ImageUrl
- inference::InferenceRequest
- inference::InferenceResponse
- inference::Message
- inference::ModelInfo
- inference::SpeechRequest
- inference::SpeechResponse
- inference::TokenUsage
- inference::TranscriptionRequest
- inference::TranscriptionResponse
- inference::TranscriptionSegment
- inference::batch::BatchItemResult
- inference::batch::BatchManager
- inference::batch::BatchProgress
- middleware::rate_limit::RateLimitRegistry
- provider::ProviderRegistry
- provider::TlsConfig
- provider::anthropic::AnthropicProvider
- provider::deepseek::DeepSeekProvider
- provider::grok::GrokProvider
- provider::groq::GroqProvider
- provider::llamacpp::LlamaCppProvider
- provider::lmstudio::LmStudioProvider
- provider::localai::LocalAiProvider
- provider::metadata::ModelCapabilities
- provider::metadata::ModelMetadata
- provider::metadata::ModelMetadataRegistry
- provider::mistral::MistralProvider
- provider::ollama::OllamaProvider
- provider::openai_compat::OpenAiCompatibleProvider
- provider::openai_remote::OpenAiProvider
- provider::openrouter::OpenRouterProvider
- provider::retry::RetryConfig
- provider::retry::RetryManager
- provider::synapse::SynapseProvider
- queue::InferenceQueue
- queue::QueuedRequest
- router::ProviderRoute
- router::Router
- server::AppState
- server::ServerConfig
- tools::ToolCall
- tools::ToolDefinition
- tools::ToolResult
Enums
- config::StrategyValue
- error::HooshError
- events::ProviderEvent
- inference::ContentPart
- inference::EmbeddingsInput
- inference::MessageContent
- inference::Role
- inference::batch::BatchStatus
- provider::ProviderType
- provider::metadata::Modality
- provider::metadata::ModelTier
- router::RoutingStrategy
- tools::ToolChoice
Traits
Functions
- cache::cache_key
- cache::warming::spawn_warming_task
- cache::warming::to_inference_requests
- context::compression::compress_messages
- events::new_event_bus
- health::new_health_map
- health::spawn_health_checker
- metrics::gather
- metrics::record_request
- metrics::set_providers_configured
- middleware::auth::auth_middleware
- middleware::auth::hash_token
- provider::build_provider_client
- provider::retry::retry_with
- server::build_app
- server::run
- tools::extract_anthropic_text
- tools::parse_anthropic_tool_calls
- tools::parse_openai_tool_calls
- tools::to_anthropic_tools
- tools::to_openai_tools