List of all items
Structs
- context::LlamaContext
- context::llama_state_seq_flags::LlamaStateSeqFlags
- context::params::LlamaContextParams
- gguf_context::GgufContext
- llama_backend::LlamaBackend
- llama_backend_device::LlamaBackendDevice
- llama_backend_numa_strategy::InvalidNumaStrategy
- llama_batch::LlamaBatch
- log::State
- log_options::LogOptions
- model::LlamaModel
- model::chat_template_result::ChatTemplateResult
- model::grammar_trigger::GrammarTrigger
- model::llama_chat_message::LlamaChatMessage
- model::llama_chat_template::LlamaChatTemplate
- model::llama_lora_adapter::LlamaLoraAdapter
- model::params::LlamaModelParams
- model::params::kv_overrides::KvOverrideValueIterator
- model::params::kv_overrides::KvOverrides
- model::params::param_override_value::UnknownKvOverrideTag
- model::split_mode::LlamaSplitModeParseError
- openai::chat_parse_state_oaicompat::ChatParseStateOaicompat
- openai::openai_chat_template_params::OpenAIChatTemplateParams
- sampling::LlamaSampler
- timing::LlamaTimings
- token::LlamaToken
- token::data::LlamaTokenData
- token::data_array::LlamaTokenDataArray
- token::logit_bias::LlamaLogitBias
- token_type::LlamaTokenAttrs
Enums
- context::kv_cache::KvCacheConversionError
- context::load_seq_state_error::LoadSeqStateError
- context::load_session_error::LoadSessionError
- context::params::KvCacheType
- context::params::LlamaAttentionType
- context::params::LlamaPoolingType
- context::params::RopeScalingType
- context::save_seq_state_error::SaveSeqStateError
- context::save_session_error::SaveSessionError
- error::ApplyChatTemplateError
- error::ChatParseError
- error::ChatTemplateError
- error::DecodeError
- error::EmbeddingsError
- error::EncodeError
- error::GrammarError
- error::LlamaContextLoadError
- error::LlamaCppError
- error::LlamaLoraAdapterInitError
- error::LlamaLoraAdapterRemoveError
- error::LlamaLoraAdapterSetError
- error::LlamaModelLoadError
- error::LogitsError
- error::MetaValError
- error::ModelParamsError
- error::NewLlamaChatMessageError
- error::SampleError
- error::SamplerAcceptError
- error::SamplingError
- error::StringToTokenError
- error::TokenSamplingError
- error::TokenToStringError
- gguf_context_error::GgufContextError
- gguf_type::GgufType
- llama_backend_device::LlamaBackendDeviceType
- llama_backend_numa_strategy::NumaStrategy
- llama_batch::BatchAddError
- log::Module
- model::add_bos::AddBos
- model::grammar_trigger::GrammarTriggerType
- model::params::param_override_value::ParamOverrideValue
- model::rope_type::RopeType
- model::split_mode::LlamaSplitMode
- model::vocab_type::LlamaTokenTypeFromIntError
- model::vocab_type::VocabType
- openai::grammar_sampler_error::GrammarSamplerError
- token_type::LlamaTokenAttr
- token_type::LlamaTokenTypeFromIntError
Functions
- ffi_error_reader::read_and_free_cpp_error
- ffi_status_is_ok::status_is_ok
- ffi_status_to_i32::status_to_i32
- ggml_time_us::ggml_time_us
- json_schema_to_grammar::json_schema_to_grammar
- llama_backend_device::list_llama_ggml_backend_devices
- llama_time_us::llama_time_us
- log::send_logs_to_tracing
- max_devices::max_devices
- mlock_supported::mlock_supported
- mmap_supported::mmap_supported
- model::chat_template_result::new_empty_chat_template_raw_result
- model::chat_template_result::parse_chat_template_raw_result
- model::rope_type::rope_type_from_raw