llama-gguf 0.14.0

A high-performance Rust implementation of llama.cpp - LLM inference engine with full GGUF support
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
//! Tensor module for llama-rs
//!
//! This module provides tensor types for representing multi-dimensional arrays
//! with support for various data types including quantized formats.

mod core;
mod dtype;
mod error;
pub mod quant;
mod storage;

pub use core::{Tensor, compute_strides};
pub use dtype::DType;
pub use error::TensorError;
pub use storage::TensorStorage;