autoagents_onnx/
lib.rs

1//! # Liquid Edge - Generic Edge Inference Runtime
2//!
3//! A lightweight, efficient inference runtime designed for edge computing environments.
4//! Supports multiple backends for running deep learning models on edge devices.
5
6pub mod device;
7pub mod error;
8pub mod model;
9pub mod runtime;
10
11pub mod chat;
12
13// Re-exports
14pub use device::{cpu, cpu_with_threads, Device};
15
16#[cfg(feature = "cuda")]
17pub use device::{cuda, cuda_default};
18
19pub use error::{EdgeError, EdgeResult};
20pub use model::Model;
21pub use runtime::{
22    inference::{OnnxBackend, OnnxModel},
23    onnx_model, InferenceInput, InferenceOutput, InferenceRuntime,
24};