Expand description
Low-level GPU inference support via ONNX Runtime ExecutionProviders.
This module handles the ort integration layer – it configures ONNX
Runtime SessionBuilder instances with the appropriate ExecutionProvider
(CUDA, CoreML, DirectML, TensorRT) and manages device string parsing for
the engine.
Feature-gated: cuda, coreml, directml, tensorrt features enable
respective providers. Auto-detection tries available providers and falls
back to CPU.
For the high-level, user-facing device enumeration and selection API, see
crate::device.
Structs§
- Device
Info - Information about an available compute device.
Enums§
- Device
Type - Supported GPU device types.
Functions§
- configure_
session_ builder - Configure an ONNX Runtime session builder with the appropriate ExecutionProvider.
- list_
devices - List all available compute devices.
- parse_
device_ string - Parse a device string (“cpu”, “cuda”, “cuda:0”, “cuda:1”, “coreml”,
“directml”, “directml:2”, “tensorrt”, “tensorrt:0”, “auto”) into a
DeviceType.