ugnos: Concurrent Time-Series Database Core in Rust
ugnos is a concurrent, embeddable time-series storage + query engine designed for durability and high-throughput ingest in Rust services.
For project goals and long-term architecture, see the whitepaper.
For latest changelog, see the CHANGELOG. README may lag behind.
What this is / what this is not (yet)
This crate is a library-grade database core intended to be embedded into a Rust process (service/agent/daemon).
- This is:
- An embeddable time-series ingest + query core with WAL/snapshots/segments (SST-like) and a structured event hook.
- Suitable for single-process usage where you own deployment, IO, and operational integration.
- This is not (yet):
- A networked database server (no HTTP/gRPC API, authn/authz, multi-tenant isolation).
- A distributed system (no replication, consensus, sharding across nodes).
- A full query language / SQL layer (queries are programmatic APIs).
- A turnkey operational product (no built-in backup orchestration, migrations tooling, or admin UI).
Features (today)
- Concurrent ingest: sharded write buffering + background flush thread.
- Durable persistence:
- WAL with explicit format versioning and per-record CRC32 checksums.
- Snapshots with explicit format versioning, payload CRC32, and atomic install (temp + rename + fsync).
- On-disk segment engine (SST-like):
- Immutable segment files with per-series columnar blocks; block-level checksums and versioning (v2 header with payload CRC32 and version field).
- Timestamp delta encoding (varint) for series blocks; configurable float encoding strategies (Raw64, GorillaXor).
- Tag dictionary encoding; optional per-block compression (LZ4, Zstd with configurable level).
- Time index per segment/block (time-range via binary search); tag index (inverted index with Roaring bitmaps) for tag filters without full scans.
- Atomic manifest (
MANIFEST.bin) tracking active segments and retention watermark. - Background compaction (L0 → L1 merge) with safe concurrent reads.
- Indexing & cardinality:
- Tag filters use the tag index (bitmap intersection); configurable series cardinality hard limit per scope with explicit error and metrics.
- Retention/TTL:
- Immediate logical deletion via tombstone watermark.
- Physical removal via compaction guarantees.
- Observability hooks:
- No stdout logging in core hot paths.
- Structured
DbEventstream viaDbConfig.event_listener.
- Bench suite reports segment size per encoding configuration; tests assert p99 query latency within target for encoded/compressed segments.
- Acceptance and break-it tests covering format layout validation, checksum/version enforcement, roundtrip correctness, and corruption detection.
Data layout on disk
All persistence lives under DbConfig.data_dir:
wal/wal.log(current WAL)wal_*.log(rotated WAL segments; may exist briefly)
snapshots/snapshot_<timestamp>.bin(atomic, checksummed snapshots)
engine/segments/MANIFEST.bin(atomic + checksummed)seg_<id>_l0.seg,seg_<id>_l1.seg, ...
Core API notes
DbCore::flush()blocks until the flush is complete.DbCore::snapshot()blocks until the snapshot is written (when enabled).- With segments enabled,
DbCore::recover():- Uses segment max-seq to replay only the WAL tail.
- Truncates
wal.logback to just the header (bounded restart cost).
- Query results are not guaranteed to be globally sorted across multiple segments; sort by timestamp if you need ordering.
Configuration
DbConfig is intended to be explicit and production-friendly:
use PathBuf;
use Duration;
use TempDir;
use ;
use ;
let dir = new.unwrap;
let mut cfg = default;
cfg.data_dir = from;
// Durability toggles
cfg.enable_wal = true;
cfg.enable_snapshots = true;
cfg.enable_segments = true; // segment engine + compaction + retention
// Tuning
cfg.wal_buffer_size = 1_000;
cfg.flush_interval = from_millis;
cfg.snapshot_interval = from_secs;
// Retention (optional):
// makes data older than now - ttl invisible,
// and compaction reclaims disk.
cfg.retention_ttl = Some;
cfg.retention_check_interval = from_secs;
// Encoding & compression (series blocks in segments):
// float (Raw64 | GorillaXor), tag dictionary, LZ4/Zstd.
cfg.segment_store.encoding.float_encoding = GorillaXor;
cfg.segment_store.encoding.compression = Zstd ;
// Cardinality (optional):
// hard limit for distinct series per scope;
// scope is derived from tags[cardinality_scope_tag_key].
// When exceeded, insert returns DbError::SeriesCardinalityLimitExceeded
// and metrics:
// - ugnos_cardinality_limit_rejections
// - ugnos_series_cardinality
// cfg.max_series_cardinality = Some(100);
// cfg.cardinality_scope_tag_key = Some("tenant".to_string());
let mut db = with_config.unwrap;
db.recover.unwrap;
Observability (event hook)
Core emits structured events via DbConfig.event_listener. With cardinality limits enabled, telemetry exposes ugnos_cardinality_limit_rejections and ugnos_series_cardinality (when using the Prometheus recorder).
use ;
use ;
;
Basic usage
use PathBuf;
use ;
use Duration;
use TempDir;
use ;
;
How to build and test (this workspace)
Benchmarks
NOWAL=1
Benchmark results are saved in target/criterion/. Benchmarks default to the in-memory engine (segments disabled) to keep IO minimal.
License
This project is licensed under either of
- MIT License (LICENSE-MIT or http://opensource.org/licenses/MIT)
- Apache License, Version 2.0 (LICENSE-APACHE or http://www.apache.org/licenses/LICENSE-2.0)
at your option.