rustqueue 0.2.0

Background jobs without infrastructure — embeddable job queue with zero external dependencies
Documentation
# RustQueue v0.2.0 Benchmark Results

**Date:** 2026-03-28
**Platform:** macOS Darwin 25.3.0, Apple Silicon
**Profile:** `cargo bench` (release, optimized)

## Single Operation Benchmarks

| Benchmark | Time (median) | vs Previous |
|-----------|------------:|------------|
| push_single_job (redb, immediate fsync) | 3.18 ms | -3.7% improved |
| push_single_job_eventual (redb, eventual) | 405 µs | no change |
| push_pull_ack_roundtrip (redb, immediate) | 13.69 ms | +6.5% regressed |
| push_pull_ack_roundtrip_eventual | 1.14 ms | -35.7% improved |

**Throughput (single-threaded sequential):**
- Immediate fsync: ~314 push/s, ~73 push+pull+ack/s
- Eventual durability: ~2,469 push/s, ~876 push+pull+ack/s

## Batch Push Benchmarks (redb, immediate)

| Batch Size | Time (median) | Jobs/sec |
|----------:|------------:|---------:|
| 10 | 46.5 ms | 215 |
| 100 | 464 ms | 216 |
| 1000 | 4.38 s | 228 |

## Concurrent Push — Raw redb

| Concurrency | Time (median) | Jobs/sec |
|-----------:|------------:|---------:|
| 10 | 41.5 ms | 241 |
| 50 | 208.6 ms | 240 |
| 100 | 423.7 ms | 236 |

## Concurrent Push — Buffered redb (write coalescing)

| Concurrency | Time (median) | Jobs/sec |
|-----------:|------------:|---------:|
| 10 | 16.5 ms | 606 |
| 50 | 16.7 ms | 2,994 |
| 100 | 5.24 ms | **19,084** |

Write coalescing delivers ~80x throughput improvement at 100 concurrent callers.

## Concurrent Push — Hybrid (DashMap + redb snapshots)

| Concurrency | Time (median) | Jobs/sec |
|-----------:|------------:|---------:|
| 10 | 32.2 µs | **310,559** |
| 50 | 115.5 µs | **432,900** |
| 100 | 270.1 µs | **370,256** |

Hybrid storage delivers microsecond-level push latency. In-memory DashMap hot path with background disk persistence.

## Key Takeaways

1. **No performance regression from v0.1.0** — the repositioning changes were documentation/examples only, no engine code was modified.
2. **Hybrid storage remains the throughput champion** — 300K+ ops/sec for concurrent push workloads.
3. **Buffered redb scales well** — 19K ops/sec at 100 concurrent callers (80x over raw redb).
4. **Immediate fsync redb** is I/O-bound at ~230 sequential ops/sec — expected for ACID durability.

## Competitor Benchmarks

Competitor benchmarks (vs RabbitMQ, Redis, BullMQ, Celery) require external services. See `docs/competitor-benchmark-2026-03-28.md` for the most recent comparison. No engine changes since that benchmark — results remain valid.