Qubit Thread Pool
Thread-pool executor services for Rust.
Overview
Qubit Thread Pool provides OS-thread based ExecutorService implementations for
synchronous work. It contains a dynamic ThreadPool for bursty workloads, a
FixedThreadPool for stable worker counts.
The crate is built on qubit-executor, so it shares the same task acceptance,
shutdown, cancellation, and TaskHandle semantics as other Qubit executor
implementations. It does not require Tokio or Rayon for normal use.
Features
- Dynamic
ThreadPoolwith separate core and maximum worker limits. - Fixed-size
FixedThreadPoolfor predictable worker counts. FixedThreadPoolimplementsDefault: it is built fromFixedThreadPoolBuilder::default()with the same defaults (worker count from available parallelism, unbounded queue, default thread name prefix). On failure to spawn workers it panics; useFixedThreadPoolBuilder::default().build()when you need aResult.- Bounded or unbounded queue configuration.
- Lazy worker creation for the dynamic pool, with optional core-worker prestart.
- Keep-alive and optional core-thread timeout for dynamic-pool workers.
- Configurable worker thread name prefixes and stack sizes.
- Worker and task lifecycle hooks for lightweight instrumentation.
ThreadPoolStatsfor observing pool configuration and runtime counters.- Shared
ExecutorServicelifecycle methods includingshutdown,stop, andwait_termination. - Criterion benchmarks and test data for comparing Qubit pools with
threadpooland Rayon.
Pool Models
ThreadPool is modeled after the common core-size / maximum-size executor
pattern. It creates workers lazily up to the core size, queues tasks after that,
and grows toward the maximum size when a bounded queue cannot accept more work.
This is useful when load is uneven and you want short bursts to use additional
threads without keeping them alive forever.
FixedThreadPool starts and maintains a fixed number of workers. It is useful
when capacity planning is simple, when worker count should be stable, or when
predictable scheduling is more important than dynamic growth.
FixedThreadPool::default() is equivalent to FixedThreadPoolBuilder::default().build() except that build errors become a panic; prefer the builder's build() when you must handle ExecutorServiceBuilderError.
Internally, the dynamic pool keeps accepted waiting work in a monitor-protected global FIFO queue. FIFO describes the waiting queue; it is not a strict task start or completion ordering guarantee. The fixed pool uses a lock-free global injector with targeted idle-worker wakeups, which keeps the fire-and-forget submit path small and predictable.
Runtime size setters on ThreadPool are intended for explicit control-plane
adjustments, such as operator-driven throttling or short-lived incident
mitigation. Normal application code should prefer choosing pool sizes at
construction time. Changing the core size at runtime updates future admission
and prestart behavior, but it does not eagerly create workers for already
queued work.
Queueing and Rejection
A pool can use either an unbounded queue or a bounded queue. Bounded queues make
back pressure explicit: when the pool cannot accept a task, submission returns
SubmissionError::Saturated instead of silently growing memory use.
A successful submit means only that the pool accepted a fire-and-forget
runnable. Use submit_callable when you need a TaskHandle for the final
result, or submit_tracked / submit_tracked_callable when you also need
status and pre-start cancellation.
Lifecycle Hooks
Both ThreadPoolBuilder and FixedThreadPoolBuilder support optional hooks for
worker and task instrumentation:
before_worker_startafter_worker_stopbefore_taskafter_task
Each hook receives the stable worker index and runs on the worker thread. Hook panics are caught and ignored so instrumentation cannot terminate workers or corrupt executor accounting. Keep hooks short; they are part of the execution hot path.
use ;
let pool = builder
.pool_size
.before_task
.after_task
.build?;
pool.submit?;
pool.shutdown;
# Ok::
Shutdown Behavior
shutdown stops accepting new tasks and lets already accepted tasks finish.
stop stops accepting new tasks and cancels work that is still queued or
not yet started. Already running OS-thread tasks are not forcefully killed; they
finish according to their own code.
wait_termination blocks the current thread after shutdown has been requested
until all accepted work has completed or been cancelled.
Quick Start
Dynamic thread pool
use io;
use ;
let pool = builder
.core_pool_size
.maximum_pool_size
.queue_capacity
.thread_name_prefix
.build?;
let handle = pool.submit_callable?;
assert_eq!;
pool.shutdown;
# Ok::
Fixed thread pool
use io;
use ;
let pool = builder
.pool_size
.queue_capacity
.build?;
let handle = pool.submit_callable?;
assert_eq!;
pool.shutdown;
# Ok::
With default builder settings you can also write let pool = FixedThreadPool::default();—same as FixedThreadPoolBuilder::default().build() but panics if worker threads cannot be spawned.
Choosing an Executor
Use ThreadPool when a service has blocking work with bursty traffic and needs
core/maximum worker tuning. Use FixedThreadPool when the desired worker count
is stable and should not grow dynamically. Use qubit-executor's scheduled
executor service when you need delayed or deadline-based task submission.
For CPU-bound divide-and-conquer work, prefer qubit-rayon-executor. For Tokio
applications, prefer qubit-tokio-executor for Tokio blocking tasks or async IO
futures. For application-level routing across all of these domains, use
qubit-execution-services.
Benchmarks
This crate includes Criterion benchmarks migrated with the thread-pool code from
the original concurrent module split. The benchmarks compare Qubit thread-pool
implementations with common open-source alternatives such as threadpool and
Rayon under representative submission patterns.
Run the benchmark suite with:
The submission-mode benchmark compares ThreadPool.submit,
ThreadPool.submit_tracked, FixedThreadPool.submit,
FixedThreadPool.submit_tracked, the external threadpool crate, and Rayon on
cpu_light, cpu_medium, and cpu_heavy tasks. CPU task costs are
deterministically varied with a bell-shaped distribution so worker scheduling,
queue contention, and wakeup behavior are visible instead of every task
completing at the same time.
Benchmark inputs and historical comparison data are kept under test-data.
Latest local run
The latest local run used
cargo bench --bench thread_pool_bench -- thread_pool_submit_modes on
2026-05-11 with Rust 1.94.1 on an Apple M3 Max with 16 hardware threads. The
tables below show Criterion mean wall-clock time for thread_pool_submit_modes;
lower is better. Each case submits 2,000 tasks. The submit_tracked cases use
the same channel-based task completion wait as submit, so the table compares
submission modes without mixing in handle wait costs.
cpu_light
| Workers | ThreadPool.submit |
ThreadPool.submit_tracked |
FixedThreadPool.submit |
FixedThreadPool.submit_tracked |
threadpool.execute |
Rayon |
|---|---|---|---|---|---|---|
| 1 | 0.444 ms | 0.546 ms | 0.388 ms | 0.439 ms | 0.386 ms | 0.144 ms |
| 4 | 0.726 ms | 1.285 ms | 0.560 ms | 0.981 ms | 0.740 ms | 0.082 ms |
| 8 | 1.758 ms | 4.561 ms | 0.967 ms | 1.402 ms | 1.065 ms | 0.142 ms |
cpu_medium
| Workers | ThreadPool.submit |
ThreadPool.submit_tracked |
FixedThreadPool.submit |
FixedThreadPool.submit_tracked |
threadpool.execute |
Rayon |
|---|---|---|---|---|---|---|
| 1 | 2.031 ms | 2.133 ms | 2.029 ms | 2.115 ms | 2.037 ms | 1.439 ms |
| 4 | 1.354 ms | 1.057 ms | 1.321 ms | 1.296 ms | 1.455 ms | 0.425 ms |
| 8 | 1.902 ms | 3.868 ms | 0.959 ms | 1.280 ms | 2.022 ms | 0.391 ms |
cpu_heavy
| Workers | ThreadPool.submit |
ThreadPool.submit_tracked |
FixedThreadPool.submit |
FixedThreadPool.submit_tracked |
threadpool.execute |
Rayon |
|---|---|---|---|---|---|---|
| 1 | 14.256 ms | 14.360 ms | 14.251 ms | 14.198 ms | 14.157 ms | 11.078 ms |
| 4 | 4.384 ms | 4.588 ms | 4.715 ms | 4.533 ms | 4.594 ms | 3.311 ms |
| 8 | 3.505 ms | 3.502 ms | 3.391 ms | 3.993 ms | 4.335 ms | 2.965 ms |
Testing
A minimal local run:
To mirror what continuous integration enforces, run the repository scripts from
the project root: ./align-ci.sh brings local tooling and configuration in line
with CI, then ./ci-check.sh runs the same checks the pipeline uses. For test
coverage, use ./coverage.sh to generate or open reports.
Contributing
Issues and pull requests are welcome.
- Open an issue for bug reports, design questions, or larger feature proposals when it helps align on direction.
- Keep pull requests scoped to one behavior change, fix, or documentation update when practical.
- Before submitting, run
./align-ci.shand then./ci-check.shso your branch matches CI rules and passes the same checks as the pipeline. - Add or update tests when you change runtime behavior, and update this README or public rustdoc when user-visible API behavior changes.
- If you change scheduling, queueing, or shutdown behavior, include tests that cover both
ThreadPoolandFixedThreadPoolwhen the behavior applies to both.
By contributing, you agree to license your contributions under the Apache License, Version 2.0, the same license as this project.
License
Copyright (c) 2026. Haixing Hu.
This project is licensed under the Apache License, Version 2.0. See the LICENSE file in the repository for the full text.
Author
Haixing Hu — Qubit Co. Ltd.
| Repository | github.com/qubit-ltd/rs-thread-pool |
| Documentation | docs.rs/qubit-thread-pool |
| Crate | crates.io/crates/qubit-thread-pool |