sockudo-ws
Ultra-low latency WebSocket library for Rust, designed for high-frequency trading (HFT) applications and real-time systems. Fully compatible with Tokio and Axum.
Used in Sockudo, a high-performance Pusher-compatible WebSocket server.
Coming soon: N-API bindings for Node.js
Performance
Rust WebSocket Libraries Benchmark
Benchmarked using web-socket-benchmark (100,000 iterations of "Hello, World!" message):
| Library | Send | Echo | Recv | Total |
|---|---|---|---|---|
| sockudo-ws | 1.2ms | 5.0ms | 3.1ms | 10.2ms |
| fastwebsockets | 3.3ms | 5.7ms | 3.0ms | 12.0ms |
| web-socket | 2.1ms | 6.8ms | 3.3ms | 12.2ms |
| soketto | 5.8ms | 17.6ms | 9.7ms | 33.1ms |
| tokio-tungstenite | 6.4ms | 18.2ms | 10.2ms | 34.8ms |
sockudo-ws is ~17% faster than the next fastest Rust WebSocket library!
vs uWebSockets (C++)
Benchmarked against uWebSockets, the industry standard for high-performance WebSockets:
| Test Case | sockudo-ws | uWebSockets | Ratio |
|---|---|---|---|
| 512 bytes, 100 connections | 232,712 msg/s | 227,973 msg/s | 1.02x |
| 1024 bytes, 100 connections | 232,072 msg/s | 224,498 msg/s | 1.03x |
| 512 bytes, 500 connections | 231,135 msg/s | 222,493 msg/s | 1.03x |
| 1024 bytes, 500 connections | 222,578 msg/s | 216,833 msg/s | 1.02x |
Benchmarked on AMD Ryzen 9 7950X, 32GB RAM, Linux 6.18
sockudo-ws matches or exceeds uWebSockets performance while providing a safe, ergonomic Rust API.
Features
- SIMD Acceleration: AVX2/AVX-512/NEON for frame masking and UTF-8 validation
- Zero-Copy Parsing: Direct buffer access without intermediate allocations
- Write Batching (Corking): Minimizes syscalls via vectored I/O
- permessage-deflate: Full compression support with shared/dedicated compressors
- Split Streams: Concurrent read/write from separate tasks
- HTTP/2 WebSocket: RFC 8441 Extended CONNECT protocol support
- HTTP/3 WebSocket: RFC 9220 WebSocket over QUIC support
- io_uring: Linux high-performance async I/O (combinable with HTTP/2 and HTTP/3)
- Autobahn Compliant: Passes all 517 Autobahn test suite cases
Installation
Add to your Cargo.toml:
[]
= { = "https://github.com/RustNSparks/sockudo-ws" }
# With compression
= { = "https://github.com/RustNSparks/sockudo-ws", = ["permessage-deflate"] }
# With HTTP/2 support
= { = "https://github.com/RustNSparks/sockudo-ws", = ["http2"] }
# With HTTP/3 support
= { = "https://github.com/RustNSparks/sockudo-ws", = ["http3"] }
# With io_uring (Linux only)
= { = "https://github.com/RustNSparks/sockudo-ws", = ["io-uring"] }
# All transports
= { = "https://github.com/RustNSparks/sockudo-ws", = ["all-transports"] }
# Everything
= { = "https://github.com/RustNSparks/sockudo-ws", = ["full"] }
Quick Start
Simple Echo Server
use ;
use ;
use TcpStream;
async
Split Streams (Concurrent Read/Write)
use ;
use mpsc;
async
Axum Integration
use ;
use ;
use TokioIo;
use ;
async
HTTP/2 WebSocket (RFC 8441)
HTTP/2 WebSocket uses the Extended CONNECT protocol for multiplexed WebSocket streams over a single TCP connection.
use H2WebSocketServer;
use ;
use ;
async
HTTP/2 Client
use H2WebSocketClient;
let client = new;
let mut ws = client.connect.await?;
ws.send.await?;
HTTP/2 Multiplexed Connections
Open multiple WebSocket streams over a single HTTP/2 connection:
let client = new;
let mut conn = client.connect_multiplexed.await?;
// Open multiple WebSocket streams on the same connection
let mut ws1 = conn.open_websocket.await?;
let mut ws2 = conn.open_websocket.await?;
HTTP/3 WebSocket (RFC 9220)
HTTP/3 WebSocket runs over QUIC, providing benefits like 0-RTT, no head-of-line blocking, and better mobile performance.
use H3WebSocketServer;
use ;
use ;
async
HTTP/3 Benefits
| Feature | Benefit |
|---|---|
| No head-of-line blocking | One slow stream doesn't block others |
| 0-RTT connection resumption | Faster reconnections |
| Better mobile performance | Handles network changes gracefully |
| Multiple streams per connection | Efficient multiplexing |
io_uring Support (Linux)
io_uring provides kernel-level async I/O with zero-copy operations. It's a transport layer that can be combined with any protocol.
io_uring with HTTP/1.1
use UringStream;
use ;
async
io_uring with HTTP/2
Combine io_uring transport with HTTP/2 protocol for maximum performance:
use UringStream;
use H2WebSocketServer;
use Config;
async
The io_uring + HTTP/2 Stack
┌─────────────────────────────┐
│ WebSocket Messages │ ← Your application code
├─────────────────────────────┤
│ WebSocketStream<H2Stream> │ ← sockudo-ws
├─────────────────────────────┤
│ HTTP/2 (h2 crate) │ ← Extended CONNECT framing
├─────────────────────────────┤
│ TLS (rustls/openssl) │ ← Required for HTTP/2
├─────────────────────────────┤
│ UringStream │ ← io_uring async I/O
├─────────────────────────────┤
│ TCP (kernel) │ ← io_uring submission queue
└─────────────────────────────┘
Unified API
All transports use the same WebSocketStream<S> API:
// HTTP/1.1 (default)
let ws = server;
// HTTP/2
let ws = server;
// HTTP/3
let ws = server;
// io_uring
let ws = server;
// Same message loop for all!
while let Some = ws.next.await
Configuration
Basic Configuration
use ;
let config = builder
.compression // SHARED_COMPRESSOR
.max_payload_length // 16KB max message
.idle_timeout // 10 second timeout
.max_backpressure // 1MB backpressure limit
.build;
// Or use uWebSockets-style defaults
let config = uws_defaults;
HTTP/2 Configuration
let config = builder
.http2_window_size // 1MB stream window
.http2_connection_window_size // 2MB connection window
.http2_max_streams // Max concurrent streams
.build;
HTTP/3 Configuration
let config = builder
.http3_idle_timeout // 30 second idle timeout
.build;
Configuration Options
| Option | Default | Description |
|---|---|---|
compression |
Disabled |
Compression mode |
max_message_size |
64MB | Maximum message size |
max_frame_size |
16MB | Maximum single frame size |
idle_timeout |
120s | Close connection after inactivity (0 = disabled) |
max_backpressure |
1MB | Max write buffer before dropping connection |
auto_ping |
true | Automatic ping/pong keepalive |
ping_interval |
30s | Seconds between pings |
write_buffer_size |
16KB | Cork buffer size |
Compression Modes
| Mode | Description |
|---|---|
Compression::Disabled |
No compression |
Compression::Dedicated |
Per-connection compressor (best ratio, more memory) |
Compression::Shared |
Shared compressor (good for many connections) |
Compression::Shared4KB |
Shared with 4KB sliding window |
Compression::Shared8KB |
Shared with 8KB sliding window |
Compression::Shared16KB |
Shared with 16KB sliding window |
Feature Flags
| Feature | Description |
|---|---|
simd |
SIMD acceleration (default) |
tokio-runtime |
Tokio async runtime (default) |
permessage-deflate |
Compression support (default) |
axum-integration |
Axum web framework support |
http2 |
HTTP/2 WebSocket (RFC 8441) |
http3 |
HTTP/3 WebSocket (RFC 9220) |
io-uring |
Linux io_uring support |
all-transports |
All transport features |
full |
Everything |
API Reference
WebSocketStream
The main WebSocket type implementing Stream + Sink:
// Create server-side stream
let ws = server;
// Create client-side stream
let ws = client;
// Send messages
ws.send.await?;
ws.send.await?;
// Receive messages
while let Some = ws.next.await
// Close connection
ws.close.await?;
Split Streams
For concurrent read/write operations:
let = ws.split;
// SplitReader
reader.next.await // Receive message
// SplitWriter
writer.send.await?;
writer.send_text.await?;
writer.send_binary.await?;
writer.close.await?;
writer.is_closed.await;
// Reunite
let ws = reunite?;
Message Types
Zero-Copy API (RawMessage)
For maximum performance, use RawMessage to avoid String allocation for text messages:
Using RawMessage with the Protocol layer:
use ;
use BytesMut;
let mut protocol = new;
let mut buffer = new;
let mut messages = Vecnew;
// Parse incoming data into RawMessage (zero String allocation)
protocol.process_raw_into?;
for msg in messages.drain
This API is ~17% faster than the standard Message API by avoiding heap allocation for text message payloads.
Running Tests
Unit Tests
With Features
Autobahn Test Suite
# Build and run server + tests
# Or manually:
# Then run Autobahn client in another terminal
Examples
Run the examples:
# Basic echo server
# Split streams (concurrent read/write)
# Axum integration
# HTTP/2 WebSocket server
# HTTP/3 WebSocket server
Architecture
sockudo-ws/
├── src/
│ ├── lib.rs # Public API, Config
│ ├── stream.rs # WebSocketStream, Split types
│ ├── protocol.rs # WebSocket protocol state machine
│ ├── frame.rs # Frame encoding/decoding
│ ├── handshake.rs # HTTP upgrade handshake
│ ├── mask.rs # SIMD frame masking
│ ├── utf8.rs # SIMD UTF-8 validation
│ ├── cork.rs # Write batching buffer
│ ├── deflate.rs # permessage-deflate compression
│ ├── simd.rs # SIMD feature detection
│ ├── http2/ # HTTP/2 WebSocket (RFC 8441)
│ │ ├── mod.rs
│ │ ├── stream.rs # H2Stream wrapper
│ │ ├── handshake.rs
│ │ ├── server.rs # H2WebSocketServer
│ │ └── client.rs # H2WebSocketClient
│ ├── http3/ # HTTP/3 WebSocket (RFC 9220)
│ │ ├── mod.rs
│ │ ├── stream.rs # H3Stream wrapper
│ │ ├── handshake.rs
│ │ ├── server.rs # H3WebSocketServer
│ │ └── client.rs # H3WebSocketClient
│ └── io_uring/ # Linux io_uring transport
│ ├── mod.rs
│ ├── stream.rs # UringStream wrapper
│ └── buffer.rs # Registered buffer pool
├── examples/
│ ├── simple_echo.rs # Basic echo server
│ ├── split_echo.rs # Concurrent read/write
│ ├── axum_echo.rs # Axum integration
│ ├── http2_echo.rs # HTTP/2 WebSocket server
│ └── http3_echo.rs # HTTP/3 WebSocket server
├── autobahn/
│ ├── server.rs # Autobahn test server
│ └── Makefile # Build and test automation
└── benches/
└── throughput.rs # Criterion benchmarks
Performance Optimizations
- SIMD Masking: Uses AVX2/AVX-512/NEON to XOR mask frames at 32-64 bytes per cycle
- SIMD UTF-8: Validates UTF-8 text at memory bandwidth speeds
- Zero-Copy: Parses frames directly from receive buffer without copying
- Cork Buffer: Batches small writes into 16KB chunks for fewer syscalls
- Vectored I/O: Uses
writev()to send multiple buffers in single syscall - io_uring: Kernel-level async I/O with submission queue batching
License
MIT
Credits
- Inspired by uWebSockets
- HTTP/2 support via h2
- HTTP/3 support via quinn and h3
- io_uring support via tokio-uring