inference-lab 0.6.2

High-performance LLM inference simulator for analyzing serving systems
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
/target
/pkg
web/node_modules
web/dist
.npmrc

# Large test datasets and models (user-specific)
wildchat_*.jsonl
tokenizer.json
*.gz

# User test configs (not examples)
test_*.toml
wildchat_*.toml
qwen_chat_template.txt

# mdBook output
docs/book/