vipune /ˈʋi.pu.ne/
███
░░░
█████ █████ ████ ████████ █████ ████ ████████ ██████
░░███ ░░███ ░░███ ░░███░░███░░███ ░███ ░░███░░███ ███░░███
░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███████
░░███ ███ ░███ ░███ ░███ ░███ ░███ ░███ ░███ ░███░░░
░░█████ █████ ░███████ ░░████████ ████ █████░░██████
░░░░░ ░░░░░ ░███░░░ ░░░░░░░░ ░░░░ ░░░░░ ░░░░░░
░███
█████
░░░░░
A minimal memory layer for AI agents.
In Finnish mythology, Antero Vipunen is a giant who sleeps underground, holding all the world's knowledge and ancient songs. vipune is your agent's sleeping giant — a local knowledge store that remembers everything.
Store semantic memories, search by meaning, and detect conflicts. Single binary CLI. No API keys required.
Features
- Semantic search - Find memories by meaning, not keywords (ONNX embeddings, bge-small-en-v1.5)
- Conflict detection - Automatically warns when adding duplicate or similar memories
- Zero configuration - Works out of the box (auto-detected git projects, sensible defaults)
- Single binary - Just one CLI tool, no daemon, no database server
- No API keys - Everything runs locally, no network dependencies
- Project scoped - Memories isolated by git repository
Installation
Platform Support
Supported: macOS ARM64, Linux x86_64, Linux ARM64
Not supported: Windows (due to ONNX Runtime compilation complexity)
Quick install (macOS and Linux)
|
The installer detects your platform, downloads the correct binary, and adds it to
~/.cargo/bin/by default. If you don't use Rust, restart your shell or runexport PATH="$HOME/.cargo/bin:$PATH"to make vipune available. Use--prefix=/usr/localfor system-wide installs.
To verify the download before running, download the checksum file and check:
Pre-built binary
macOS Apple Silicon (arm64)
Download and extract:
&&
Linux x86_64
Download and extract:
&&
Linux ARM64
Download and extract:
&&
Build from source
Requires the Rust toolchain (1.85+). On Linux, you may also need libgomp1 and libc6.
Latest release (recommended)
Or clone and build manually
&&
The binary will be at ./target/release/vipune. Install it:
&&
Or add to PATH temporarily:
Uninstall
Or via cargo: cargo uninstall vipune
Optionally, clear all data:
Air-gapped / Offline Usage
vipune downloads the embedding model from HuggingFace Hub on first run. The model is pinned to a specific SHA to ensure reproducibility:
- Model ID:
BAAI/bge-small-en-v1.5 - Pinned revision SHA:
5c38ec7c405ec4b44b94cc5a9bb96e735b38267a
For air-gapped environments, pre-fetch the model before going offline:
# Install huggingface-cli first if you don't have it
# Download the pinned model to the cache directory
The model will be cached in the HF Hub cache layout at ~/.cache/huggingface/hub/ and vipune will use it without network access. You can verify the cache before going offline:
Note: When upgrading vipune, the pinned revision may change. Check the release notes and re-download the new revision if the SHA has changed.
Library consumers: The constants vipune::EMBED_MODEL_ID and vipune::EMBED_MODEL_REVISION are exported for tracking the pinned model version programmatically.
Quick Start
Add a memory:
Search by semantic meaning:
Add with metadata (optional):
CLI Commands
| Command | Description |
|---|---|
vipune add <text> |
Store a memory (with type, status, conflict detection) |
vipune search <query> |
Find memories by meaning (filters by type/status) |
vipune get <id> |
Retrieve a memory by ID |
vipune list |
List memories (default: active only) |
vipune delete <id> |
Delete a memory |
vipune update <id> [text] |
Update content, metadata, type, or status |
vipune validate <text> |
Check if text is within embedding token limits |
vipune version |
Show version |
Complete CLI reference • Quickstart guide • Search guide • Architecture
Library Usage
vipune can also be used as a Rust crate for programmatic integration:
# Cargo.toml
[]
= "0.3"
use ;
// Initialize memory store
let config = default;
let mut store = new.expect;
// Add a memory
let project_id = "my-project";
let memory_id = store.add
.expect;
// Search memories
let results = store.search
.expect;
for memory in results
v0.3 additions: MemoryType, MemoryStatus, and filter parameters are available for type-aware memory management. See the CLI reference for details.
See the crate documentation at docs.rs for complete API reference.
Configuration
vipune works with zero configuration. All paths use the user's home directory:
Default paths:
- Database:
~/.vipune/memories.db - Model cache:
~/.vipune/models/ - Config file:
~/.config/vipune/config.toml
Environment variables (override defaults):
VIPUNE_DATABASE_PATH- SQLite database locationVIPUNE_EMBEDDING_MODEL- HuggingFace model ID (default:BAAI/bge-small-en-v1.5)VIPUNE_MODEL_CACHE- Model download cache directoryVIPUNE_PROJECT- Project identifier (overrides auto-detection)VIPUNE_SIMILARITY_THRESHOLD- Conflict detection threshold, 0.0-1.0 (default:0.85)VIPUNE_RECENCY_WEIGHT- Recency bias in search results, 0.0-1.0 (default:0.3)
Config file (~/.config/vipune/config.toml):
= "~/.vipune/memories.db"
= "BAAI/bge-small-en-v1.5"
= "~/.vipune/models"
= 0.85
= 0.3
MCP Server
vipune can act as an MCP (Model Context Protocol) server, enabling AI agents like Claude Code and Cursor to use it as a native memory provider.
Note: MCP is enabled by default. Library users can use
default-features = falsefor sync-only builds.
Setup (Claude Code)
Add to your Claude Code configuration (~/.claude/settings.json or project .claude.json):
Setup (Cursor)
Add to your Cursor MCP configuration with the same JSON structure.
Available Tools
- store_memory: Store information for later recall
- search_memories: Find memories by meaning
- list_memories: List recent memories
Agent Integration
vipune works with any agent that can run shell commands — no plugins, adapters, or API keys required. Configure your agent with a few lines of instructions, grant shell command permissions, and the agent can use vipune search and vipune add to maintain persistent memory across tasks.
→ See Agent Integration Guide for per-tool setup instructions (Claude Code, Cursor, Windsurf, Cline, Roo Code, GitHub Copilot, Goose, Aider, OpenCode, Zed, and more).
See also: Search Guide for agent-friendly query patterns.
Exit codes for agent workflows:
0- Success1- Error (missing file, invalid input, etc.)2- Conflicts detected (similar memories found)3- Content too long (exceeds embedding token limit)
Recency Scoring
Search results can be weighted by recency using the --recency flag or VIPUNE_RECENCY_WEIGHT config:
# Increase recency bias (recent memories rank higher)
# Pure semantic similarity (no recency bias)
The final score combines semantic similarity and recency time decay:
score = (1 - recency_weight) * similarity + recency_weight * time_score- Default balance: 70% semantic, 30% recency
License
Apache-2.0 © Janni Turunen