vectorless 0.1.2

Hierarchical, reasoning-native document intelligence engine
Documentation
# Vectorless Configuration Example
# Copy this file to config.toml and fill in your API keys

[indexer]
# Word count threshold for splitting sections into subsections
subsection_threshold = 300

# Maximum tokens to send in a single segmentation request
max_segment_tokens = 3000

# Maximum tokens for each summary
max_summary_tokens = 200

[summary]
# API key - get from your provider
# api_key = "sk-..."

# API endpoint
# OpenAI: https://api.openai.com/v1
# ZAI General: https://api.z.ai/api/paas/v4
# ZAI Coding: https://api.z.ai/api/coding/paas/v4
endpoint = "https://api.openai.com/v1"

# Model for summarization (use cheaper models for indexing)
model = "gpt-4o-mini"

# Maximum tokens for summary generation
max_tokens = 200

# Temperature for summary generation
temperature = 0.0

[retrieval]
# API key (optional, defaults to summary.api_key)
# api_key = "sk-..."

# API endpoint for retrieval
endpoint = "https://api.openai.com/v1"

# Model for retrieval navigation (use smarter models for better results)
model = "gpt-4o"

# Retriever type: llm_navigate, beam_search, mcts, multi_doc, hybrid
retriever_type = "llm_navigate"

# Number of top results to return
top_k = 3

# Maximum tokens for retrieval context
max_tokens = 1000

# Temperature for retrieval
temperature = 0.0

[storage]
# Workspace directory for persisted documents
#
# Structure:
# workspace/
# ├── _meta.json           # Lightweight index
# ├── {doc_id_1}.json      # Document 1
# └── {doc_id_2}.json      # Document 2
workspace_dir = "./workspace"