squid-rs 0.6.0

An AI-powered command-line tool for code reviews and suggestions.
# =============================================================================
# squid Configuration Example
# =============================================================================
# Copy this file to .env and update with your settings
# Choose ONE of the configurations below based on your LLM provider

# -----------------------------------------------------------------------------
# Logging Configuration
# -----------------------------------------------------------------------------
LOG_LEVEL=error

# =============================================================================
# OPTION 1: LM Studio (Recommended for Local Development)
# =============================================================================
# 1. Download LM Studio from https://lmstudio.ai/
# 2. Download model: lmstudio-community/Qwen2.5-Coder-7B-Instruct-MLX-4bit
# 3. Load the model in LM Studio
# 4. Start local server (↔️ icon → "Start Server")
# 5. Use these settings:

API_URL=http://127.0.0.1:1234/v1
API_MODEL=local-model
API_KEY=not-needed

# -----------------------------------------------------------------------------
# OPTION 2: Ollama (Lightweight CLI Alternative)
# -----------------------------------------------------------------------------
# 1. Install: brew install ollama (macOS) or see https://ollama.com/
# 2. Start service: ollama serve
# 3. Pull model: ollama pull qwen2.5-coder
# 4. Use these settings:

# API_URL=http://localhost:11434/v1
# API_MODEL=qwen2.5-coder
# API_KEY=not-needed

# Alternative Ollama models:
# API_MODEL=codellama          # Code generation focused
# API_MODEL=deepseek-coder     # Code understanding
# API_MODEL=llama3.1           # General purpose

# -----------------------------------------------------------------------------
# OPTION 3: OpenAI API (Cloud Service)
# -----------------------------------------------------------------------------
# 1. Get API key from https://platform.openai.com/api-keys
# 2. Add credits to your account
# 3. Use these settings:

# API_URL=https://api.openai.com/v1
# API_MODEL=gpt-4
# API_KEY=sk-your-actual-api-key-here

# Alternative OpenAI models:
# API_MODEL=gpt-4-turbo        # Faster, cheaper GPT-4
# API_MODEL=gpt-3.5-turbo      # Fastest, cheapest

# -----------------------------------------------------------------------------
# OPTION 4: Mistral API (Cloud Service)
# -----------------------------------------------------------------------------
# 1. Get API key from https://console.mistral.ai/
# 2. Choose a model (devstral-2512, mistral-large-latest, etc.)
# 3. Use these settings:

# API_URL=https://api.mistral.ai/v1
# API_MODEL=devstral-2512
# API_KEY=your-mistral-api-key-here

# Alternative Mistral models:
# API_MODEL=mistral-large-latest    # Most capable
# API_MODEL=mistral-small-latest    # Fast and efficient
# API_MODEL=codestral-latest        # Code-specialized

# -----------------------------------------------------------------------------
# OPTION 5: Other OpenAI-Compatible Services
# -----------------------------------------------------------------------------

# OpenRouter (https://openrouter.ai/)
# API_URL=https://openrouter.ai/api/v1
# API_MODEL=anthropic/claude-3-sonnet
# API_KEY=your-openrouter-key

# Together AI (https://together.ai/)
# API_URL=https://api.together.xyz/v1
# API_MODEL=codellama/CodeLlama-34b-Instruct-hf
# API_KEY=your-together-ai-key

# Custom OpenAI-compatible endpoint
# API_URL=http://your-custom-endpoint:port/v1
# API_MODEL=your-model-name
# API_KEY=your-api-key-if-needed

# =============================================================================
# Notes
# =============================================================================
# - Only ONE configuration should be active (uncomment the one you want)
# - LM Studio and Ollama don't require API keys (use "not-needed")
# - For best code-related results, use qwen2.5-coder model
# - Streaming is supported with all providers (use -s or --stream flag)