check:
cargo check
cargo build
cargo test
# ===== BENCHMARKING =====
# Run all benchmarks
bench:
cargo bench --bench tensor_ops --
cargo bench --bench neural_networks --
cargo bench --bench gpu_comparison --
# Run specific benchmark
bench-name name:
cargo bench --bench {{name}} --
# Benchmark CPU only (no features)
bench-cpu:
cargo bench --bench tensor_ops --no-default-features --
cargo bench --bench neural_networks --no-default-features --
cargo bench --bench gpu_comparison --no-default-features --
# Benchmark with Accelerate (macOS BLAS)
bench-accel:
cargo bench --features accelerate --no-default-features --bench tensor_ops --
cargo bench --features accelerate --no-default-features --bench neural_networks --
cargo bench --features accelerate --no-default-features --bench gpu_comparison --
# Benchmark GPU comparison
bench-gpu:
cargo bench --features gpu --bench gpu_comparison --
# Save benchmark results for comparison
bench-save:
cargo bench --bench tensor_ops -- --save-baseline main
cargo bench --bench neural_networks -- --save-baseline main
cargo bench --bench gpu_comparison -- --save-baseline main
# Compare against saved baseline
bench-compare:
cargo bench --bench tensor_ops -- --baseline main
cargo bench --bench neural_networks -- --baseline main
cargo bench --bench gpu_comparison -- --baseline main
bench-report:
open './target/criterion/report/index.html'
# ===== LLM ASSISTED DEV =====
ask-gpu model:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/claude-on-gpu-integration.md responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "`cat sys-prompts/gpu-sys-prompt.md`" -o reasoning_effort high > responses/gpu-recommendations.md
nvim responses/gpu-recommendations.md
ask-codex:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/openai/gpt-5.1-codex" --system "`cat sys-prompts/sys-prompt.md`" -o reasoning_effort high > responses/recommendations.md
@echo "Finished!"
ask-codex-mini:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/openai/gpt-5.1-codex-mini" --system "`cat sys-prompts/sys-prompt.md`" -o reasoning_effort high > responses/recommendations.md
@echo "Finished!"
ask-readme model:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "`cat sys-prompts/readme-sys-prompt.md`" > responses/updates-to-readme.md
@echo "Finished!"
ask-custom model prompt filename:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "{{prompt}}" > responses/{{filename}}.md
@echo "Finished!"
ask-err-file model file:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
-rm err-recommendations.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/{{file}} tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "`cat sys-prompts/err-sys-prompt.md`" > responses/err-recommendations.md
@echo "Finished!"
ask-err model:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
-rm responses/err-recommendations.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "`cat sys-prompts/err-sys-prompt.md`" > responses/err-recommendations.md
@echo "Finished!"
ask-status model:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "`cat sys-prompts/status-sys-prompt.md`" > responses/status-report.md
@echo "Finished!"
ask model:
@echo "Cleaning!"
-rm err.txt tests.txt
-rm responses/context.md
@echo "Asking!"
-cargo build &> err.txt
-cargo test &> tests.txt
files-to-prompt Cargo.toml src/ tests/ err.txt tests.txt responses/status-report.md README.md > responses/context.md
cat responses/context.md | llm --model "openrouter/{{model}}" --system "`cat sys-prompts/sys-prompt.md`" > responses/recommendations.md
@echo "Finished!"