.PHONY: help build test mcp-test fmt clippy doc mdbook paper clean coverage rust-export compare qubo-testdata export-schemas release run-plan run-issue run-pipeline run-pipeline-forever run-review run-review-forever board-next board-claim board-ack board-move issue-context issue-guards pr-context pr-wait-ci worktree-issue worktree-pr diagrams jl-testdata cli cli-demo copilot-review
RUNNER ?= codex
CLAUDE_MODEL ?= opus
CODEX_MODEL ?= gpt-5.4
help:
@echo "Available targets:"
@echo " build - Build the project"
@echo " test - Run all tests"
@echo " mcp-test - Run MCP server tests"
@echo " fmt - Format code with rustfmt"
@echo " fmt-check - Check code formatting"
@echo " clippy - Run clippy lints"
@echo " doc - Build mdBook documentation"
@echo " diagrams - Generate SVG diagrams from Typst (light + dark)"
@echo " mdbook - Build and serve mdBook (with live reload)"
@echo " paper - Build Typst paper from checked-in fixtures (requires typst)"
@echo " coverage - Generate coverage report (requires cargo-llvm-cov)"
@echo " clean - Clean build artifacts"
@echo " check - Quick check (fmt + clippy + test)"
@echo " rust-export - Generate Rust mapping JSON exports"
@echo " compare - Generate and compare Rust mapping exports"
@echo " export-schemas - Export problem schemas to JSON"
@echo " qubo-testdata - Regenerate QUBO test data (requires uv)"
@echo " jl-testdata - Regenerate Julia parity test data (requires julia)"
@echo " release V=x.y.z - Tag and push a new release (triggers CI publish)"
@echo " cli - Build the pred CLI tool"
@echo " cli-demo - Run closed-loop CLI demo (build + exercise all commands)"
@echo " run-plan - Execute a plan with Codex or Claude (latest plan in docs/plans/)"
@echo " run-issue N=<number> - Run issue-to-pr --execute for a GitHub issue"
@echo " run-pipeline [N=<number>] - Pick a Ready issue, implement, move to Review pool"
@echo " run-pipeline-forever - Loop: poll Ready column for new issues, run-pipeline when new ones appear"
@echo " run-review [N=<number>] - Pick PR from Review pool, fix comments/CI, run agentic tests"
@echo " run-review-forever - Loop: poll Review pool for eligible PRs, dispatch run-review"
@echo " board-next MODE=<ready|review|final-review> [NUMBER=<n>] [FORMAT=text|json] - Get the next eligible queued project item"
@echo " board-claim MODE=<ready|review> [NUMBER=<n>] [FORMAT=text|json] - Claim and move the next eligible queued project item"
@echo " board-ack MODE=<ready|review|final-review> ITEM=<id> - Acknowledge a queued project item"
@echo " board-move ITEM=<id> STATUS=<status> - Move a project item to a named status"
@echo " issue-context ISSUE=<number> [REPO=<owner/repo>] - Fetch structured issue preflight JSON"
@echo " issue-guards ISSUE=<number> [REPO=<owner/repo>] - Backward-compatible alias for issue-context"
@echo " pr-context PR=<number> [REPO=<owner/repo>] - Fetch structured PR snapshot JSON"
@echo " pr-wait-ci PR=<number> [REPO=<owner/repo>] - Poll CI until terminal state and print JSON"
@echo " worktree-issue ISSUE=<number> SLUG=<slug> - Create an issue worktree from origin/main"
@echo " worktree-pr PR=<number> [REPO=<owner/repo>] - Checkout a PR into an isolated worktree"
@echo " copilot-review - Request Copilot code review on current PR"
@echo ""
@echo " Set RUNNER=claude to use Claude instead of Codex (default: codex)"
@echo " Override CODEX_MODEL or CLAUDE_MODEL to pick a different model"
build:
cargo build --features ilp-highs
test:
cargo test --features "ilp-highs example-db" -- --include-ignored
mcp-test:
cargo test --features mcp -p problemreductions-cli mcp
fmt:
cargo fmt --all
fmt-check:
cargo fmt --all -- --check
clippy:
cargo clippy --all-targets --features ilp-highs -- -D warnings
doc:
cargo run --example export_graph
cargo run --example export_schemas
mdbook build docs
RUSTDOCFLAGS="--default-theme=dark" cargo doc --features ilp-highs --no-deps
rm -rf docs/book/api
cp -r target/doc docs/book/api
TYPST_DOC_DIAGRAMS := $(wildcard docs/src/static/*.typ)
TYPST_PAPER_DIAGRAMS := $(wildcard docs/paper/static/*.typ)
diagrams:
@for src in $(TYPST_DOC_DIAGRAMS); do \
base=$$(basename $$src .typ); \
echo "Compiling $$base (doc)..."; \
typst compile $$src --root=. --input dark=false docs/src/static/$$base.svg; \
typst compile $$src --root=. --input dark=true docs/src/static/$$base-dark.svg; \
done
mdbook:
@echo "Exporting graph..."
@cargo run --example export_graph 2>&1 | tail -1
@echo "Exporting schemas..."
@cargo run --example export_schemas 2>&1 | tail -1
@echo "Building API docs..."
@RUSTDOCFLAGS="--default-theme=dark" cargo doc --features ilp-highs --no-deps 2>&1 | tail -1
@echo "Building mdBook..."
@mdbook build
rm -rf book/api
cp -r target/doc book/api
@-lsof -ti:3001 | xargs kill 2>/dev/null || true
@echo "Serving at http://localhost:3001"
python3 -m http.server 3001 -d book &
@sleep 1 && (command -v xdg-open >/dev/null && xdg-open http://localhost:3001 || open http://localhost:3001)
export-schemas:
cargo run --example export_schemas
paper:
cargo run --features "example-db" --example export_examples
cargo run --example export_petersen_mapping
cargo run --example export_graph
cargo run --example export_schemas
typst compile --root . docs/paper/reductions.typ docs/paper/reductions.pdf
coverage:
@command -v cargo-llvm-cov >/dev/null 2>&1 || { echo "Installing cargo-llvm-cov..."; cargo install cargo-llvm-cov; }
cargo llvm-cov --features ilp-highs --workspace --html --open
clean:
cargo clean
check: fmt-check clippy test
@echo "✅ All checks passed!"
qubo-testdata:
cd scripts && uv run python generate_qubo_tests.py
jl-testdata:
cd scripts/jl && julia --project=. generate_testdata.jl
release:
ifndef V
$(error Usage: make release V=x.y.z)
endif
@echo "Releasing v$(V)..."
sed -i 's/^version = ".*"/version = "$(V)"/' Cargo.toml
sed -i 's/^version = ".*"/version = "$(V)"/' problemreductions-macros/Cargo.toml
sed -i 's/^version = ".*"/version = "$(V)"/' problemreductions-cli/Cargo.toml
sed -i 's/problemreductions-macros = { version = "[^"]*"/problemreductions-macros = { version = "$(V)"/' Cargo.toml
sed -i 's/problemreductions = { version = "[^"]*"/problemreductions = { version = "$(V)"/' problemreductions-cli/Cargo.toml
cargo check
git add Cargo.toml problemreductions-macros/Cargo.toml problemreductions-cli/Cargo.toml
git commit -m "release: v$(V)"
git tag -a "v$(V)" -m "Release v$(V)"
git push origin main --tags
@echo "v$(V) pushed — CI will publish to crates.io"
cli:
cargo install --path problemreductions-cli
mcp:
cargo install --path problemreductions-cli --features mcp
GRAPHS := diamond bull house petersen
MODES := unweighted weighted triangular
rust-export:
@mkdir -p tests/julia
@for graph in $(GRAPHS); do \
for mode in $(MODES); do \
echo "Exporting $$graph ($$mode)..."; \
cargo run --example export_mapping_stages -- $$graph $$mode; \
done; \
done
compare: rust-export
@echo ""
@echo "=== Julia vs Rust Comparison ==="
@for graph in $(GRAPHS); do \
echo ""; \
echo "=== $$graph ==="; \
echo "-- unweighted --"; \
echo "Julia: $$(jq '{nodes: .num_grid_nodes, overhead: .mis_overhead, tape: .num_tape_entries}' tests/data/$${graph}_unweighted_trace.json)"; \
echo "Rust: $$(jq '{nodes: .stages[3].num_nodes, overhead: .total_overhead, tape: ((.crossing_tape | length) + (.simplifier_tape | length))}' tests/data/$${graph}_rust_unweighted.json)"; \
echo "-- weighted --"; \
echo "Julia: $$(jq '{nodes: .num_grid_nodes, overhead: .mis_overhead, tape: .num_tape_entries}' tests/data/$${graph}_weighted_trace.json)"; \
echo "Rust: $$(jq '{nodes: .stages[3].num_nodes, overhead: .total_overhead, tape: ((.crossing_tape | length) + (.simplifier_tape | length))}' tests/data/$${graph}_rust_weighted.json)"; \
echo "-- triangular --"; \
echo "Julia: $$(jq '{nodes: .num_grid_nodes, overhead: .mis_overhead, tape: .num_tape_entries}' tests/data/$${graph}_triangular_trace.json)"; \
echo "Rust: $$(jq '{nodes: .stages[3].num_nodes, overhead: .total_overhead, tape: ((.crossing_tape | length) + (.simplifier_tape | length))}' tests/data/$${graph}_rust_triangular.json)"; \
done
INSTRUCTIONS ?=
OUTPUT ?= run-plan-output.log
AGENT_TYPE ?= $(RUNNER)
PLAN_FILE ?= $(shell ls -t docs/plans/*.md 2>/dev/null | head -1)
run-plan:
@. scripts/make_helpers.sh; \
NL=$$'\n'; \
BRANCH=$$(git branch --show-current); \
PLAN_FILE="$(PLAN_FILE)"; \
if [ "$(AGENT_TYPE)" = "claude" ]; then \
PROCESS="1. Read the plan file$${NL}2. Execute the plan — it specifies which skill(s) to use$${NL}3. Push: git push origin $$BRANCH$${NL}4. If a PR already exists for this branch, skip. Otherwise create one."; \
else \
PROCESS="1. Read the plan file$${NL}2. If the plan references repo-local workflow docs under .claude/skills/*/SKILL.md, open and follow them directly. Treat slash-command names as aliases for those files.$${NL}3. Execute the tasks step by step. For each task, implement and test before moving on.$${NL}4. Push: git push origin $$BRANCH$${NL}5. If a PR already exists for this branch, skip. Otherwise create one."; \
fi; \
PROMPT="Execute the plan in '$$PLAN_FILE'."; \
if [ "$(AGENT_TYPE)" != "claude" ]; then \
PROMPT="$${PROMPT}$${NL}$${NL}Repo-local skills live in .claude/skills/*/SKILL.md. Treat any slash-command references in the plan as aliases for those skill files."; \
fi; \
if [ -n "$(INSTRUCTIONS)" ]; then \
PROMPT="$${PROMPT}$${NL}$${NL}## Additional Instructions$${NL}$(INSTRUCTIONS)"; \
fi; \
PROMPT="$${PROMPT}$${NL}$${NL}## Process$${NL}$${PROCESS}$${NL}$${NL}## Rules$${NL}- Tests should be strong enough to catch regressions.$${NL}- Do not modify tests to make them pass.$${NL}- Test failure must be reported."; \
echo "=== Prompt ===" && echo "$$PROMPT" && echo "===" ; \
RUNNER="$(AGENT_TYPE)" run_agent "$(OUTPUT)" "$$PROMPT"
N ?=
run-issue:
@. scripts/make_helpers.sh; \
if [ -z "$(N)" ]; then echo "Usage: make run-issue N=<issue-number>"; exit 1; fi; \
PROMPT=$$(skill_prompt issue-to-pr "/issue-to-pr $(N) --execute" "process GitHub issue $(N) with --execute behavior"); \
run_agent "issue-$(N)-output.log" "$$PROMPT"
PRED := cargo run -p problemreductions-cli --release --
CLI_DEMO_DIR := /tmp/pred-cli-demo
cli-demo: cli
@echo "=== pred CLI closed-loop demo ==="
@rm -rf $(CLI_DEMO_DIR) && mkdir -p $(CLI_DEMO_DIR)
@set -e; \
PRED="./target/release/pred"; \
\
echo ""; \
echo "--- 1. list: all registered problems ---"; \
$$PRED list; \
$$PRED list -o $(CLI_DEMO_DIR)/problems.json; \
\
echo ""; \
echo "--- 2. show: inspect MIS (variants, fields, reductions) ---"; \
$$PRED show MIS; \
$$PRED show MIS -o $(CLI_DEMO_DIR)/mis_info.json; \
\
echo ""; \
echo "--- 3. to: explore 2-hop outgoing neighborhood ---"; \
$$PRED to MIS --hops 2; \
$$PRED to MIS --hops 2 -o $(CLI_DEMO_DIR)/mis_hops.json; \
\
echo ""; \
echo "--- 4. from: incoming neighbors ---"; \
$$PRED from QUBO --hops 1; \
\
echo ""; \
echo "--- 5. path: find reduction paths ---"; \
$$PRED path MIS QUBO; \
$$PRED path MIS QUBO -o $(CLI_DEMO_DIR)/path_mis_qubo.json; \
$$PRED path Factoring SpinGlass; \
$$PRED path MIS QUBO --cost minimize:num_variables; \
\
echo ""; \
echo "--- 6. path --all: enumerate all paths ---"; \
$$PRED path MIS QUBO --all; \
$$PRED path MIS QUBO --all -o $(CLI_DEMO_DIR)/all_paths/; \
\
echo ""; \
echo "--- 7. export-graph: full reduction graph ---"; \
$$PRED export-graph -o $(CLI_DEMO_DIR)/graph.json; \
\
echo ""; \
echo "--- 8. create: build problem instances ---"; \
$$PRED create MIS --graph 0-1,1-2,2-3,3-4,4-0 -o $(CLI_DEMO_DIR)/mis.json; \
$$PRED create MIS --graph 0-1,1-2,2-3 --weights 2,1,3,1 -o $(CLI_DEMO_DIR)/mis_weighted.json; \
$$PRED create SAT --num-vars 3 --clauses "1,2;-1,3;2,-3" -o $(CLI_DEMO_DIR)/sat.json; \
$$PRED create 3SAT --num-vars 4 --clauses "1,2,3;-1,2,-3;1,-2,3" -o $(CLI_DEMO_DIR)/3sat.json; \
$$PRED create QUBO --matrix "1,-0.5;-0.5,2" -o $(CLI_DEMO_DIR)/qubo.json; \
$$PRED create KColoring --k 3 --graph 0-1,1-2,2-0 -o $(CLI_DEMO_DIR)/kcol.json; \
$$PRED create SpinGlass --graph 0-1,1-2 -o $(CLI_DEMO_DIR)/sg.json; \
$$PRED create MaxCut --graph 0-1,1-2,2-0 -o $(CLI_DEMO_DIR)/maxcut.json; \
$$PRED create MVC --graph 0-1,1-2,2-3 -o $(CLI_DEMO_DIR)/mvc.json; \
$$PRED create MaximumMatching --graph 0-1,1-2,2-3 -o $(CLI_DEMO_DIR)/matching.json; \
$$PRED create Factoring --target 15 --m 4 --n 4 -o $(CLI_DEMO_DIR)/factoring.json; \
$$PRED create Factoring --target 21 --m 3 --n 3 -o $(CLI_DEMO_DIR)/factoring2.json; \
\
echo ""; \
echo "--- 9. evaluate: test configurations ---"; \
$$PRED evaluate $(CLI_DEMO_DIR)/mis.json --config 1,0,1,0,0; \
$$PRED evaluate $(CLI_DEMO_DIR)/mis.json --config 1,1,0,0,0; \
$$PRED evaluate $(CLI_DEMO_DIR)/sat.json --config 0,1,1; \
$$PRED evaluate $(CLI_DEMO_DIR)/mis.json --config 1,0,1,0,0 -o $(CLI_DEMO_DIR)/eval.json; \
\
echo ""; \
echo "--- 10. solve: direct ILP (auto-reduces to ILP) ---"; \
$$PRED solve $(CLI_DEMO_DIR)/mis.json; \
$$PRED solve $(CLI_DEMO_DIR)/mis.json -o $(CLI_DEMO_DIR)/sol_ilp.json; \
\
echo ""; \
echo "--- 11. solve: brute-force ---"; \
$$PRED solve $(CLI_DEMO_DIR)/mis.json --solver brute-force; \
\
echo ""; \
echo "--- 12. solve: weighted MIS ---"; \
$$PRED solve $(CLI_DEMO_DIR)/mis_weighted.json; \
\
echo ""; \
echo "--- 13. reduce: MIS → QUBO (auto-discover path) ---"; \
$$PRED reduce $(CLI_DEMO_DIR)/mis.json --to QUBO -o $(CLI_DEMO_DIR)/bundle_qubo.json; \
\
echo ""; \
echo "--- 14. solve bundle: brute-force on reduced QUBO ---"; \
$$PRED solve $(CLI_DEMO_DIR)/bundle_qubo.json --solver brute-force; \
\
echo ""; \
echo "--- 15. reduce --via: use explicit path file ---"; \
$$PRED reduce $(CLI_DEMO_DIR)/mis.json --via $(CLI_DEMO_DIR)/path_mis_qubo.json -o $(CLI_DEMO_DIR)/bundle_via.json; \
\
echo ""; \
echo "--- 16. solve bundle with ILP: MIS → MVC → ILP ---"; \
$$PRED reduce $(CLI_DEMO_DIR)/mis.json --to MVC -o $(CLI_DEMO_DIR)/bundle_mvc.json; \
$$PRED solve $(CLI_DEMO_DIR)/bundle_mvc.json --solver ilp; \
\
echo ""; \
echo "--- 17. solve: other problem types ---"; \
$$PRED solve $(CLI_DEMO_DIR)/sat.json --solver brute-force; \
$$PRED solve $(CLI_DEMO_DIR)/kcol.json --solver brute-force; \
$$PRED solve $(CLI_DEMO_DIR)/maxcut.json --solver brute-force; \
$$PRED solve $(CLI_DEMO_DIR)/mvc.json; \
\
echo ""; \
echo "--- 18. closed-loop: create → reduce → solve → verify ---"; \
echo "Creating a 6-vertex graph..."; \
$$PRED create MIS --graph 0-1,1-2,2-3,3-4,4-5,0-5,1-4 -o $(CLI_DEMO_DIR)/big.json; \
echo "Solving with ILP..."; \
$$PRED solve $(CLI_DEMO_DIR)/big.json -o $(CLI_DEMO_DIR)/big_sol.json; \
echo "Reducing to QUBO and solving with brute-force..."; \
$$PRED reduce $(CLI_DEMO_DIR)/big.json --to QUBO -o $(CLI_DEMO_DIR)/big_qubo.json; \
$$PRED solve $(CLI_DEMO_DIR)/big_qubo.json --solver brute-force -o $(CLI_DEMO_DIR)/big_qubo_sol.json; \
echo "Verifying both solutions have the same evaluation..."; \
ILP_EVAL=$$(jq -r '.evaluation' $(CLI_DEMO_DIR)/big_sol.json); \
BF_EVAL=$$(jq -r '.evaluation' $(CLI_DEMO_DIR)/big_qubo_sol.json); \
echo " ILP solution evaluation: $$ILP_EVAL"; \
echo " Brute-force (via QUBO) evaluation: $$BF_EVAL"; \
if [ "$$ILP_EVAL" = "$$BF_EVAL" ]; then \
echo " ✅ Solutions agree!"; \
else \
echo " ❌ Solutions disagree!" && exit 1; \
fi; \
\
echo ""; \
echo "--- 19. show with alias and variant slash syntax ---"; \
$$PRED show MIS/UnitDiskGraph; \
\
echo ""; \
echo "--- 20. completions: generate shell completions ---"; \
$$PRED completions bash > /dev/null && echo "bash completions: OK"; \
$$PRED completions zsh > /dev/null && echo "zsh completions: OK"; \
$$PRED completions fish > /dev/null && echo "fish completions: OK"; \
\
echo ""; \
echo "=== Demo complete: $$(ls $(CLI_DEMO_DIR)/*.json | wc -l | tr -d ' ') JSON files in $(CLI_DEMO_DIR) ==="
@echo "=== All 20 steps passed ✅ ==="
run-pipeline:
@. scripts/make_helpers.sh; \
if [ -n "$(N)" ]; then \
issue="$(N)"; \
else \
status=0; \
tmp_state=$$(mktemp); \
selection=$$(board_next_json ready "" "" "$$tmp_state") || status=$$?; \
rm -f "$$tmp_state"; \
if [ "$$status" -eq 1 ]; then \
echo "No Ready issues are currently eligible."; \
exit 1; \
elif [ "$$status" -ne 0 ]; then \
exit "$$status"; \
fi; \
issue=$$(printf '%s\n' "$$selection" | python3 -c "import sys,json; data=json.load(sys.stdin); print(data['issue_number'] or data['number'])"); \
fi; \
PROMPT=$$(skill_prompt_with_context project-pipeline "/project-pipeline $$issue" "process GitHub issue $$issue" "Selected queue item" "$$selection"); \
run_agent "pipeline-output.log" "$$PROMPT"
run-pipeline-forever:
@. scripts/make_helpers.sh; \
MAKE=$(MAKE) watch_and_dispatch ready run-pipeline "Ready issues"
board-next:
@if [ -z "$(MODE)" ]; then \
echo "MODE=ready|review|final-review is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
state_file=$${STATE_FILE:-/tmp/problemreductions-$(MODE)-state.json}; \
case "$(MODE)" in \
review|final-review) \
repo=$${REPO:-$$(gh repo view --json nameWithOwner --jq .nameWithOwner)}; \
poll_project_items "$(MODE)" "$$state_file" "$$repo" "$(NUMBER)" "$(if $(FORMAT),$(FORMAT),text)"; \
;; \
*) \
poll_project_items "$(MODE)" "$$state_file" "" "$(NUMBER)" "$(if $(FORMAT),$(FORMAT),text)"; \
;; \
esac
board-claim:
@if [ -z "$(MODE)" ]; then \
echo "MODE=ready|review is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
state_file=$${STATE_FILE:-/tmp/problemreductions-$(MODE)-state.json}; \
case "$(MODE)" in \
review) \
repo=$${REPO:-$$(gh repo view --json nameWithOwner --jq .nameWithOwner)}; \
claim_project_items "$(MODE)" "$$state_file" "$$repo" "$(NUMBER)" "$(if $(FORMAT),$(FORMAT),json)"; \
;; \
ready) \
claim_project_items "$(MODE)" "$$state_file" "" "$(NUMBER)" "$(if $(FORMAT),$(FORMAT),json)"; \
;; \
*) \
echo "MODE=ready|review is required"; \
exit 2; \
;; \
esac
board-ack:
@if [ -z "$(MODE)" ] || [ -z "$(ITEM)" ]; then \
echo "MODE=ready|review|final-review and ITEM=<project-item-id> are required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
state_file=$${STATE_FILE:-/tmp/problemreductions-$(MODE)-state.json}; \
ack_polled_item "$$state_file" "$(ITEM)"
board-move:
@if [ -z "$(ITEM)" ] || [ -z "$(STATUS)" ]; then \
echo "ITEM=<project-item-id> and STATUS=<backlog|ready|in-progress|review-pool|under-review|final-review|on-hold|done> are required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
move_board_item "$(ITEM)" "$(STATUS)"
issue-context:
@if [ -z "$(ISSUE)" ]; then \
echo "ISSUE=<number> is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
repo=$${REPO:-CodingThrust/problem-reductions}; \
issue_context "$$repo" "$(ISSUE)"
issue-guards:
@if [ -z "$(ISSUE)" ]; then \
echo "ISSUE=<number> is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
repo=$${REPO:-CodingThrust/problem-reductions}; \
issue_guards "$$repo" "$(ISSUE)"
pr-context:
@if [ -z "$(PR)" ]; then \
echo "PR=<number> is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
repo=$${REPO:-$$(gh repo view --json nameWithOwner --jq .nameWithOwner)}; \
pr_snapshot "$$repo" "$(PR)"
pr-wait-ci:
@if [ -z "$(PR)" ]; then \
echo "PR=<number> is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
repo=$${REPO:-$$(gh repo view --json nameWithOwner --jq .nameWithOwner)}; \
timeout=$${TIMEOUT:-900}; \
interval=$${INTERVAL:-30}; \
pr_wait_ci "$$repo" "$(PR)" "$$timeout" "$$interval"
worktree-issue:
@if [ -z "$(ISSUE)" ] || [ -z "$(SLUG)" ]; then \
echo "ISSUE=<number> and SLUG=<slug> are required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
base=$${BASE:-origin/main}; \
create_issue_worktree "$(ISSUE)" "$(SLUG)" "$$base"
worktree-pr:
@if [ -z "$(PR)" ]; then \
echo "PR=<number> is required"; \
exit 2; \
fi
@. scripts/make_helpers.sh; \
repo=$${REPO:-$$(gh repo view --json nameWithOwner --jq .nameWithOwner)}; \
checkout_pr_worktree "$$repo" "$(PR)"
run-review:
@. scripts/make_helpers.sh; \
repo=$${REPO:-$$(gh repo view --json nameWithOwner --jq .nameWithOwner)}; \
pr="$(N)"; \
selection=$$(review_pipeline_context "$$repo" "$$pr"); \
status_name=$$(printf '%s\n' "$$selection" | python3 -c "import sys,json; print(json.load(sys.stdin)['status'])"); \
if [ "$$status_name" = "empty" ]; then \
echo "No Review pool PRs are currently eligible."; \
exit 1; \
fi; \
if [ "$$status_name" = "ready" ]; then \
pr=$$(printf '%s\n' "$$selection" | python3 -c "import sys,json; print(json.load(sys.stdin)['selection']['pr_number'])"); \
slash_cmd="/review-pipeline $$pr"; \
codex_desc="process PR else \
slash_cmd="/review-pipeline"; \
codex_desc="inspect the review pipeline bundle and resolve the next action"; \
fi; \
PROMPT=$$(skill_prompt_with_context review-pipeline "$$slash_cmd" "$$codex_desc" "Review pipeline context" "$$selection"); \
run_agent "review-output.log" "$$PROMPT"
run-review-forever:
@. scripts/make_helpers.sh; \
REPO=$$(gh repo view --json nameWithOwner --jq .nameWithOwner); \
MAKE=$(MAKE) watch_and_dispatch review run-review "Review pool PRs" "$$REPO"
copilot-review:
@PR=$$(gh pr view --json number --jq .number 2>/dev/null) || { echo "No PR found for current branch"; exit 1; }; \
echo "Requesting Copilot review on PR gh copilot-review $$PR