Skip to main content

apm/cmd/
work.rs

1use anyhow::Result;
2use apm_core::{config::Config, ticket};
3use std::path::Path;
4use std::sync::atomic::{AtomicUsize, Ordering};
5use std::sync::Arc;
6use std::time::{Duration, Instant};
7
8fn log(msg: &str) {
9    let ts = chrono::Local::now().format("%H:%M:%S");
10    println!("[{ts}] {msg}");
11}
12
13pub fn run(root: &Path, skip_permissions: bool, dry_run: bool, daemon: bool, interval_secs: u64, epic: Option<String>) -> Result<()> {
14    if daemon && dry_run {
15        anyhow::bail!("--daemon and --dry-run cannot be used together");
16    }
17
18    let config = Config::load(root)?;
19    let max_concurrent = config.agents.max_concurrent.max(1);
20    let epic_filter: Option<String> = epic.or_else(|| config.work.epic.clone());
21
22    if dry_run {
23        return run_dry(root, &config, epic_filter.as_deref());
24    }
25
26    let sig_count = Arc::new(AtomicUsize::new(0));
27    let sig_count_clone = Arc::clone(&sig_count);
28    let _ = ctrlc::set_handler(move || {
29        sig_count_clone.fetch_add(1, Ordering::Relaxed);
30    });
31
32    let mut workers: Vec<(String, Option<String>, apm_core::start::ManagedChild, std::path::PathBuf)> = Vec::new();
33    let mut started_ids: Vec<String> = Vec::new();
34    let mut no_more = false;
35    // next_poll only used in daemon mode
36    let mut next_poll = Instant::now();
37    let mut drain_announced = false;
38
39    loop {
40        let sigs = sig_count.load(Ordering::Relaxed);
41
42        if daemon {
43            if sigs >= 2 {
44                log(&format!("Forced exit; {} worker(s) may still be running", workers.len()));
45                break;
46            }
47            if sigs == 1 {
48                if workers.is_empty() {
49                    log("Daemon stopped.");
50                    break;
51                }
52                if !drain_announced {
53                    log(&format!(
54                        "Graceful shutdown: waiting for {} worker(s) to finish (Ctrl+C again to exit immediately)",
55                        workers.len()
56                    ));
57                    drain_announced = true;
58                }
59                // Reap finished workers during drain.
60                workers.retain_mut(|(id, _epic_id, child, _pid_path)| {
61                    let done = matches!(child.try_wait(), Ok(Some(_)));
62                    if done {
63                        log(&format!("Worker for ticket #{id} finished"));
64                    }
65                    !done
66                });
67                if workers.is_empty() {
68                    log("All workers finished; exiting.");
69                    break;
70                }
71                std::thread::sleep(Duration::from_millis(500));
72                continue;
73            }
74        } else if sigs >= 1 {
75            break;
76        }
77
78        // Reap finished workers.
79        let mut reaped = false;
80        workers.retain_mut(|(id, _epic_id, child, _pid_path)| {
81            let done = matches!(child.try_wait(), Ok(Some(_)));
82            if done {
83                log(&format!("Worker for ticket #{id} finished"));
84                reaped = true;
85            }
86            !done
87        });
88
89        // In daemon mode: a reaped worker opens a slot — check immediately.
90        if daemon && reaped {
91            next_poll = Instant::now();
92            no_more = false;
93        }
94
95        if !daemon && no_more && workers.is_empty() {
96            break;
97        }
98
99        // In daemon mode: if no_more and not yet time to poll again, sleep and continue.
100        if daemon && no_more {
101            let now = Instant::now();
102            if now < next_poll {
103                std::thread::sleep(Duration::from_millis(500));
104                continue;
105            }
106            // Poll interval elapsed — try again.
107            no_more = false;
108        }
109
110        if !no_more && workers.len() < max_concurrent {
111            let (blocked_epics, default_blocked) = {
112                let epic_ids: Vec<Option<String>> = workers.iter()
113                    .map(|(_, eid, _, _)| eid.clone())
114                    .collect();
115                let blocked = config.blocked_epics(&epic_ids);
116                let def_blocked = config.is_default_branch_blocked(&epic_ids);
117                (blocked, def_blocked)
118            };
119            match super::start::spawn_next_worker(root, true, skip_permissions, epic_filter.as_deref(), &blocked_epics, default_blocked) {
120                Ok(None) => {
121                    if daemon {
122                        let secs = interval_secs;
123                        log(&format!("No actionable tickets; next check in {secs}s"));
124                        next_poll = Instant::now() + Duration::from_secs(interval_secs);
125                    }
126                    no_more = true;
127                }
128                Ok(Some((id, epic_id, child, pid_path))) => {
129                    log(&format!(
130                        "Dispatched worker for ticket #{id}"
131                    ));
132                    started_ids.push(id.clone());
133                    workers.push((id, epic_id, child, pid_path));
134                    no_more = false;
135                }
136                Err(e) => {
137                    eprintln!("warning: dispatch failed: {e:#}");
138                    no_more = true;
139                    std::thread::sleep(Duration::from_secs(30));
140                }
141            }
142        } else {
143            std::thread::sleep(Duration::from_millis(500));
144        }
145    }
146
147    // Wait for all remaining workers in non-daemon mode (they were already
148    // reaped in the loop above for daemon mode; non-daemon exits when empty).
149    // In daemon mode workers run independently — we just stop dispatching.
150
151    if started_ids.is_empty() {
152        println!("No tickets to work.");
153        return Ok(());
154    }
155
156    if daemon {
157        // Don't print summary — workers are still running independently.
158        return Ok(());
159    }
160
161    let tickets = ticket::load_all_from_git(root, &config.tickets.dir)?;
162    let good_states: Vec<&str> = config.workflow.states.iter()
163        .filter(|s| s.terminal)
164        .map(|s| s.id.as_str())
165        .collect();
166    let mut any_bad = false;
167    println!("\nSummary:");
168    for id in &started_ids {
169        if let Some(t) = tickets.iter().find(|t| t.frontmatter.id == *id) {
170            let state = &t.frontmatter.state;
171            let ok = good_states.contains(&state.as_str());
172            if !ok { any_bad = true; }
173            println!("  #{id} {} — {state}", t.frontmatter.title);
174        }
175    }
176
177    if any_bad {
178        std::process::exit(1);
179    }
180    Ok(())
181}
182
183fn run_dry(root: &Path, config: &Config, epic_filter: Option<&str>) -> Result<()> {
184    let pw = config.workflow.prioritization.priority_weight;
185    let ew = config.workflow.prioritization.effort_weight;
186    let rw = config.workflow.prioritization.risk_weight;
187    let max_concurrent = config.agents.max_concurrent.max(1);
188
189    let startable: Vec<&str> = config.workflow.states.iter()
190        .filter(|s| s.transitions.iter().any(|tr| tr.trigger == "command:start"))
191        .map(|s| s.id.as_str())
192        .collect();
193    let actionable_owned = config.actionable_states_for("agent");
194    let actionable: Vec<&str> = actionable_owned.iter().map(|s| s.as_str()).collect();
195
196    let tickets = ticket::load_all_from_git(root, &config.tickets.dir)?;
197    let mut candidates: Vec<&ticket::Ticket> = tickets
198        .iter()
199        .filter(|t| {
200            let state = t.frontmatter.state.as_str();
201            actionable.contains(&state)
202                && (startable.is_empty() || startable.contains(&state))
203                && epic_filter
204                    .map_or(true, |id| t.frontmatter.epic.as_deref() == Some(id))
205        })
206        .collect();
207    candidates.sort_by(|a, b| {
208        b.score(pw, ew, rw)
209            .partial_cmp(&a.score(pw, ew, rw))
210            .unwrap_or(std::cmp::Ordering::Equal)
211    });
212
213    if candidates.is_empty() {
214        println!("dry-run: no actionable tickets");
215    } else {
216        for t in candidates.into_iter().take(max_concurrent) {
217            println!(
218                "dry-run: would start next: #{} [{}] {}",
219                t.frontmatter.id, t.frontmatter.state, t.frontmatter.title
220            );
221        }
222    }
223    Ok(())
224}
225
226#[cfg(test)]
227mod tests {
228    use super::*;
229
230    #[test]
231    fn daemon_dry_run_is_error() {
232        // We can't call run() against a real git repo here, but we can verify
233        // the guard fires before any I/O by passing a non-existent path and
234        // ensuring the error message mentions the flag combination.
235        let result = run(
236            std::path::Path::new("/nonexistent"),
237            false,
238            true,  // dry_run
239            true,  // daemon
240            30,
241            None,
242        );
243        let err = result.unwrap_err();
244        let msg = err.to_string();
245        assert!(
246            msg.contains("--daemon") && msg.contains("--dry-run"),
247            "unexpected error: {msg}"
248        );
249    }
250
251    #[test]
252    fn sig_count_increments_correctly() {
253        let sig_count = Arc::new(AtomicUsize::new(0));
254        assert_eq!(sig_count.load(Ordering::Relaxed), 0);
255
256        sig_count.fetch_add(1, Ordering::Relaxed);
257        assert_eq!(sig_count.load(Ordering::Relaxed), 1);
258
259        sig_count.fetch_add(1, Ordering::Relaxed);
260        let sigs = sig_count.load(Ordering::Relaxed);
261        assert!(sigs >= 2);
262    }
263}