agent-procs 0.3.1

Concurrent process runner for AI agents
Documentation

agent-procs

Concurrent process runner for AI agents. Processes run in a background daemon and persist across CLI invocations.

Install

cargo install agent-procs

Quick start

# Start a process
agent-procs run "npm run dev" --name server

# Wait for it to be ready
agent-procs wait server --until "Listening on" --timeout 30

# Check output
agent-procs logs server --tail 50

# See what's running
agent-procs status

# Stop it
agent-procs stop server

Config file

Create an agent-procs.yaml to manage multiple processes together:

session: myproject                          # optional — isolates this project's processes
proxy: true                                 # optional — enables reverse proxy
proxy_port: 9095                            # optional — pin proxy to a specific port

processes:
  db:
    cmd: docker compose up postgres
    ready: "ready to accept connections"
  api:
    cmd: ./start-api-server
    cwd: ./backend
    env:
      DATABASE_URL: postgres://localhost:5432/mydb
    ready: "Listening on :8080"
    port: 8080
    depends_on: [db]

Processes start in dependency order; independent ones run concurrently.

agent-procs up                    # start all
agent-procs up --only db,api      # start specific ones
agent-procs down                  # stop all

Field reference

Per-process fields:

Field Required Description
cmd yes Shell command to execute
cwd no Working directory (relative to config file location)
env no Environment variables (key: value map)
ready no Stdout pattern that signals the process is ready
depends_on no List of process names that must be ready first
port no Port number — injected as PORT env var

Top-level fields:

Field Required Description
session no Session name (overridden by --session CLI flag)
proxy no Enable reverse proxy (default: false)
proxy_port no Pin proxy to a specific port (default: auto-assign from 9090-9190)

Reverse proxy

Give processes stable named URLs instead of port numbers. Opt-in via proxy: true in config or --proxy on the CLI.

$ agent-procs up
Proxy listening on http://localhost:9090
started api (http://api.localhost:9090)
started web (http://web.localhost:9090)
  • Processes without an explicit port get one auto-assigned (4000-4999 range)
  • PORT and HOST=127.0.0.1 are injected into the process env (user env takes precedence)
  • Each session gets its own proxy port, so two projects can both have api without conflict

Ad-hoc usage without a config file:

agent-procs run "node server.js" --name api --port 3001 --proxy
# → http://api.localhost:9090

Commands

Command Description
run <cmd> [--name N] [--port P] [--proxy] Spawn a background process
stop <name> Stop a process
stop-all Stop all processes
restart <name> Restart a process
status [--json] Show all process statuses
logs <name> [--tail N] [--follow] [--stderr] [--all] View process output
wait <name> --until <pattern> [--regex] [--timeout N] Wait for output pattern
wait <name> --exit [--timeout N] Wait for process to exit
up [--only X,Y] [--config path] [--proxy] Start from config file
down Stop config-managed processes
session list List active sessions
session clean Remove stale sessions
ui Open terminal UI
completions <shell> Generate shell completions (bash, zsh, fish, powershell)

Sessions

Use --session to isolate process groups (e.g. per-project):

agent-procs --session projectA run "make serve" --name app
agent-procs --session projectB run "make serve" --name app
agent-procs --session projectA status   # only shows projectA's processes

Architecture

The CLI communicates with a per-session background daemon over a Unix domain socket. The daemon manages process lifecycles, captures stdout/stderr to log files, and handles wait conditions. The daemon auto-starts on first use and exits when all processes are stopped.

Exit codes

Code Meaning
0 Success
1 Error (timeout, connection failure, unexpected response)
2 No logs found for target process