devai - Command Agent File Runner
# install
# Will fix all code comment in all matching file
# How: It will run the installed Command Agent file ".devai/defaults/proof-comments.md" on all source files matching "./src/m*.rs"
# IMPORTANT: Make sure everything is committed before usage.
ONE Command Agent Markdown File that defines the full agent flow:
itemsget expanded from the-ffile matches (more ways to generate items later).-> Datascripting for getting full control over what data to put in the context.-> Instructiontemplating (Handlebars) to have full control over the prompt layout.-> Outputscripting to get full control over how to manage the AI output.
Data, Instruction, Output (and more later) are all defined in a single file (see below), which is called the Command Agent File
Supports all models/providers supported by the genai crate (see below for more information).
You can customize the model and concurrency in .devai/config.toml.
IMPORTANT: Make sure to run this command line when everything is committed, so that overwritten files can be reverted easily.
STILL IN HEAVY DEVELOPMENT... But it's starting to get pretty cool.
P.S. If possible, try to refrain from publishing devai-custom type crates, as this might be more confusing than helpful. However, any other name is great.
API Keys
devai uses the genai crate, and therefore the simplest way to provide the API keys for each provider is via environment variables in the terminal when running devai.
Here are the environment variable names used:
OPENAI_API_KEY
ANTHROPIC_API_KEY
MODEL_GEMINI
GEMINI_API_KEY
GROQ_API_KEY
COHERE_API_KEY
Usage & Concept
Usage: devai run proof-comments -f "./src/main.rs"
(or have any glob like -f "./src/**/*.rs" )
- This will initialize the
.devai/defaultsfolder with the "Command Agent Markdown"proof-comments.md(see .devai/defaults/proof-comments.md`) and run it with genai as follows:-f "./src/**/*.rs": The-fcommand line argument takes a glob and will create an "item" for each file, which can then be accessed in the# Datascripting section.# Data, which contains arhaiblock that will get executed with theitemvalue (the file reference in our example above).- With
rhai, there are some utility functions to list files, load file content, and such that can then be used in the instruction section.
- With
# Instruction, which is a Handlebars template section, has access toitemas well as the output of the# Datasection, accessible as thedatavariable.- This will be sent to the AI.
# Output, which now executes anotherrhaiblock, using theitem,data, andai_output, which is the string returned by the AI.- It can save files in place or create new files.
- Later, it will even be able to queue new devai work.
- By default, this will run with
gpt-4o-miniand look for theOPENAI_API_KEYenvironment variable. - It supports all AI providers supported by the genai crate.
- Here are the environment variable names per provider:
OPENAI_API_KEY,ANTHROPIC_API_KEY,COHERE_API_KEY,GEMINI_API_KEY,GROQ_API_KEY. - On Mac, if the environment variable is not present, it will attempt to prompt and get/save it from the keychain, under the devai group.
- Here are the environment variable names per provider:
Example of a Agent Command File
.devai/defaults/proof-comments.md (see .devai/defaults/proof-comments.md`)
Config
On devai run or devai init a .devai/config.toml will be created with the following:
[]
# Required (any model rust genai crate support).
= "gpt-4o-mini"
[]
# Default to 1 if absent. Great way to increase speed when remote AI services.
= 1
Future Plan
- Support for the
# Itemssection withyamlorRhai. - More
Rhaimodules/functions. - Support for
# Before All,# Before,# After, and# After All(allRhai). --dry-reqwill perform a dry run of the request by just saving the content of the request in a file.--dry-reswill perform a real AI request but just capture the AI response in a file (the request will be captured as well).--capturewill perform the normal run but capture the request and response in the request/response file.