Expand description
Evaluation framework for the TraitClaw AI agent framework.
Provides EvalSuite, TestCase, and Metric abstractions for
measuring agent quality. Includes built-in metrics for relevancy
and keyword matching.
§Quick Start
use traitclaw_eval::{EvalSuite, TestCase, KeywordMetric, Metric};
let suite = EvalSuite::new("quality_tests")
.add_case(TestCase::new("greeting", "Say hello")
.expect_contains("hello"));
assert_eq!(suite.name(), "quality_tests");
assert_eq!(suite.cases().len(), 1);
let metric = KeywordMetric;
let score = metric.score("Say hello", "Hello! How can I help?", &["hello"]);
assert!(score > 0.0);Re-exports§
pub use export::EvalReportExport;pub use metrics::JudgeProvider;pub use metrics::LlmJudgeMetric;pub use metrics::SchemaValidationMetric;pub use metrics::ToolUsageMetric;pub use runner::AsyncMetric;pub use runner::EvalAgent;pub use runner::EvalRunner;pub use runner::SyncMetricAdapter;
Modules§
- export
- Report export — JSON and CSV serialization for
EvalReport. - metrics
- Specialized metrics for LLM output evaluation.
- runner
- Async
Metrictrait andEvalRunnerexecution engine.
Structs§
- Eval
Report - An evaluation report.
- Eval
Suite - A suite of evaluation test cases.
- Keyword
Metric - Built-in keyword matching metric.
- Length
Relevancy Metric - Built-in length-based relevancy metric.
- Test
Case - A single evaluation test case.
- Test
Result - A single test case result.
Traits§
- Metric
- Trait for evaluation metrics.