Expand description
Evaluation pipeline for content contributors and live peers.
Per CP-015 section 5: Validators download content from Arweave, index it locally in an isolated substrate, run test queries, and measure retrieval quality using precision@10, NDCG@10, and MRR.
Per CP-015 section 13: Validators test live peers by connecting over Tor, sending test queries, and measuring search quality and latency.
Structs§
- Evaluation
Result - Result of evaluating a single contributor’s content.
- Peer
Quality Metrics - Quality metrics for a live peer, including latency and availability.
- Peer
Test Result - Result of testing a live peer.
- Quality
Metrics - Quality metrics for a contributor’s content, measured against ground truth.
Functions§
- composite_
score - Composite quality score for a contributor evaluation result.
- evaluate_
contributor - Evaluate a contributor by downloading their content from Arweave, indexing it in an isolated substrate, and running test queries.
- evaluate_
local_ graph - Evaluate a local graph store directly by running test queries against it.
- evaluate_
peer - Evaluate a live peer by connecting over Tor, sending test queries, and measuring search quality and latency.
- mrr
- Mean Reciprocal Rank: the reciprocal of the rank of the first relevant result.
- ndcg_
at_ k - Normalized Discounted Cumulative Gain at K.
- peer_
composite_ score - Composite quality score for a peer test result.
- precision_
at_ k - Precision at K: fraction of the top-K results that are relevant.
- update_
contributor_ ratings - Update contributor ratings based on evaluation results using pairwise
OpenSkillcomparisons. - update_
peer_ ratings - Update peer ratings based on peer test results using pairwise
OpenSkillcomparisons.