Expand description
Evaluation metrics for machine learning models
This crate provides a comprehensive set of metrics for evaluating machine learning models, including classification, regression, and clustering metrics.
§Examples
use sklears_metrics::classification::{accuracy_score, precision_score, recall_score, f1_score};
use scirs2_core::ndarray::array;
let y_true = array![0, 1, 2, 0, 1, 2];
let y_pred = array![0, 2, 1, 0, 0, 1];
let accuracy = accuracy_score(&y_true, &y_pred).unwrap();
println!("Accuracy: {:.3}", accuracy);§Classification Metrics
accuracy_score: Classification accuracyprecision_score,recall_score,f1_score: Precision, recall, and F1 scorefbeta_score: F-beta score with configurable betabalanced_accuracy_score: Accuracy adjusted for imbalanced datasetscohen_kappa_score: Cohen’s kappa coefficientmatthews_corrcoef: Matthews correlation coefficientconfusion_matrix: Confusion matrixlog_loss: Logarithmic losshamming_loss: Hamming lossjaccard_score: Jaccard similarity coefficientzero_one_loss: Zero-one classification losshinge_loss: Hinge loss for SVMbrier_score_loss: Brier score for probabilistic predictionstop_k_accuracy_score: Top-k accuracy for multi-class classificationtop_2_accuracy_score,top_3_accuracy_score,top_5_accuracy_score: Convenience functionsmultilabel_exact_match_ratio: Exact match ratio for multi-label classificationmultilabel_accuracy_score: Jaccard similarity for multi-label classificationmultilabel_hamming_loss: Hamming loss for multi-label classificationmultilabel_ranking_loss: Ranking loss for multi-label classificationmultilabel_average_precision_score: Average precision for multi-label classificationdemographic_parity_difference: Fairness metric measuring difference in positive prediction ratesdemographic_parity_ratio: Fairness metric measuring ratio of positive prediction ratesequalized_odds_difference: Fairness metric measuring difference in true positive ratesequal_opportunity_difference: Alias for equalized_odds_differenceexpected_calibration_error: Expected Calibration Error (ECE) for model calibration assessmentreliability_diagram: Data for creating reliability diagrams (calibration plots)cost_sensitive_accuracy: Cost-sensitive accuracy for imbalanced classificationcost_sensitive_loss: Cost-sensitive loss functionexpected_cost: Expected cost given prediction probabilitiescost_weighted_f1_score: Cost-weighted F1 scorespherical_score: Spherical scoring rule for probabilistic classificationquadratic_score: Quadratic scoring rule for probabilistic classificationcross_entropy: Cross-entropy loss for multi-class classificationkl_divergence: Kullback-Leibler divergence between probability distributionsjs_divergence: Jensen-Shannon divergence (symmetric version of KL divergence)hierarchical_precision: Hierarchical precision with partial credit for ancestor matcheshierarchical_recall: Hierarchical recall with partial credit for ancestor matcheshierarchical_f1_score: Hierarchical F1-score for tree-structured labelstree_distance_loss: Tree distance-based loss for hierarchical classificationhierarchical_accuracy: Hierarchical accuracy with optional partial credit
§Regression Metrics
mean_absolute_error: Mean absolute error (MAE)mean_squared_error: Mean squared error (MSE)root_mean_squared_error: Root mean squared error (RMSE)r2_score: Coefficient of determination (R²)mean_absolute_percentage_error: Mean absolute percentage error (MAPE)explained_variance_score: Explained variance scoremax_error: Maximum errormedian_absolute_error: Median absolute errormean_squared_log_error: Mean squared logarithmic errorroot_mean_squared_log_error: Root mean squared logarithmic errormean_gamma_deviance: Mean Gamma deviancemean_poisson_deviance: Mean Poisson deviancemean_tweedie_deviance: Mean Tweedie deviance with power parametermean_pinball_loss: Pinball loss for quantile regressiond2_absolute_error_score: D² score based on absolute errord2_pinball_score: D² score based on pinball lossd2_tweedie_score: D² score based on Tweedie deviancehuber_loss: Huber loss for robust regressionquantile_loss: Quantile loss (alias for pinball loss)median_absolute_percentage_error: Median absolute percentage error (MdAPE)theil_u_statistic: Theil’s U statistic for forecast accuracysymmetric_mean_absolute_percentage_error: Symmetric MAPE (sMAPE)mean_absolute_scaled_error: Mean Absolute Scaled Error (MASE) for time seriesrobust_r2_score: Robust R-squared variants (median, trimmed, Huber, MAD-based)median_r2_score: Median-based robust R-squaredtrimmed_r2_score: Trimmed R-squared (excluding outliers)huber_r2_score: Huber loss-based robust R-squaredmad_r2_score: Median Absolute Deviation-based R-squaredcontinuous_ranked_probability_score: CRPS for distributional forecastscrps_ensemble: CRPS for ensemble forecasts (convenience function)crps_gaussian: CRPS for Gaussian predictions with analytical formulaenergy_score: Energy score for multivariate distributional forecastsdirection_accuracy: Directional accuracy percentage for time series forecastingdirectional_symmetry: Balance between upward and downward predictionshit_rate: Hit rate for directional predictions above a thresholddirectional_theil_u: Directional adaptation of Theil’s U statistictrend_accuracy: Trend accuracy using sliding window comparisonprediction_interval_coverage: Proportion of true values within prediction intervalsmean_interval_width: Average width of prediction intervalsinterval_score: Proper scoring rule combining coverage and widthprobability_integral_transform: PIT for distributional forecast evaluationpit_uniformity_test: Test PIT values for uniform distributiondiebold_mariano_test: Statistical test for comparing forecast accuracyseasonal_mase: Seasonal Mean Absolute Scaled Error using seasonal naive forecastseasonal_naive_forecast_error: Relative error compared to seasonal naive forecastseasonal_autocorrelation: Autocorrelation coefficient at seasonal lagseasonal_strength: Strength of seasonal component (0 to 1)persistence_model_comparison: Compare against simple, seasonal, and trend persistence models
§Ranking Metrics
auc: Area Under the Curveaverage_precision_score: Average precision from precision-recall curvecoverage_error: Coverage error for multi-label rankingdcg_score: Discounted Cumulative Gainndcg_score: Normalized Discounted Cumulative Gainroc_auc_score: Area Under the ROC Curve (binary)roc_auc_score_multiclass: Multi-class ROC AUC with OvR/OvO strategiesprecision_recall_auc_score: Area Under the Precision-Recall Curveroc_curve: Compute ROC curve coordinatesprecision_recall_curve: Compute precision-recall curve coordinatesmean_average_precision: Mean Average Precision (MAP) for information retrievalmean_reciprocal_rank: Mean Reciprocal Rank (MRR) for ranking evaluation
§Clustering Metrics
adjusted_rand_score: Adjusted Rand indexadjusted_mutual_info_score: Adjusted mutual informationcalinski_harabasz_score: Calinski-Harabasz index (variance ratio criterion)completeness_score: Completeness metricdavies_bouldin_score: Davies-Bouldin indexfowlkes_mallows_score: Fowlkes-Mallows indexhomogeneity_score: Homogeneity metrichomogeneity_completeness_v_measure: All three metrics in one callmutual_info_score: Mutual informationnormalized_mutual_info_score: Normalized mutual informationrand_score: Rand indexsilhouette_score: Silhouette coefficientv_measure_score: V-measure (harmonic mean of homogeneity and completeness)dunn_index: Dunn index for cluster separation assessmentgap_statistic: Gap statistic for optimal number of clusterswithin_cluster_sum_of_squares: Within-cluster sum of squares for cluster compactnessbetween_cluster_sum_of_squares: Between-cluster sum of squares for cluster separationbootstrap_stability: Bootstrap stability for clustering robustness assessmentjaccard_stability: Jaccard stability coefficient between clusteringsconsensus_clustering_stability: Consensus clustering stability metricperturbation_stability: Clustering stability under data perturbationsparameter_stability: Clustering stability across parameter variationsentropy: Shannon entropy for discrete distributionsconditional_entropy: Conditional entropy H(Y|X)mutual_information: Mutual information I(X; Y) between variablesnormalized_mutual_information_symmetric: Normalized mutual information (symmetric)joint_entropy: Joint entropy H(X, Y) of two variablesvariation_of_information: Variation of information distance between clusteringsinformation_gain: Information gain (entropy reduction)information_gain_ratio: Normalized information gainintra_cluster_coherence: Average pairwise similarity within clustersinter_cluster_separation: Average dissimilarity between clusterscluster_coherence_score: Combined coherence score with configurable weightingsemantic_coherence: Semantic coherence for text clustering using word co-occurrencexie_beni_index: Validity index for fuzzy clustering (compactness to separation ratio)ball_hall_index: Internal validity measure (average within-cluster distance)hartigan_index: Stopping rule for k-means clusteringkrzanowski_lai_index: Stopping rule using rate of change in WCSSbic_clustering: Bayesian Information Criterion for clusteringaic_clustering: Akaike Information Criterion for clusteringsugar_james_index: Model selection criterion based on distortion and degrees of freedom
§Computer Vision Metrics
psnr: Peak Signal-to-Noise Ratio for image quality assessmentssim: Structural Similarity Index for perceptual image qualityiou_boxes: Intersection over Union for bounding boxesiou_masks: Intersection over Union for segmentation masksmean_average_precision: Mean Average Precision (mAP) for object detectionmean_iou: Mean Intersection over Union for semantic segmentationpixel_accuracy: Pixel-wise accuracy for segmentation tasksDetection: Structure for object detection resultsGroundTruth: Structure for ground truth annotations
§Natural Language Processing Metrics
bleu_score: BLEU score for machine translation evaluationrouge_n_score: ROUGE-N score for summarization evaluationrouge_l_score: ROUGE-L score using longest common subsequenceperplexity: Perplexity for language model evaluationjaccard_similarity: Jaccard similarity coefficient for textcosine_similarity_tfidf: Cosine similarity using TF-IDF vectorsedit_distance: Levenshtein distance between stringsnormalized_edit_distance: Normalized edit distance between stringsSmoothingFunction: Enumeration of smoothing methods for BLEU
§Survival Analysis Metrics
concordance_index: C-index for survival analysistime_dependent_auc: Time-dependent AUC for survival predictionsbrier_score_survival: Brier score for survival analysisintegrated_brier_score: Integrated Brier score over timekaplan_meier_survival: Kaplan-Meier survival function estimationlog_rank_test: Log-rank test for comparing survival curves
§Pairwise Metrics
euclidean_distances: Compute euclidean distances between samplesnan_euclidean_distances: Euclidean distances ignoring NaN valuespairwise_distances: Compute distances with various metrics (Euclidean, Manhattan, Chebyshev, Minkowski, Cosine, Hamming)pairwise_distances_argmin: Find minimum distances and indicespairwise_distances_argmin_min: Find both argmin and min valuespairwise_kernels: Compute kernel matrix (Linear, Polynomial, RBF, Sigmoid, Cosine)wasserstein_distance: Earth Mover’s Distance between 1D distributionsmahalanobis_distances: Mahalanobis distance accounting for correlationscosine_similarity: Cosine similarity matrix between samplesnormalized_compression_distance: Universal metric based on compression algorithmsnormalized_compression_distance_matrix: NCD between all pairs of sequencesapproximate_kolmogorov_complexity: Estimate Kolmogorov complexity using compressioninformation_distance: Information distance between byte sequencesstring_kernel_similarity: Subsequence-based similarity for stringsstring_kernel_matrix: Pairwise string kernel similarities
§Statistical Tests
mcnemar_test: Compare two binary classifiers using McNemar’s testfriedman_test: Compare multiple algorithms across datasetswilcoxon_signed_rank_test: Non-parametric test for paired samplespermutation_test: General permutation test frameworktransfer_entropy: Directed information transfer between time seriesbidirectional_transfer_entropy: Transfer entropy in both directionsnet_transfer_entropy: Net information flow between time seriesmulti_lag_transfer_entropy: Transfer entropy across multiple lagspartial_transfer_entropy: Transfer entropy conditioning on third variable
§Temporal and Dynamic Metrics
TemporalMetricsAnalyzer: Main analyzer for temporal metric patternsdetect_concept_drift: Detect concept drift using multiple methods (KS test, Page-Hinkley, ADWIN, SPC, PSI)analyze_temporal_trend: Analyze trends with seasonal decomposition and change point detectioncalculate_adaptive_weights: Calculate time-decaying weights for temporal datatemporal_stability: Measure temporal stability of metrics over timetrack_metric_evolution: Track how metrics evolve with rolling statisticsConceptDriftResult: Results of drift detection with confidence and magnitudeTemporalTrendAnalysis: Comprehensive trend analysis with seasonality detectionMetricEvolution: Evolution tracking with change point detectionWindowConfig: Configuration for sliding windows and drift detection parameters
§Interpretability Metrics
calculate_faithfulness_removal: Faithfulness using feature removal/occlusioncalculate_faithfulness_permutation: Faithfulness using feature permutationcalculate_explanation_stability: Stability analysis using correlation, cosine similarity, rank correlationcalculate_comprehensibility: Comprehensibility assessment with sparsity and complexity measurescalculate_trustworthiness: Comprehensive trustworthiness combining multiple metricsevaluate_ranking_quality: Quality assessment for feature importance rankingsFaithfulnessResult: Detailed faithfulness evaluation with confidence intervalsStabilityResult: Pairwise stability analysis across explanationsComprehensibilityResult: Comprehensibility assessment with entropy and consistencyTrustworthinessResult: Combined trustworthiness score with individual componentsRankingQualityResult: Feature importance ranking validation and consistency
§Multi-Objective Evaluation
pareto_frontier: Find Pareto optimal solutions (non-dominated models)topsis_ranking: TOPSIS multi-criteria decision analysisweighted_sum_ranking: Weighted sum approach for model rankingtrade_off_analysis: Analyze trade-offs between competing metricsutility_optimization: Optimize custom utility functionsmulti_objective_evaluation: Comprehensive multi-objective evaluationMultiObjectiveResult: Structure containing complete evaluation resultsMultiObjectiveConfig: Configuration for multi-objective evaluation
§Uncertainty Quantification
bootstrap_confidence_interval: Bootstrap confidence intervals for any metricbca_bootstrap_confidence_interval: Bias-corrected and accelerated bootstrapbayesian_accuracy_credible_interval: Bayesian credible intervals for accuracycorrelation_confidence_interval: Analytical confidence intervals for correlationmse_confidence_interval: Confidence intervals for mean squared erroruncertainty_propagation: Uncertainty propagation for composite metricsbootstrap_metric_comparison: Bootstrap hypothesis testing for metric comparisonUncertaintyResult: Structure containing uncertainty quantification resultsUncertaintyConfig: Configuration for uncertainty quantification
§Type Safety and Compile-Time Validation
TypedMetric: Type-safe metric wrapper with phantom typesMetricCategory: Trait for defining metric categoriesClassification,Regression,Clustering: Phantom types for metric categoriesMetric: Trait for computable metrics with type safetyMetricSuite: Type-safe collection of metrics from the same categoryCompositeMetric: Type-safe composition of metrics from different categoriesMetricTransform: Trait for type-safe metric transformationsMetricBuilder: Builder pattern for type-safe metric constructionValidatedMetric: Compile-time validation using const genericsZeroCostMetric: Zero-cost abstraction for metric computation
§Performance Enhancements
HighPerformanceMetricsComputer: All-in-one high-performance metrics computationAdaptiveMetricsComputer: Adaptive algorithm selection based on data characteristicsCacheFriendlyAccumulator: Cache-aligned data structures for metric accumulationLockFreeMetricsAccumulator: Lock-free concurrent metrics accumulationPrefetchingMetricsComputer: Memory prefetching for improved cache performanceCacheOptimizedMatrixOps: Cache-conscious matrix operationsProfileGuidedOptimizer: Profile-guided optimization and performance analysisMemoryPrefetcher: Memory prefetching utilities for cache optimization
§Validation Framework
MetricValidator: Comprehensive validation framework for metric correctnessSyntheticDataGenerator: Generate synthetic data for testing metric implementationsReferenceTestCase: Reference test cases with known expected resultsValidationResult: Results of metric validation with error analysisComprehensiveValidationReport: Complete validation report with multiple test typesStabilityAnalysis: Bootstrap-based stability analysis for metric robustnessMetamericAnalysis: Parameter sensitivity analysis for understanding metric behaviorStandardReferenceDatasets: Standard test cases for common metrics
§Automated Benchmarking
BenchmarkSuite: Comprehensive benchmarking suite for metrics validationadd_classification_benchmark: Add classification metric benchmarks with datasetsadd_regression_benchmark: Add regression metric benchmarks with datasetsadd_clustering_benchmark: Add clustering metric benchmarks with datasetsrun_all: Execute all benchmarks with performance and accuracy testingrun_scalability_test: Test metric performance across different data sizesBenchmarkResult: Detailed benchmark results with timing and accuracyBenchmarkReport: Comprehensive benchmarking report with statisticsDataset: Standard datasets (Iris, Wine, BreastCancer, BostonHousing, etc.)ReferenceImplementations: Reference metric implementations for validation
§Interactive Visualization
create_roc_curve_data: Generate ROC curve data with AUC calculationcreate_precision_recall_data: Generate PR curve data with average precisioncreate_calibration_plot_data: Generate calibration plots for probability assessmentRocCurveData: ROC curve visualization with HTML plot generationPrecisionRecallData: PR curve visualization with interactive featuresConfusionMatrixVisualization: Confusion matrix heatmaps with normalizationCalibrationPlot: Calibration plots with Brier score and ECE metricsLearningCurve: Learning curve visualization for training progressFeatureImportanceViz: Feature importance bar charts with rankingMetricDashboard: Comprehensive metric comparison dashboardPlotConfig: Configuration for plot styling and interactivity
§Federated Learning Metrics
privacy_preserving_aggregation: Differentially private metric aggregationcommunication_efficient_aggregation: Weighted aggregation with compression effectsdemographic_parity_across_clients: Fairness evaluation using coefficient of variationequalized_odds_across_clients: Equalized odds difference across federated clientscommunication_efficiency: Measure efficiency as improvement per unit communication costclient_contribution_score: Assess individual client contributions using Shapley-like valuesshapley_client_contributions: Calculate exact Shapley values for client coalitionsprivacy_budget_allocation: Track differential privacy budget across federated roundsanalyze_convergence: Convergence analysis for federated training with stability measurescomprehensive_federated_evaluation: Complete federated learning evaluation frameworksecure_aggregation: Secure multiparty computation for metric aggregationFederatedConfig: Configuration for federated learning evaluation parametersFederatedEvaluationResult: Comprehensive results including global metrics, fairness, efficiencyConvergenceMetrics: Convergence rate, stability, and rounds to convergence analysisPrivacyComposition: Privacy composition methods (Basic, Advanced, RDP)
§Adversarial Robustness Metrics
adversarial_accuracy: Accuracy on adversarial examples compared to true labelsattack_success_rate: Fraction of examples where predictions changed due to perturbationsrobust_accuracy: Accuracy within a specific perturbation budget constraintcertified_accuracy: Accuracy with formal certification guarantees against perturbationsaverage_perturbation_magnitude: Average L-p norm of adversarial perturbationsrobustness_score: Weighted combination of clean and adversarial accuracyadversarial_transferability: Success rate of adversarial examples across different modelsgradient_based_robustness: Local intrinsic dimensionality using gradient informationadaptive_attack_resistance: Resistance to adaptive attacks accounting for gradient maskingempirical_robustness: Robustness estimation using random noise perturbationsarea_under_robustness_curve: AURC metric for robustness across perturbation budgetscomprehensive_adversarial_evaluation: Complete adversarial evaluation frameworkAdversarialConfig: Configuration for perturbation budgets, norms, and attack parametersAdversarialResult: Comprehensive results including multiple attack types and metricsAttackResult: Individual attack results with success rates and perturbation statisticsNormType: Perturbation norm types (L-infinity, L2, L1, L0)AttackType: Supported attack methods (FGSM, PGD, C&W, AutoAttack, etc.)
§Automated Reporting
generate_metric_report: Create comprehensive metric reports with statistical analysisgenerate_model_comparison_report: Automated comparison reports with significance testingMetricReport: Complete report structure with metadata, summaries, and recommendationsMetricSummary: Individual metric summaries with interpretations and confidence intervalsExecutiveSummary: High-level summary for stakeholders with key findings and business impactModelComparison: Pairwise model comparison with statistical and practical significanceStatisticalAnalysis: Sample size analysis, confidence intervals, and power analysisRecommendation: Automated recommendations with priority levels and implementation guidancePerformanceTrends: Trend analysis and performance regression detectionReportConfig: Configuration for report generation including formats and significance thresholdsReportFormat: Output formats (HTML, Markdown, Text, JSON)PerformanceGrade: Performance grading system (Excellent, Good, Fair, Poor, Critical)SignificanceTest: Statistical significance test results with p-values and effect sizesPracticalSignificance: Practical significance assessment beyond statistical significance
§Fluent API and Builder Patterns
MetricsBuilder: Fluent API for metric computation with method chainingMetricPreset: Configuration presets for common use cases (ClassificationBasic, RegressionBasic, etc.)MetricConfig: Configuration for metric computation with confidence intervals and averagingMetricResults: Serializable metric results with metadata and confidence intervalsConfigBuilder: Builder pattern for advanced metric configurationquick_classification_metrics: Convenience function for rapid classification evaluationquick_regression_metrics: Convenience function for rapid regression evaluationAveragingStrategy: Strategies for multi-class averaging (Macro, Micro, Weighted, Binary)ZeroDivisionStrategy: Handling strategies for division by zero in metrics
§Async Streaming Metrics (feature = “async”)
StreamingMetricsComputer: Async streaming metric computation with backpressure handlingMetricStream: Stream wrapper for real-time metric computation with reactive updatesChannelMetricsComputer: Channel-based async metric computation for producer-consumer patternsStreamingConfig: Configuration for async streaming operations (chunk size, concurrency, windowing)MetricAccumulator: Incremental metric accumulation for streaming data with sliding windowsstreaming_accuracy: Convenience function for streaming accuracy computationstreaming_classification_metrics: Convenience function for comprehensive streaming classification evaluation
§Modular Framework
Metric: Core trait for all metrics with type-safe input/outputComposableMetric: Trait for metrics that can be combined and transformedMetricAggregator: Trait for aggregating multiple metric resultsMetricPipeline: Pipeline for composing multiple metrics and aggregatorsMetricMiddleware: Middleware system for metric processing pipelinesMetricRegistry: Dynamic registration and discovery of metricsScoringFunction: Extensible scoring function systemMetricPlugin: Plugin architecture for extending framework capabilitiesPluginManager: Manager for loading and organizing metric plugins
Re-exports§
pub use scoring::check_multimetric_scoring;pub use scoring::check_scoring;pub use scoring::get_scorer;pub use scoring::get_scorer_names;pub use scoring::make_scorer;pub use scoring::Scorer;pub use scoring::ScorerConfig;pub use scoring::ScoringMetric;pub use classification::ClassificationReport;pub use classification::ConfusionMatrixDisplay;pub use classification::DetCurveDisplay;pub use classification::MetricsDisplay;pub use classification::PrecisionRecallDisplay;pub use classification::RocCurveDisplay;
Modules§
- advanced_
metrics - Advanced classification metrics
- adversarial_
robustness - Adversarial Robustness Metrics
- automated_
reporting - Automated Reporting for Machine Learning Metrics
- basic_
metrics - Basic classification metrics
- benchmarking
- Automated Benchmarking System for Machine Learning Metrics
- classification
- Classification metrics
- clustering
- Clustering evaluation metrics and validation utilities
- computer_
vision - Computer Vision Metrics
- display_
utils - Display utilities for classification metrics
- fairness_
metrics - Fairness metrics for algorithmic bias evaluation
- federated_
learning - Federated Learning Metrics
- fluent_
api - Fluent API for Machine Learning Metrics
- interpretability
- Interpretability Metrics for Machine Learning Model Explanations
- mathematical_
foundations - Mathematical Foundations and Derivations
- memory_
efficiency - Memory efficiency improvements for metrics computation
- modular_
framework - Modular Metric Framework with Trait-Based Design
- multi_
objective - Multi-Objective Evaluation Framework
- multilabel_
metrics - Multi-label classification metrics
- nlp
- Natural Language Processing Metrics
- optimized
- Optimized Metric Implementations for High-Performance Computing
- pairwise
- Pairwise distance and similarity metrics
- performance_
enhancements - Advanced Performance Enhancements for Metrics Computation
- probabilistic_
metrics - Probabilistic classification metrics
- ranking
- regression
- Regression metrics for model evaluation
- scoring
- Scoring utilities for model evaluation
- statistical_
tests - Statistical tests for model comparison and evaluation
- survival
- Survival Analysis Metrics
- temporal
- Temporal and Dynamic Metrics for Machine Learning
- thread_
local_ optimization - Thread-Local Storage Optimization for Metric Updates
- type_
safety - Type Safety and Compile-Time Validation for Metrics
- uncertainty
- Uncertainty Quantification for Machine Learning Metrics
- validation
- Validation Framework for Machine Learning Metrics
- visualization
- Interactive Visualization Utilities for Machine Learning Metrics
Enums§
- Metrics
Error - Common error type for metrics
Type Aliases§
- Metrics
Result - Type alias for metrics results