gcp_bigquery_client/model/
ranking_metrics.rs

1//! Evaluation metrics used by weighted-ALS models specified by feedback_type=implicit.
2
3#[derive(Debug, Default, Clone, Serialize, Deserialize)]
4#[serde(rename_all = "camelCase")]
5pub struct RankingMetrics {
6    /// Calculates a precision per user for all the items by ranking them and then averages all the precisions across all the users.
7    pub mean_average_precision: Option<f64>,
8    /// A metric to determine the goodness of a ranking calculated from the predicted confidence by comparing it to an ideal rank measured by the original ratings.
9    pub normalized_discounted_cumulative_gain: Option<f64>,
10    /// Similar to the mean squared error computed in regression and explicit recommendation models except instead of computing the rating directly, the output from evaluate is computed against a preference which is 1 or 0 depending on if the rating exists or not.
11    pub mean_squared_error: Option<f64>,
12    /// Determines the goodness of a ranking by computing the percentile rank from the predicted confidence and dividing it by the original rank.
13    pub average_rank: Option<f64>,
14}