Type Definition opencv::types::PtrOfBoost [−][src]
Implementations
Trait Implementations
Stores algorithm parameters in a file storage
simplified API for language bindings Stores algorithm parameters in a file storage Read more
Returns true if the Algorithm is empty (e.g. in the very beginning or after unsuccessful read
Saves the algorithm to a file. In order to make this method work, the derived class must implement Algorithm::write(FileStorage& fs). Read more
Returns the algorithm string identifier. This string is used as top level xml/yml node tag when the object is saved to a file or string. Read more
Type of the boosting algorithm. See Boost::Types. Default value is Boost::REAL. Read more
The number of weak classifiers. Default value is 100. Read more
A threshold between 0 and 1 used to save computational time. Samples with summary weight
inline formula
do not participate in the next iteration of training. Set this parameter to 0 to turn off this functionality. Default value is 0.95. Read moreType of the boosting algorithm. See Boost::Types. Default value is Boost::REAL. Read more
The number of weak classifiers. Default value is 100. Read more
A threshold between 0 and 1 used to save computational time. Samples with summary weight
inline formula
do not participate in the next iteration of training. Set this parameter to 0 to turn off this functionality. Default value is 0.95. Read moreCluster possible values of a categorical variable into K<=maxCategories clusters to find a suboptimal split. If a discrete variable, on which the training procedure tries to make a split, takes more than maxCategories values, the precise best subset estimation may take a very long time because the algorithm is exponential. Instead, many decision trees engines (including our implementation) try to find sub-optimal split in this case by clustering all the samples into maxCategories clusters that is some categories are merged together. The clustering is applied only in n > 2-class classification problems for categorical variables with N > max_categories possible values. In case of regression and 2-class classification the optimal split can be found efficiently without employing clustering, thus the parameter is not used in these cases. Default value is 10. Read more
The maximum possible depth of the tree. That is the training algorithms attempts to split a node while its depth is less than maxDepth. The root node has zero depth. The actual depth may be smaller if the other termination criteria are met (see the outline of the training procedure @ref ml_intro_trees “here”), and/or if the tree is pruned. Default value is INT_MAX. Read more
If the number of samples in a node is less than this parameter then the node will not be split. Read more
If CVFolds > 1 then algorithms prunes the built decision tree using K-fold cross-validation procedure where K is equal to CVFolds. Default value is 10. Read more
If true then surrogate splits will be built. These splits allow to work with missing data and compute variable importance correctly. Default value is false. Read more
If true then a pruning will be harsher. This will make a tree more compact and more resistant to the training data noise but a bit less accurate. Default value is true. Read more
If true then pruned branches are physically removed from the tree. Otherwise they are retained and it is possible to get results from the original unpruned (or pruned less aggressively) tree. Default value is true. Read more
Termination criteria for regression trees. If all absolute differences between an estimated value in a node and values of train samples in this node are less than this parameter then the node will not be split further. Default value is 0.01f Read more
Cluster possible values of a categorical variable into K<=maxCategories clusters to find a suboptimal split. If a discrete variable, on which the training procedure tries to make a split, takes more than maxCategories values, the precise best subset estimation may take a very long time because the algorithm is exponential. Instead, many decision trees engines (including our implementation) try to find sub-optimal split in this case by clustering all the samples into maxCategories clusters that is some categories are merged together. The clustering is applied only in n > 2-class classification problems for categorical variables with N > max_categories possible values. In case of regression and 2-class classification the optimal split can be found efficiently without employing clustering, thus the parameter is not used in these cases. Default value is 10. Read more
The maximum possible depth of the tree. That is the training algorithms attempts to split a node while its depth is less than maxDepth. The root node has zero depth. The actual depth may be smaller if the other termination criteria are met (see the outline of the training procedure @ref ml_intro_trees “here”), and/or if the tree is pruned. Default value is INT_MAX. Read more
If the number of samples in a node is less than this parameter then the node will not be split. Read more
If CVFolds > 1 then algorithms prunes the built decision tree using K-fold cross-validation procedure where K is equal to CVFolds. Default value is 10. Read more
If true then surrogate splits will be built. These splits allow to work with missing data and compute variable importance correctly. Default value is false. Read more
If true then a pruning will be harsher. This will make a tree more compact and more resistant to the training data noise but a bit less accurate. Default value is true. Read more
If true then pruned branches are physically removed from the tree. Otherwise they are retained and it is possible to get results from the original unpruned (or pruned less aggressively) tree. Default value is true. Read more
Termination criteria for regression trees. If all absolute differences between an estimated value in a node and values of train samples in this node are less than this parameter then the node will not be split further. Default value is 0.01f Read more
The array of a priori class probabilities, sorted by the class label value. Read more
Returns all the splits Read more
Trains the statistical model Read more
fn train(
&mut self,
samples: &dyn ToInputArray,
layout: i32,
responses: &dyn ToInputArray
) -> Result<bool>
fn train(
&mut self,
samples: &dyn ToInputArray,
layout: i32,
responses: &dyn ToInputArray
) -> Result<bool>
Trains the statistical model Read more
Returns the number of variables in training samples
Returns true if the model is trained
Returns true if the model is classifier
fn calc_error(
&self,
data: &Ptr<dyn TrainData>,
test: bool,
resp: &mut dyn ToOutputArray
) -> Result<f32>
fn calc_error(
&self,
data: &Ptr<dyn TrainData>,
test: bool,
resp: &mut dyn ToOutputArray
) -> Result<f32>
Computes error on the training or test dataset Read more
fn predict(
&self,
samples: &dyn ToInputArray,
results: &mut dyn ToOutputArray,
flags: i32
) -> Result<f32>
fn predict(
&self,
samples: &dyn ToInputArray,
results: &mut dyn ToOutputArray,
flags: i32
) -> Result<f32>
Predicts response(s) for the provided sample(s) Read more