module documentation
Classification evaluation.
Class |
|
Class that stores the results of a binary classification evaluation. |
Class |
|
Binary classification evaluation. |
Class |
|
Binary evaluation config. |
Class |
|
Base class for classification evaluation methods. |
Class |
|
Base class for configuring ClassificationEvaluation instances. |
Class |
|
Class that stores the results of a classification evaluation. |
Class |
|
Standard classification evaluation. |
Class |
|
Simple classification evaluation config. |
Class |
|
Top-k classification evaluation. |
Class |
|
Top-k classification evaluation config. |
Function | evaluate |
Evaluates the classification predictions in the given collection with respect to the specified ground truth labels. |
Function | _evaluate |
Undocumented |
Function | _parse |
Undocumented |
Function | _to |
Undocumented |
def evaluate_classifications(samples, pred_field, gt_field='ground_truth', eval_key=None, classes=None, missing=None, method=None, custom_metrics=None, progress=None, **kwargs):
(source)
¶
Evaluates the classification predictions in the given collection with respect to the specified ground truth labels.
By default, this method simply compares the ground truth and prediction for each sample, but other strategies such as binary evaluation and top-k matching can be configured via the method parameter.
You can customize the evaluation method by passing additional parameters for the method's config class as kwargs.
The natively provided method values and their associated configs are:
- "simple":
SimpleEvaluationConfig
- "top-k":
TopKEvaluationConfig
- "binary":
BinaryEvaluationConfig
If an eval_key is specified, then this method will record some statistics on each sample:
- When evaluating sample-level fields, an eval_key field will be populated on each sample recording whether that sample's prediction is correct.
- When evaluating frame-level fields, an eval_key field will be populated on each frame recording whether that frame's prediction is correct. In addition, an eval_key field will be populated on each sample that records the average accuracy of the frame predictions of the sample.
Parameters | |
samples | a fiftyone.core.collections.SampleCollection |
pred | the name of the field containing the predicted
fiftyone.core.labels.Classification instances |
gt | the name of the field containing the ground
truth fiftyone.core.labels.Classification instances |
evalNone | an evaluation key to use to refer to this evaluation |
classes:None | the list of possible classes. If not provided, the observed ground truth/predicted labels are used |
missing:None | a missing label string. Any None-valued labels are given this label for results purposes |
method:None | a string specifying the evaluation method to use. The supported values are fo.evaluation_config.classification_backends.keys() and the default is fo.evaluation_config.default_classification_backend |
customNone | an optional list of custom metrics to compute or dict mapping metric names to kwargs dicts |
progress:None | whether to render a progress bar (True/False), use the default value fiftyone.config.show_progress_bars (None), or a progress callback function to invoke instead |
**kwargs | optional keyword arguments for the constructor of the
ClassificationEvaluationConfig being used |
Returns | |
a ClassificationResults |