- Resource: Evaluation
- EvaluationMetrics
- ClassificationMetrics
- PrCurve
- ConfidenceMetricsEntry
- ConfusionMatrix
- Row
- ConfusionMatrixEntry
- ObjectDetectionMetrics
- Methods
Resource: Evaluation
Describes an evaluation between a machine learning model's predictions and ground truth labels. Created when an EvaluationJob
runs successfully.
JSON representation | |
---|---|
{ "name": string, "config": { object ( |
Fields | |
---|---|
name |
Output only. Resource name of an evaluation. The name has the following format: "projects/{project_id}/datasets/{dataset_id}/evaluations/{evaluation_id}' |
config |
Output only. Options used in the evaluation job that created this evaluation. |
evaluationJobRunTime |
Output only. Timestamp for when the evaluation job that created this evaluation ran. A timestamp in RFC3339 UTC "Zulu" format, accurate to nanoseconds. Example: |
createTime |
Output only. Timestamp for when this evaluation was created. A timestamp in RFC3339 UTC "Zulu" format, accurate to nanoseconds. Example: |
evaluationMetrics |
Output only. Metrics comparing predictions to ground truth labels. |
annotationType |
Output only. Type of task that the model version being evaluated performs, as defined in the
|
evaluatedItemCount |
Output only. The number of items in the ground truth dataset that were used for this evaluation. Only populated when the evaulation is for certain AnnotationTypes. |
EvaluationMetrics
JSON representation | |
---|---|
{ // Union field |
Fields | ||
---|---|---|
Union field metrics . Common metrics covering most general cases. metrics can be only one of the following: |
||
classificationMetrics |
|
|
objectDetectionMetrics |
|
ClassificationMetrics
Metrics calculated for a classification model.
JSON representation | |
---|---|
{ "prCurve": { object ( |
Fields | |
---|---|
prCurve |
Precision-recall curve based on ground truth labels, predicted labels, and scores for the predicted labels. |
confusionMatrix |
Confusion matrix of predicted labels vs. ground truth labels. |
PrCurve
JSON representation | |
---|---|
{ "annotationSpec": { object ( |
Fields | |
---|---|
annotationSpec |
The annotation spec of the label for which the precision-recall curve calculated. If this field is empty, that means the precision-recall curve is an aggregate curve for all labels. |
areaUnderCurve |
Area under the precision-recall curve. Not to be confused with area under a receiver operating characteristic (ROC) curve. |
confidenceMetricsEntries[] |
Entries that make up the precision-recall graph. Each entry is a "point" on the graph drawn for a different |
meanAveragePrecision |
Mean average prcision of this curve. |
ConfidenceMetricsEntry
JSON representation | |
---|---|
{ "confidenceThreshold": number, "recall": number, "precision": number, "f1Score": number, "recallAt1": number, "precisionAt1": number, "f1ScoreAt1": number, "recallAt5": number, "precisionAt5": number, "f1ScoreAt5": number } |
Fields | |
---|---|
confidenceThreshold |
Threshold used for this entry. For classification tasks, this is a classification threshold: a predicted label is categorized as positive or negative (in the context of this point on the PR curve) based on whether the label's score meets this threshold. For image object detection (bounding box) tasks, this is the [intersection-over-union (IOU)](/vision/automl/object-detection/docs/evaluate#intersection-over-union) threshold for the context of this point on the PR curve. |
recall |
Recall value. |
precision |
Precision value. |
f1Score |
Harmonic mean of recall and precision. |
recallAt1 |
Recall value for entries with label that has highest score. |
precisionAt1 |
Precision value for entries with label that has highest score. |
f1ScoreAt1 |
The harmonic mean of |
recallAt5 |
Recall value for entries with label that has highest 5 scores. |
precisionAt5 |
Precision value for entries with label that has highest 5 scores. |
f1ScoreAt5 |
The harmonic mean of |
ConfusionMatrix
Confusion matrix of the model running the classification. Only applicable when the metrics entry aggregates multiple labels. Not applicable when the entry is for a single label.
JSON representation | |
---|---|
{
"row": [
{
object ( |
Fields | |
---|---|
row[] |
|
Row
A row in the confusion matrix. Each entry in this row has the same ground truth label.
JSON representation | |
---|---|
{ "annotationSpec": { object ( |
Fields | |
---|---|
annotationSpec |
The annotation spec of the ground truth label for this row. |
entries[] |
A list of the confusion matrix entries. One entry for each possible predicted label. |
ConfusionMatrixEntry
JSON representation | |
---|---|
{
"annotationSpec": {
object ( |
Fields | |
---|---|
annotationSpec |
The annotation spec of a predicted label. |
itemCount |
Number of items predicted to have this label. (The ground truth label for these items is the |
ObjectDetectionMetrics
Metrics calculated for an image object detection (bounding box) model.
JSON representation | |
---|---|
{
"prCurve": {
object ( |
Fields | |
---|---|
prCurve |
Precision-recall curve. |
Methods |
|
---|---|
|
Gets an evaluation by resource name (to search, use projects.evaluations.search ). |