google.cloud.automl.v1.classification.proto Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of proto-google-cloud-automl-v1 Show documentation
Show all versions of proto-google-cloud-automl-v1 Show documentation
PROTO library for proto-google-cloud-automl-v1
// Copyright 2024 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
syntax = "proto3";
package google.cloud.automl.v1;
option csharp_namespace = "Google.Cloud.AutoML.V1";
option go_package = "cloud.google.com/go/automl/apiv1/automlpb;automlpb";
option java_multiple_files = true;
option java_outer_classname = "ClassificationProto";
option java_package = "com.google.cloud.automl.v1";
option php_namespace = "Google\\Cloud\\AutoMl\\V1";
option ruby_package = "Google::Cloud::AutoML::V1";
// Type of the classification problem.
enum ClassificationType {
// An un-set value of this enum.
CLASSIFICATION_TYPE_UNSPECIFIED = 0;
// At most one label is allowed per example.
MULTICLASS = 1;
// Multiple labels are allowed for one example.
MULTILABEL = 2;
}
// Contains annotation details specific to classification.
message ClassificationAnnotation {
// Output only. A confidence estimate between 0.0 and 1.0. A higher value
// means greater confidence that the annotation is positive. If a user
// approves an annotation as negative or positive, the score value remains
// unchanged. If a user creates an annotation, the score is 0 for negative or
// 1 for positive.
float score = 1;
}
// Model evaluation metrics for classification problems.
// Note: For Video Classification this metrics only describe quality of the
// Video Classification predictions of "segment_classification" type.
message ClassificationEvaluationMetrics {
// Metrics for a single confidence threshold.
message ConfidenceMetricsEntry {
// Output only. Metrics are computed with an assumption that the model
// never returns predictions with score lower than this value.
float confidence_threshold = 1;
// Output only. Metrics are computed with an assumption that the model
// always returns at most this many predictions (ordered by their score,
// descendingly), but they all still need to meet the confidence_threshold.
int32 position_threshold = 14;
// Output only. Recall (True Positive Rate) for the given confidence
// threshold.
float recall = 2;
// Output only. Precision for the given confidence threshold.
float precision = 3;
// Output only. False Positive Rate for the given confidence threshold.
float false_positive_rate = 8;
// Output only. The harmonic mean of recall and precision.
float f1_score = 4;
// Output only. The Recall (True Positive Rate) when only considering the
// label that has the highest prediction score and not below the confidence
// threshold for each example.
float recall_at1 = 5;
// Output only. The precision when only considering the label that has the
// highest prediction score and not below the confidence threshold for each
// example.
float precision_at1 = 6;
// Output only. The False Positive Rate when only considering the label that
// has the highest prediction score and not below the confidence threshold
// for each example.
float false_positive_rate_at1 = 9;
// Output only. The harmonic mean of [recall_at1][google.cloud.automl.v1.ClassificationEvaluationMetrics.ConfidenceMetricsEntry.recall_at1] and [precision_at1][google.cloud.automl.v1.ClassificationEvaluationMetrics.ConfidenceMetricsEntry.precision_at1].
float f1_score_at1 = 7;
// Output only. The number of model created labels that match a ground truth
// label.
int64 true_positive_count = 10;
// Output only. The number of model created labels that do not match a
// ground truth label.
int64 false_positive_count = 11;
// Output only. The number of ground truth labels that are not matched
// by a model created label.
int64 false_negative_count = 12;
// Output only. The number of labels that were not created by the model,
// but if they would, they would not match a ground truth label.
int64 true_negative_count = 13;
}
// Confusion matrix of the model running the classification.
message ConfusionMatrix {
// Output only. A row in the confusion matrix.
message Row {
// Output only. Value of the specific cell in the confusion matrix.
// The number of values each row has (i.e. the length of the row) is equal
// to the length of the `annotation_spec_id` field or, if that one is not
// populated, length of the [display_name][google.cloud.automl.v1.ClassificationEvaluationMetrics.ConfusionMatrix.display_name] field.
repeated int32 example_count = 1;
}
// Output only. IDs of the annotation specs used in the confusion matrix.
// For Tables CLASSIFICATION
// [prediction_type][google.cloud.automl.v1p1beta.TablesModelMetadata.prediction_type]
// only list of [annotation_spec_display_name-s][] is populated.
repeated string annotation_spec_id = 1;
// Output only. Display name of the annotation specs used in the confusion
// matrix, as they were at the moment of the evaluation. For Tables
// CLASSIFICATION
// [prediction_type-s][google.cloud.automl.v1p1beta.TablesModelMetadata.prediction_type],
// distinct values of the target column at the moment of the model
// evaluation are populated here.
repeated string display_name = 3;
// Output only. Rows in the confusion matrix. The number of rows is equal to
// the size of `annotation_spec_id`.
// `row[i].example_count[j]` is the number of examples that have ground
// truth of the `annotation_spec_id[i]` and are predicted as
// `annotation_spec_id[j]` by the model being evaluated.
repeated Row row = 2;
}
// Output only. The Area Under Precision-Recall Curve metric. Micro-averaged
// for the overall evaluation.
float au_prc = 1;
// Output only. The Area Under Receiver Operating Characteristic curve metric.
// Micro-averaged for the overall evaluation.
float au_roc = 6;
// Output only. The Log Loss metric.
float log_loss = 7;
// Output only. Metrics for each confidence_threshold in
// 0.00,0.05,0.10,...,0.95,0.96,0.97,0.98,0.99 and
// position_threshold = INT32_MAX_VALUE.
// ROC and precision-recall curves, and other aggregated metrics are derived
// from them. The confidence metrics entries may also be supplied for
// additional values of position_threshold, but from these no aggregated
// metrics are computed.
repeated ConfidenceMetricsEntry confidence_metrics_entry = 3;
// Output only. Confusion matrix of the evaluation.
// Only set for MULTICLASS classification problems where number
// of labels is no more than 10.
// Only set for model level evaluation, not for evaluation per label.
ConfusionMatrix confusion_matrix = 4;
// Output only. The annotation spec ids used for this evaluation.
repeated string annotation_spec_id = 5;
}