| // Code generated by protoc-gen-go. DO NOT EDIT. |
| // source: google/cloud/automl/v1beta1/classification.proto |
| |
| package automl |
| |
| import ( |
| fmt "fmt" |
| math "math" |
| |
| proto "github.com/golang/protobuf/proto" |
| _ "google.golang.org/genproto/googleapis/api/annotations" |
| ) |
| |
| // Reference imports to suppress errors if they are not otherwise used. |
| var _ = proto.Marshal |
| var _ = fmt.Errorf |
| var _ = math.Inf |
| |
| // This is a compile-time assertion to ensure that this generated file |
| // is compatible with the proto package it is being compiled against. |
| // A compilation error at this line likely means your copy of the |
| // proto package needs to be updated. |
| const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package |
| |
| // Type of the classification problem. |
| type ClassificationType int32 |
| |
| const ( |
| // An un-set value of this enum. |
| ClassificationType_CLASSIFICATION_TYPE_UNSPECIFIED ClassificationType = 0 |
| // At most one label is allowed per example. |
| ClassificationType_MULTICLASS ClassificationType = 1 |
| // Multiple labels are allowed for one example. |
| ClassificationType_MULTILABEL ClassificationType = 2 |
| ) |
| |
| var ClassificationType_name = map[int32]string{ |
| 0: "CLASSIFICATION_TYPE_UNSPECIFIED", |
| 1: "MULTICLASS", |
| 2: "MULTILABEL", |
| } |
| |
| var ClassificationType_value = map[string]int32{ |
| "CLASSIFICATION_TYPE_UNSPECIFIED": 0, |
| "MULTICLASS": 1, |
| "MULTILABEL": 2, |
| } |
| |
| func (x ClassificationType) String() string { |
| return proto.EnumName(ClassificationType_name, int32(x)) |
| } |
| |
| func (ClassificationType) EnumDescriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{0} |
| } |
| |
| // Contains annotation details specific to classification. |
| type ClassificationAnnotation struct { |
| // Output only. A confidence estimate between 0.0 and 1.0. A higher value |
| // means greater confidence that the annotation is positive. If a user |
| // approves an annotation as negative or positive, the score value remains |
| // unchanged. If a user creates an annotation, the score is 0 for negative or |
| // 1 for positive. |
| Score float32 `protobuf:"fixed32,1,opt,name=score,proto3" json:"score,omitempty"` |
| XXX_NoUnkeyedLiteral struct{} `json:"-"` |
| XXX_unrecognized []byte `json:"-"` |
| XXX_sizecache int32 `json:"-"` |
| } |
| |
| func (m *ClassificationAnnotation) Reset() { *m = ClassificationAnnotation{} } |
| func (m *ClassificationAnnotation) String() string { return proto.CompactTextString(m) } |
| func (*ClassificationAnnotation) ProtoMessage() {} |
| func (*ClassificationAnnotation) Descriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{0} |
| } |
| |
| func (m *ClassificationAnnotation) XXX_Unmarshal(b []byte) error { |
| return xxx_messageInfo_ClassificationAnnotation.Unmarshal(m, b) |
| } |
| func (m *ClassificationAnnotation) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { |
| return xxx_messageInfo_ClassificationAnnotation.Marshal(b, m, deterministic) |
| } |
| func (m *ClassificationAnnotation) XXX_Merge(src proto.Message) { |
| xxx_messageInfo_ClassificationAnnotation.Merge(m, src) |
| } |
| func (m *ClassificationAnnotation) XXX_Size() int { |
| return xxx_messageInfo_ClassificationAnnotation.Size(m) |
| } |
| func (m *ClassificationAnnotation) XXX_DiscardUnknown() { |
| xxx_messageInfo_ClassificationAnnotation.DiscardUnknown(m) |
| } |
| |
| var xxx_messageInfo_ClassificationAnnotation proto.InternalMessageInfo |
| |
| func (m *ClassificationAnnotation) GetScore() float32 { |
| if m != nil { |
| return m.Score |
| } |
| return 0 |
| } |
| |
| // Contains annotation details specific to video classification. |
| type VideoClassificationAnnotation struct { |
| // Output only. Expresses the type of video classification. Possible values: |
| // |
| // * `segment` - Classification done on a specified by user |
| // time segment of a video. AnnotationSpec is answered to be present |
| // in that time segment, if it is present in any part of it. The video |
| // ML model evaluations are done only for this type of classification. |
| // |
| // * `shot`- Shot-level classification. |
| // AutoML Video Intelligence determines the boundaries |
| // for each camera shot in the entire segment of the video that user |
| // specified in the request configuration. AutoML Video Intelligence |
| // then returns labels and their confidence scores for each detected |
| // shot, along with the start and end time of the shot. |
| // WARNING: Model evaluation is not done for this classification type, |
| // the quality of it depends on training data, but there are no |
| // metrics provided to describe that quality. |
| // |
| // * `1s_interval` - AutoML Video Intelligence returns labels and their |
| // confidence scores for each second of the entire segment of the video |
| // that user specified in the request configuration. |
| // WARNING: Model evaluation is not done for this classification type, |
| // the quality of it depends on training data, but there are no |
| // metrics provided to describe that quality. |
| Type string `protobuf:"bytes,1,opt,name=type,proto3" json:"type,omitempty"` |
| // Output only . The classification details of this annotation. |
| ClassificationAnnotation *ClassificationAnnotation `protobuf:"bytes,2,opt,name=classification_annotation,json=classificationAnnotation,proto3" json:"classification_annotation,omitempty"` |
| // Output only . The time segment of the video to which the |
| // annotation applies. |
| TimeSegment *TimeSegment `protobuf:"bytes,3,opt,name=time_segment,json=timeSegment,proto3" json:"time_segment,omitempty"` |
| XXX_NoUnkeyedLiteral struct{} `json:"-"` |
| XXX_unrecognized []byte `json:"-"` |
| XXX_sizecache int32 `json:"-"` |
| } |
| |
| func (m *VideoClassificationAnnotation) Reset() { *m = VideoClassificationAnnotation{} } |
| func (m *VideoClassificationAnnotation) String() string { return proto.CompactTextString(m) } |
| func (*VideoClassificationAnnotation) ProtoMessage() {} |
| func (*VideoClassificationAnnotation) Descriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{1} |
| } |
| |
| func (m *VideoClassificationAnnotation) XXX_Unmarshal(b []byte) error { |
| return xxx_messageInfo_VideoClassificationAnnotation.Unmarshal(m, b) |
| } |
| func (m *VideoClassificationAnnotation) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { |
| return xxx_messageInfo_VideoClassificationAnnotation.Marshal(b, m, deterministic) |
| } |
| func (m *VideoClassificationAnnotation) XXX_Merge(src proto.Message) { |
| xxx_messageInfo_VideoClassificationAnnotation.Merge(m, src) |
| } |
| func (m *VideoClassificationAnnotation) XXX_Size() int { |
| return xxx_messageInfo_VideoClassificationAnnotation.Size(m) |
| } |
| func (m *VideoClassificationAnnotation) XXX_DiscardUnknown() { |
| xxx_messageInfo_VideoClassificationAnnotation.DiscardUnknown(m) |
| } |
| |
| var xxx_messageInfo_VideoClassificationAnnotation proto.InternalMessageInfo |
| |
| func (m *VideoClassificationAnnotation) GetType() string { |
| if m != nil { |
| return m.Type |
| } |
| return "" |
| } |
| |
| func (m *VideoClassificationAnnotation) GetClassificationAnnotation() *ClassificationAnnotation { |
| if m != nil { |
| return m.ClassificationAnnotation |
| } |
| return nil |
| } |
| |
| func (m *VideoClassificationAnnotation) GetTimeSegment() *TimeSegment { |
| if m != nil { |
| return m.TimeSegment |
| } |
| return nil |
| } |
| |
| // Model evaluation metrics for classification problems. |
| // Note: For Video Classification this metrics only describe quality of the |
| // Video Classification predictions of "segment_classification" type. |
| type ClassificationEvaluationMetrics struct { |
| // Output only. The Area Under Precision-Recall Curve metric. Micro-averaged |
| // for the overall evaluation. |
| AuPrc float32 `protobuf:"fixed32,1,opt,name=au_prc,json=auPrc,proto3" json:"au_prc,omitempty"` |
| // Output only. The Area Under Precision-Recall Curve metric based on priors. |
| // Micro-averaged for the overall evaluation. |
| // Deprecated. |
| BaseAuPrc float32 `protobuf:"fixed32,2,opt,name=base_au_prc,json=baseAuPrc,proto3" json:"base_au_prc,omitempty"` // Deprecated: Do not use. |
| // Output only. The Area Under Receiver Operating Characteristic curve metric. |
| // Micro-averaged for the overall evaluation. |
| AuRoc float32 `protobuf:"fixed32,6,opt,name=au_roc,json=auRoc,proto3" json:"au_roc,omitempty"` |
| // Output only. The Log Loss metric. |
| LogLoss float32 `protobuf:"fixed32,7,opt,name=log_loss,json=logLoss,proto3" json:"log_loss,omitempty"` |
| // Output only. Metrics for each confidence_threshold in |
| // 0.00,0.05,0.10,...,0.95,0.96,0.97,0.98,0.99 and |
| // position_threshold = INT32_MAX_VALUE. |
| // ROC and precision-recall curves, and other aggregated metrics are derived |
| // from them. The confidence metrics entries may also be supplied for |
| // additional values of position_threshold, but from these no aggregated |
| // metrics are computed. |
| ConfidenceMetricsEntry []*ClassificationEvaluationMetrics_ConfidenceMetricsEntry `protobuf:"bytes,3,rep,name=confidence_metrics_entry,json=confidenceMetricsEntry,proto3" json:"confidence_metrics_entry,omitempty"` |
| // Output only. Confusion matrix of the evaluation. |
| // Only set for MULTICLASS classification problems where number |
| // of labels is no more than 10. |
| // Only set for model level evaluation, not for evaluation per label. |
| ConfusionMatrix *ClassificationEvaluationMetrics_ConfusionMatrix `protobuf:"bytes,4,opt,name=confusion_matrix,json=confusionMatrix,proto3" json:"confusion_matrix,omitempty"` |
| // Output only. The annotation spec ids used for this evaluation. |
| AnnotationSpecId []string `protobuf:"bytes,5,rep,name=annotation_spec_id,json=annotationSpecId,proto3" json:"annotation_spec_id,omitempty"` |
| XXX_NoUnkeyedLiteral struct{} `json:"-"` |
| XXX_unrecognized []byte `json:"-"` |
| XXX_sizecache int32 `json:"-"` |
| } |
| |
| func (m *ClassificationEvaluationMetrics) Reset() { *m = ClassificationEvaluationMetrics{} } |
| func (m *ClassificationEvaluationMetrics) String() string { return proto.CompactTextString(m) } |
| func (*ClassificationEvaluationMetrics) ProtoMessage() {} |
| func (*ClassificationEvaluationMetrics) Descriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{2} |
| } |
| |
| func (m *ClassificationEvaluationMetrics) XXX_Unmarshal(b []byte) error { |
| return xxx_messageInfo_ClassificationEvaluationMetrics.Unmarshal(m, b) |
| } |
| func (m *ClassificationEvaluationMetrics) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { |
| return xxx_messageInfo_ClassificationEvaluationMetrics.Marshal(b, m, deterministic) |
| } |
| func (m *ClassificationEvaluationMetrics) XXX_Merge(src proto.Message) { |
| xxx_messageInfo_ClassificationEvaluationMetrics.Merge(m, src) |
| } |
| func (m *ClassificationEvaluationMetrics) XXX_Size() int { |
| return xxx_messageInfo_ClassificationEvaluationMetrics.Size(m) |
| } |
| func (m *ClassificationEvaluationMetrics) XXX_DiscardUnknown() { |
| xxx_messageInfo_ClassificationEvaluationMetrics.DiscardUnknown(m) |
| } |
| |
| var xxx_messageInfo_ClassificationEvaluationMetrics proto.InternalMessageInfo |
| |
| func (m *ClassificationEvaluationMetrics) GetAuPrc() float32 { |
| if m != nil { |
| return m.AuPrc |
| } |
| return 0 |
| } |
| |
| // Deprecated: Do not use. |
| func (m *ClassificationEvaluationMetrics) GetBaseAuPrc() float32 { |
| if m != nil { |
| return m.BaseAuPrc |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics) GetAuRoc() float32 { |
| if m != nil { |
| return m.AuRoc |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics) GetLogLoss() float32 { |
| if m != nil { |
| return m.LogLoss |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics) GetConfidenceMetricsEntry() []*ClassificationEvaluationMetrics_ConfidenceMetricsEntry { |
| if m != nil { |
| return m.ConfidenceMetricsEntry |
| } |
| return nil |
| } |
| |
| func (m *ClassificationEvaluationMetrics) GetConfusionMatrix() *ClassificationEvaluationMetrics_ConfusionMatrix { |
| if m != nil { |
| return m.ConfusionMatrix |
| } |
| return nil |
| } |
| |
| func (m *ClassificationEvaluationMetrics) GetAnnotationSpecId() []string { |
| if m != nil { |
| return m.AnnotationSpecId |
| } |
| return nil |
| } |
| |
| // Metrics for a single confidence threshold. |
| type ClassificationEvaluationMetrics_ConfidenceMetricsEntry struct { |
| // Output only. Metrics are computed with an assumption that the model |
| // never returns predictions with score lower than this value. |
| ConfidenceThreshold float32 `protobuf:"fixed32,1,opt,name=confidence_threshold,json=confidenceThreshold,proto3" json:"confidence_threshold,omitempty"` |
| // Output only. Metrics are computed with an assumption that the model |
| // always returns at most this many predictions (ordered by their score, |
| // descendingly), but they all still need to meet the confidence_threshold. |
| PositionThreshold int32 `protobuf:"varint,14,opt,name=position_threshold,json=positionThreshold,proto3" json:"position_threshold,omitempty"` |
| // Output only. Recall (True Positive Rate) for the given confidence |
| // threshold. |
| Recall float32 `protobuf:"fixed32,2,opt,name=recall,proto3" json:"recall,omitempty"` |
| // Output only. Precision for the given confidence threshold. |
| Precision float32 `protobuf:"fixed32,3,opt,name=precision,proto3" json:"precision,omitempty"` |
| // Output only. False Positive Rate for the given confidence threshold. |
| FalsePositiveRate float32 `protobuf:"fixed32,8,opt,name=false_positive_rate,json=falsePositiveRate,proto3" json:"false_positive_rate,omitempty"` |
| // Output only. The harmonic mean of recall and precision. |
| F1Score float32 `protobuf:"fixed32,4,opt,name=f1_score,json=f1Score,proto3" json:"f1_score,omitempty"` |
| // Output only. The Recall (True Positive Rate) when only considering the |
| // label that has the highest prediction score and not below the confidence |
| // threshold for each example. |
| RecallAt1 float32 `protobuf:"fixed32,5,opt,name=recall_at1,json=recallAt1,proto3" json:"recall_at1,omitempty"` |
| // Output only. The precision when only considering the label that has the |
| // highest prediction score and not below the confidence threshold for each |
| // example. |
| PrecisionAt1 float32 `protobuf:"fixed32,6,opt,name=precision_at1,json=precisionAt1,proto3" json:"precision_at1,omitempty"` |
| // Output only. The False Positive Rate when only considering the label that |
| // has the highest prediction score and not below the confidence threshold |
| // for each example. |
| FalsePositiveRateAt1 float32 `protobuf:"fixed32,9,opt,name=false_positive_rate_at1,json=falsePositiveRateAt1,proto3" json:"false_positive_rate_at1,omitempty"` |
| // Output only. The harmonic mean of [recall_at1][google.cloud.automl.v1beta1.ClassificationEvaluationMetrics.ConfidenceMetricsEntry.recall_at1] and [precision_at1][google.cloud.automl.v1beta1.ClassificationEvaluationMetrics.ConfidenceMetricsEntry.precision_at1]. |
| F1ScoreAt1 float32 `protobuf:"fixed32,7,opt,name=f1_score_at1,json=f1ScoreAt1,proto3" json:"f1_score_at1,omitempty"` |
| // Output only. The number of model created labels that match a ground truth |
| // label. |
| TruePositiveCount int64 `protobuf:"varint,10,opt,name=true_positive_count,json=truePositiveCount,proto3" json:"true_positive_count,omitempty"` |
| // Output only. The number of model created labels that do not match a |
| // ground truth label. |
| FalsePositiveCount int64 `protobuf:"varint,11,opt,name=false_positive_count,json=falsePositiveCount,proto3" json:"false_positive_count,omitempty"` |
| // Output only. The number of ground truth labels that are not matched |
| // by a model created label. |
| FalseNegativeCount int64 `protobuf:"varint,12,opt,name=false_negative_count,json=falseNegativeCount,proto3" json:"false_negative_count,omitempty"` |
| // Output only. The number of labels that were not created by the model, |
| // but if they would, they would not match a ground truth label. |
| TrueNegativeCount int64 `protobuf:"varint,13,opt,name=true_negative_count,json=trueNegativeCount,proto3" json:"true_negative_count,omitempty"` |
| XXX_NoUnkeyedLiteral struct{} `json:"-"` |
| XXX_unrecognized []byte `json:"-"` |
| XXX_sizecache int32 `json:"-"` |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) Reset() { |
| *m = ClassificationEvaluationMetrics_ConfidenceMetricsEntry{} |
| } |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) String() string { |
| return proto.CompactTextString(m) |
| } |
| func (*ClassificationEvaluationMetrics_ConfidenceMetricsEntry) ProtoMessage() {} |
| func (*ClassificationEvaluationMetrics_ConfidenceMetricsEntry) Descriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{2, 0} |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) XXX_Unmarshal(b []byte) error { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfidenceMetricsEntry.Unmarshal(m, b) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfidenceMetricsEntry.Marshal(b, m, deterministic) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) XXX_Merge(src proto.Message) { |
| xxx_messageInfo_ClassificationEvaluationMetrics_ConfidenceMetricsEntry.Merge(m, src) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) XXX_Size() int { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfidenceMetricsEntry.Size(m) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) XXX_DiscardUnknown() { |
| xxx_messageInfo_ClassificationEvaluationMetrics_ConfidenceMetricsEntry.DiscardUnknown(m) |
| } |
| |
| var xxx_messageInfo_ClassificationEvaluationMetrics_ConfidenceMetricsEntry proto.InternalMessageInfo |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetConfidenceThreshold() float32 { |
| if m != nil { |
| return m.ConfidenceThreshold |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetPositionThreshold() int32 { |
| if m != nil { |
| return m.PositionThreshold |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetRecall() float32 { |
| if m != nil { |
| return m.Recall |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetPrecision() float32 { |
| if m != nil { |
| return m.Precision |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetFalsePositiveRate() float32 { |
| if m != nil { |
| return m.FalsePositiveRate |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetF1Score() float32 { |
| if m != nil { |
| return m.F1Score |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetRecallAt1() float32 { |
| if m != nil { |
| return m.RecallAt1 |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetPrecisionAt1() float32 { |
| if m != nil { |
| return m.PrecisionAt1 |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetFalsePositiveRateAt1() float32 { |
| if m != nil { |
| return m.FalsePositiveRateAt1 |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetF1ScoreAt1() float32 { |
| if m != nil { |
| return m.F1ScoreAt1 |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetTruePositiveCount() int64 { |
| if m != nil { |
| return m.TruePositiveCount |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetFalsePositiveCount() int64 { |
| if m != nil { |
| return m.FalsePositiveCount |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetFalseNegativeCount() int64 { |
| if m != nil { |
| return m.FalseNegativeCount |
| } |
| return 0 |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfidenceMetricsEntry) GetTrueNegativeCount() int64 { |
| if m != nil { |
| return m.TrueNegativeCount |
| } |
| return 0 |
| } |
| |
| // Confusion matrix of the model running the classification. |
| type ClassificationEvaluationMetrics_ConfusionMatrix struct { |
| // Output only. IDs of the annotation specs used in the confusion matrix. |
| // For Tables CLASSIFICATION |
| // |
| // [prediction_type][google.cloud.automl.v1beta1.TablesModelMetadata.prediction_type] |
| // only list of [annotation_spec_display_name-s][] is populated. |
| AnnotationSpecId []string `protobuf:"bytes,1,rep,name=annotation_spec_id,json=annotationSpecId,proto3" json:"annotation_spec_id,omitempty"` |
| // Output only. Display name of the annotation specs used in the confusion |
| // matrix, as they were at the moment of the evaluation. For Tables |
| // CLASSIFICATION |
| // |
| // [prediction_type-s][google.cloud.automl.v1beta1.TablesModelMetadata.prediction_type], |
| // distinct values of the target column at the moment of the model |
| // evaluation are populated here. |
| DisplayName []string `protobuf:"bytes,3,rep,name=display_name,json=displayName,proto3" json:"display_name,omitempty"` |
| // Output only. Rows in the confusion matrix. The number of rows is equal to |
| // the size of `annotation_spec_id`. |
| // `row[i].value[j]` is the number of examples that have ground truth of the |
| // `annotation_spec_id[i]` and are predicted as `annotation_spec_id[j]` by |
| // the model being evaluated. |
| Row []*ClassificationEvaluationMetrics_ConfusionMatrix_Row `protobuf:"bytes,2,rep,name=row,proto3" json:"row,omitempty"` |
| XXX_NoUnkeyedLiteral struct{} `json:"-"` |
| XXX_unrecognized []byte `json:"-"` |
| XXX_sizecache int32 `json:"-"` |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) Reset() { |
| *m = ClassificationEvaluationMetrics_ConfusionMatrix{} |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) String() string { |
| return proto.CompactTextString(m) |
| } |
| func (*ClassificationEvaluationMetrics_ConfusionMatrix) ProtoMessage() {} |
| func (*ClassificationEvaluationMetrics_ConfusionMatrix) Descriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{2, 1} |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) XXX_Unmarshal(b []byte) error { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix.Unmarshal(m, b) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix.Marshal(b, m, deterministic) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) XXX_Merge(src proto.Message) { |
| xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix.Merge(m, src) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) XXX_Size() int { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix.Size(m) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) XXX_DiscardUnknown() { |
| xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix.DiscardUnknown(m) |
| } |
| |
| var xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix proto.InternalMessageInfo |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) GetAnnotationSpecId() []string { |
| if m != nil { |
| return m.AnnotationSpecId |
| } |
| return nil |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) GetDisplayName() []string { |
| if m != nil { |
| return m.DisplayName |
| } |
| return nil |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix) GetRow() []*ClassificationEvaluationMetrics_ConfusionMatrix_Row { |
| if m != nil { |
| return m.Row |
| } |
| return nil |
| } |
| |
| // Output only. A row in the confusion matrix. |
| type ClassificationEvaluationMetrics_ConfusionMatrix_Row struct { |
| // Output only. Value of the specific cell in the confusion matrix. |
| // The number of values each row has (i.e. the length of the row) is equal |
| // to the length of the `annotation_spec_id` field or, if that one is not |
| // populated, length of the [display_name][google.cloud.automl.v1beta1.ClassificationEvaluationMetrics.ConfusionMatrix.display_name] field. |
| ExampleCount []int32 `protobuf:"varint,1,rep,packed,name=example_count,json=exampleCount,proto3" json:"example_count,omitempty"` |
| XXX_NoUnkeyedLiteral struct{} `json:"-"` |
| XXX_unrecognized []byte `json:"-"` |
| XXX_sizecache int32 `json:"-"` |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) Reset() { |
| *m = ClassificationEvaluationMetrics_ConfusionMatrix_Row{} |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) String() string { |
| return proto.CompactTextString(m) |
| } |
| func (*ClassificationEvaluationMetrics_ConfusionMatrix_Row) ProtoMessage() {} |
| func (*ClassificationEvaluationMetrics_ConfusionMatrix_Row) Descriptor() ([]byte, []int) { |
| return fileDescriptor_7b436fefe6ae5367, []int{2, 1, 0} |
| } |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) XXX_Unmarshal(b []byte) error { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix_Row.Unmarshal(m, b) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix_Row.Marshal(b, m, deterministic) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) XXX_Merge(src proto.Message) { |
| xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix_Row.Merge(m, src) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) XXX_Size() int { |
| return xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix_Row.Size(m) |
| } |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) XXX_DiscardUnknown() { |
| xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix_Row.DiscardUnknown(m) |
| } |
| |
| var xxx_messageInfo_ClassificationEvaluationMetrics_ConfusionMatrix_Row proto.InternalMessageInfo |
| |
| func (m *ClassificationEvaluationMetrics_ConfusionMatrix_Row) GetExampleCount() []int32 { |
| if m != nil { |
| return m.ExampleCount |
| } |
| return nil |
| } |
| |
| func init() { |
| proto.RegisterEnum("google.cloud.automl.v1beta1.ClassificationType", ClassificationType_name, ClassificationType_value) |
| proto.RegisterType((*ClassificationAnnotation)(nil), "google.cloud.automl.v1beta1.ClassificationAnnotation") |
| proto.RegisterType((*VideoClassificationAnnotation)(nil), "google.cloud.automl.v1beta1.VideoClassificationAnnotation") |
| proto.RegisterType((*ClassificationEvaluationMetrics)(nil), "google.cloud.automl.v1beta1.ClassificationEvaluationMetrics") |
| proto.RegisterType((*ClassificationEvaluationMetrics_ConfidenceMetricsEntry)(nil), "google.cloud.automl.v1beta1.ClassificationEvaluationMetrics.ConfidenceMetricsEntry") |
| proto.RegisterType((*ClassificationEvaluationMetrics_ConfusionMatrix)(nil), "google.cloud.automl.v1beta1.ClassificationEvaluationMetrics.ConfusionMatrix") |
| proto.RegisterType((*ClassificationEvaluationMetrics_ConfusionMatrix_Row)(nil), "google.cloud.automl.v1beta1.ClassificationEvaluationMetrics.ConfusionMatrix.Row") |
| } |
| |
| func init() { |
| proto.RegisterFile("google/cloud/automl/v1beta1/classification.proto", fileDescriptor_7b436fefe6ae5367) |
| } |
| |
| var fileDescriptor_7b436fefe6ae5367 = []byte{ |
| // 881 bytes of a gzipped FileDescriptorProto |
| 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xac, 0x55, 0xd1, 0x6e, 0xdb, 0x36, |
| 0x14, 0x9d, 0xe4, 0x38, 0xad, 0xaf, 0xdd, 0xd6, 0x61, 0xb2, 0x4c, 0x75, 0xdb, 0xc5, 0x4b, 0x5f, |
| 0x8c, 0x60, 0x93, 0xe3, 0x0e, 0x7d, 0xf1, 0x9e, 0x1c, 0xcf, 0x1d, 0x8c, 0x39, 0x99, 0x41, 0xbb, |
| 0x05, 0x3a, 0x04, 0x10, 0x18, 0x9a, 0x56, 0x05, 0x48, 0xa2, 0x20, 0x52, 0x49, 0xf3, 0x11, 0xfb, |
| 0x80, 0xfd, 0xc9, 0xbe, 0x61, 0xdf, 0xb0, 0x2f, 0xd8, 0xe3, 0xb0, 0x0f, 0x18, 0x48, 0xca, 0x56, |
| 0x9c, 0x38, 0xc6, 0x86, 0xf5, 0xcd, 0xf7, 0x9e, 0x73, 0xee, 0x3d, 0x97, 0xba, 0x34, 0xe1, 0xd8, |
| 0xe7, 0xdc, 0x0f, 0x59, 0x9b, 0x86, 0x3c, 0x9b, 0xb5, 0x49, 0x26, 0x79, 0x14, 0xb6, 0x2f, 0x3b, |
| 0x17, 0x4c, 0x92, 0x4e, 0x9b, 0x86, 0x44, 0x88, 0x60, 0x1e, 0x50, 0x22, 0x03, 0x1e, 0xbb, 0x49, |
| 0xca, 0x25, 0x47, 0xcf, 0x8c, 0xc2, 0xd5, 0x0a, 0xd7, 0x28, 0xdc, 0x5c, 0xd1, 0x38, 0xda, 0x54, |
| 0x4e, 0xb2, 0x28, 0xe1, 0x29, 0x09, 0x4d, 0xa1, 0xc6, 0xf3, 0x9c, 0x4b, 0x92, 0xa0, 0x4d, 0xe2, |
| 0x98, 0x4b, 0xdd, 0x45, 0x18, 0xf4, 0xf0, 0x18, 0x9c, 0xfe, 0x4a, 0xfb, 0xde, 0x92, 0x82, 0xf6, |
| 0xa0, 0x2c, 0x28, 0x4f, 0x99, 0x63, 0x35, 0xad, 0x96, 0x8d, 0x4d, 0x70, 0xf8, 0xb7, 0x05, 0x2f, |
| 0xde, 0x05, 0x33, 0xc6, 0xef, 0xd5, 0x21, 0xd8, 0x92, 0xd7, 0x89, 0x91, 0x55, 0xb0, 0xfe, 0x8d, |
| 0x52, 0x78, 0xba, 0x3a, 0xa6, 0x57, 0x78, 0x71, 0xec, 0xa6, 0xd5, 0xaa, 0xbe, 0x7a, 0xed, 0x6e, |
| 0x18, 0xd9, 0xbd, 0xaf, 0x1b, 0x76, 0xe8, 0x7d, 0x3e, 0x7e, 0x84, 0x9a, 0x0c, 0x22, 0xe6, 0x09, |
| 0xe6, 0x47, 0x2c, 0x96, 0x4e, 0x49, 0xb7, 0x69, 0x6d, 0x6c, 0x33, 0x0d, 0x22, 0x36, 0x31, 0x7c, |
| 0x5c, 0x95, 0x45, 0x70, 0xf8, 0x2b, 0xc0, 0xc1, 0xaa, 0x87, 0xc1, 0x25, 0x09, 0x33, 0xfd, 0xeb, |
| 0x94, 0xc9, 0x34, 0xa0, 0x02, 0x7d, 0x0e, 0xdb, 0x24, 0xf3, 0x92, 0x94, 0x2e, 0x4e, 0x8c, 0x64, |
| 0xe3, 0x94, 0xa2, 0x43, 0xa8, 0x5e, 0x10, 0xc1, 0xbc, 0x1c, 0x53, 0xd3, 0xda, 0x27, 0xb6, 0x63, |
| 0xe1, 0x8a, 0x4a, 0xf7, 0x34, 0xc7, 0x48, 0x53, 0x4e, 0x9d, 0xed, 0x85, 0x14, 0x73, 0x8a, 0x9e, |
| 0xc2, 0xc3, 0x90, 0xfb, 0x5e, 0xc8, 0x85, 0x70, 0x1e, 0x68, 0xe0, 0x41, 0xc8, 0xfd, 0x11, 0x17, |
| 0x02, 0xfd, 0x62, 0x81, 0x43, 0x79, 0x3c, 0x0f, 0x66, 0x2c, 0xa6, 0xcc, 0x8b, 0x8c, 0x07, 0x8f, |
| 0xc5, 0x32, 0xbd, 0x76, 0x4a, 0xcd, 0x52, 0xab, 0xfa, 0x6a, 0xf2, 0x1f, 0x4e, 0xf4, 0xce, 0x34, |
| 0x6e, 0x7f, 0x59, 0x3c, 0xcf, 0x0c, 0x54, 0x69, 0xbc, 0x4f, 0xd7, 0xe6, 0xd1, 0x15, 0xd4, 0x15, |
| 0x92, 0x09, 0xf5, 0x71, 0x23, 0x22, 0xd3, 0xe0, 0xa3, 0xb3, 0xa5, 0x4f, 0x7c, 0xf4, 0xbf, 0x6d, |
| 0xe8, 0xa2, 0xa7, 0xba, 0x26, 0x7e, 0x42, 0x57, 0x13, 0xe8, 0x6b, 0x40, 0xc5, 0x2e, 0x79, 0x22, |
| 0x61, 0xd4, 0x0b, 0x66, 0x4e, 0xb9, 0x59, 0x6a, 0x55, 0x70, 0xbd, 0x40, 0x26, 0x09, 0xa3, 0xc3, |
| 0x59, 0xe3, 0x8f, 0x2d, 0xd8, 0x5f, 0x3f, 0x19, 0xea, 0xc0, 0xde, 0x8d, 0x03, 0x95, 0x1f, 0x52, |
| 0x26, 0x3e, 0xf0, 0x70, 0x96, 0x7f, 0xcc, 0xdd, 0x02, 0x9b, 0x2e, 0x20, 0xf4, 0x0d, 0xa0, 0x84, |
| 0x8b, 0x40, 0x77, 0x2e, 0x04, 0x8f, 0x9b, 0x56, 0xab, 0x8c, 0x77, 0x16, 0x48, 0x41, 0xdf, 0x87, |
| 0xed, 0x94, 0x51, 0x12, 0x86, 0x66, 0x09, 0x70, 0x1e, 0xa1, 0xe7, 0x50, 0x49, 0x52, 0x46, 0x03, |
| 0x35, 0x95, 0x5e, 0x53, 0x1b, 0x17, 0x09, 0xe4, 0xc2, 0xee, 0x9c, 0x84, 0x82, 0x79, 0xa6, 0xe0, |
| 0x25, 0xf3, 0x52, 0x22, 0x99, 0xf3, 0x50, 0xf3, 0x76, 0x34, 0x34, 0xce, 0x11, 0x4c, 0x24, 0x53, |
| 0x4b, 0x33, 0xef, 0x78, 0xe6, 0xea, 0x6e, 0x99, 0xa5, 0x99, 0x77, 0x26, 0x2a, 0x44, 0x2f, 0x00, |
| 0x4c, 0x4b, 0x8f, 0xc8, 0x8e, 0x53, 0x36, 0x9d, 0x4c, 0xa6, 0x27, 0x3b, 0xe8, 0x25, 0x3c, 0x5a, |
| 0xb6, 0xd5, 0x0c, 0xb3, 0x8c, 0xb5, 0x65, 0x52, 0x91, 0x5e, 0xc3, 0x17, 0x6b, 0xec, 0x68, 0x7a, |
| 0x45, 0xd3, 0xf7, 0xee, 0x58, 0x52, 0xb2, 0x26, 0xd4, 0x16, 0xae, 0x34, 0xd7, 0xac, 0x33, 0xe4, |
| 0xce, 0x14, 0xc3, 0x85, 0x5d, 0x99, 0x66, 0x37, 0xea, 0x52, 0x9e, 0xc5, 0xd2, 0x81, 0xa6, 0xd5, |
| 0x2a, 0xe1, 0x1d, 0x05, 0x2d, 0x6a, 0xf6, 0x15, 0x80, 0x8e, 0x61, 0xef, 0x96, 0x11, 0x23, 0xa8, |
| 0x6a, 0x01, 0x5a, 0x71, 0x71, 0x4b, 0x11, 0x33, 0x9f, 0xdc, 0x50, 0xd4, 0x6e, 0x28, 0xce, 0x72, |
| 0xc8, 0x28, 0x16, 0x9e, 0x6e, 0x09, 0x1e, 0x15, 0x9e, 0x56, 0xf8, 0x8d, 0xbf, 0x2c, 0x78, 0xd2, |
| 0xff, 0x57, 0x0b, 0x6a, 0xad, 0x5f, 0x50, 0xf4, 0x15, 0xd4, 0x66, 0x81, 0x48, 0x42, 0x72, 0xed, |
| 0xc5, 0x24, 0x62, 0xfa, 0x2a, 0x57, 0x70, 0x35, 0xcf, 0x9d, 0x91, 0x88, 0xa1, 0x0b, 0x28, 0xa5, |
| 0xfc, 0xca, 0xb1, 0xf5, 0x25, 0x1f, 0x7f, 0xca, 0xdb, 0xe5, 0x62, 0x7e, 0x85, 0x55, 0xf1, 0xc6, |
| 0x11, 0x94, 0x30, 0xbf, 0x52, 0x1b, 0xc1, 0x3e, 0x92, 0x28, 0x09, 0x17, 0x93, 0x2b, 0xdb, 0x65, |
| 0x5c, 0xcb, 0x93, 0x7a, 0xe8, 0xa3, 0xf7, 0x80, 0x56, 0xfb, 0x4c, 0xd5, 0x5f, 0xfe, 0x4b, 0x38, |
| 0xe8, 0x8f, 0x7a, 0x93, 0xc9, 0xf0, 0xcd, 0xb0, 0xdf, 0x9b, 0x0e, 0x7f, 0x3a, 0xf3, 0xa6, 0xef, |
| 0xc7, 0x03, 0xef, 0xed, 0xd9, 0x64, 0x3c, 0xe8, 0x0f, 0xdf, 0x0c, 0x07, 0xdf, 0xd7, 0x3f, 0x43, |
| 0x8f, 0x01, 0x4e, 0xdf, 0x8e, 0xa6, 0x43, 0xcd, 0xac, 0x5b, 0xcb, 0x78, 0xd4, 0x3b, 0x19, 0x8c, |
| 0xea, 0xf6, 0xc9, 0x6f, 0x16, 0x1c, 0x50, 0x1e, 0x6d, 0x9a, 0xf1, 0x64, 0x77, 0xb5, 0xf9, 0x58, |
| 0x3d, 0x6c, 0x3f, 0xf7, 0x72, 0x85, 0xcf, 0x43, 0x12, 0xfb, 0x2e, 0x4f, 0xfd, 0xb6, 0xcf, 0x62, |
| 0xfd, 0xe8, 0xb5, 0x0d, 0x44, 0x92, 0x40, 0xac, 0x7d, 0x41, 0xbf, 0x33, 0xe1, 0xef, 0xf6, 0xb3, |
| 0x1f, 0x34, 0xf1, 0xbc, 0xaf, 0x48, 0xe7, 0xbd, 0x4c, 0xf2, 0xd3, 0xf0, 0xfc, 0x9d, 0x21, 0xfd, |
| 0x69, 0x7f, 0x69, 0xd0, 0x6e, 0x57, 0xc3, 0xdd, 0xae, 0xc6, 0x47, 0xdd, 0x6e, 0x4e, 0xb8, 0xd8, |
| 0xd6, 0xcd, 0xbe, 0xfd, 0x27, 0x00, 0x00, 0xff, 0xff, 0x30, 0x50, 0x8f, 0x30, 0xfc, 0x07, 0x00, |
| 0x00, |
| } |