Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: tensorflow/core/framework/summary.proto
package org.tensorflow.framework;
/**
*
* Serialization format for histogram module in
* core/lib/histogram/histogram.h
*
*
* Protobuf type {@code tensorflow.HistogramProto}
*/
public final class HistogramProto extends
org.nd4j.shade.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:tensorflow.HistogramProto)
HistogramProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use HistogramProto.newBuilder() to construct.
private HistogramProto(org.nd4j.shade.protobuf.GeneratedMessageV3.Builder builder) {
super(builder);
}
private HistogramProto() {
bucketLimit_ = emptyDoubleList();
bucket_ = emptyDoubleList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new HistogramProto();
}
@java.lang.Override
public final org.nd4j.shade.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private HistogramProto(
org.nd4j.shade.protobuf.CodedInputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
this();
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
int mutable_bitField0_ = 0;
org.nd4j.shade.protobuf.UnknownFieldSet.Builder unknownFields =
org.nd4j.shade.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 9: {
min_ = input.readDouble();
break;
}
case 17: {
max_ = input.readDouble();
break;
}
case 25: {
num_ = input.readDouble();
break;
}
case 33: {
sum_ = input.readDouble();
break;
}
case 41: {
sumSquares_ = input.readDouble();
break;
}
case 49: {
if (!((mutable_bitField0_ & 0x00000001) != 0)) {
bucketLimit_ = newDoubleList();
mutable_bitField0_ |= 0x00000001;
}
bucketLimit_.addDouble(input.readDouble());
break;
}
case 50: {
int length = input.readRawVarint32();
int limit = input.pushLimit(length);
if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) {
bucketLimit_ = newDoubleList();
mutable_bitField0_ |= 0x00000001;
}
while (input.getBytesUntilLimit() > 0) {
bucketLimit_.addDouble(input.readDouble());
}
input.popLimit(limit);
break;
}
case 57: {
if (!((mutable_bitField0_ & 0x00000002) != 0)) {
bucket_ = newDoubleList();
mutable_bitField0_ |= 0x00000002;
}
bucket_.addDouble(input.readDouble());
break;
}
case 58: {
int length = input.readRawVarint32();
int limit = input.pushLimit(length);
if (!((mutable_bitField0_ & 0x00000002) != 0) && input.getBytesUntilLimit() > 0) {
bucket_ = newDoubleList();
mutable_bitField0_ |= 0x00000002;
}
while (input.getBytesUntilLimit() > 0) {
bucket_.addDouble(input.readDouble());
}
input.popLimit(limit);
break;
}
default: {
if (!parseUnknownField(
input, unknownFields, extensionRegistry, tag)) {
done = true;
}
break;
}
}
}
} catch (org.nd4j.shade.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new org.nd4j.shade.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) != 0)) {
bucketLimit_.makeImmutable(); // C
}
if (((mutable_bitField0_ & 0x00000002) != 0)) {
bucket_.makeImmutable(); // C
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final org.nd4j.shade.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.tensorflow.framework.SummaryProtos.internal_static_tensorflow_HistogramProto_descriptor;
}
@java.lang.Override
protected org.nd4j.shade.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.tensorflow.framework.SummaryProtos.internal_static_tensorflow_HistogramProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.tensorflow.framework.HistogramProto.class, org.tensorflow.framework.HistogramProto.Builder.class);
}
public static final int MIN_FIELD_NUMBER = 1;
private double min_;
/**
* double min = 1;
*/
public double getMin() {
return min_;
}
public static final int MAX_FIELD_NUMBER = 2;
private double max_;
/**
* double max = 2;
*/
public double getMax() {
return max_;
}
public static final int NUM_FIELD_NUMBER = 3;
private double num_;
/**
* double num = 3;
*/
public double getNum() {
return num_;
}
public static final int SUM_FIELD_NUMBER = 4;
private double sum_;
/**
* double sum = 4;
*/
public double getSum() {
return sum_;
}
public static final int SUM_SQUARES_FIELD_NUMBER = 5;
private double sumSquares_;
/**
* double sum_squares = 5;
*/
public double getSumSquares() {
return sumSquares_;
}
public static final int BUCKET_LIMIT_FIELD_NUMBER = 6;
private org.nd4j.shade.protobuf.Internal.DoubleList bucketLimit_;
/**
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
*
* repeated double bucket_limit = 6 [packed = true];
*/
public int getBucketLimitCount() {
return bucketLimit_.size();
}
/**
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
*
* repeated double bucket_limit = 6 [packed = true];
*/
public int getBucketLimitCount() {
return bucketLimit_.size();
}
/**
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*
* Parallel arrays encoding the bucket boundaries and the bucket values.
* bucket(i) is the count for the bucket i. The range for
* a bucket is:
* i == 0: -DBL_MAX .. bucket_limit(0)
* i != 0: bucket_limit(i-1) .. bucket_limit(i)
*