Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: tensorflow/core/framework/versions.proto
package org.tensorflow.framework;
/**
*
* Version information for a piece of serialized data
* There are different types of versions for each type of data
* (GraphDef, etc.), but they all have the same common shape
* described here.
* Each consumer has "consumer" and "min_producer" versions (specified
* elsewhere). A consumer is allowed to consume this data if
* producer >= min_producer
* consumer >= min_consumer
* consumer not in bad_consumers
*
*
* Protobuf type {@code tensorflow.VersionDef}
*/
public final class VersionDef extends
org.nd4j.shade.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:tensorflow.VersionDef)
VersionDefOrBuilder {
private static final long serialVersionUID = 0L;
// Use VersionDef.newBuilder() to construct.
private VersionDef(org.nd4j.shade.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private VersionDef() {
producer_ = 0;
minConsumer_ = 0;
badConsumers_ = java.util.Collections.emptyList();
}
@java.lang.Override
public final org.nd4j.shade.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private VersionDef(
org.nd4j.shade.protobuf.CodedInputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
this();
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
int mutable_bitField0_ = 0;
org.nd4j.shade.protobuf.UnknownFieldSet.Builder unknownFields =
org.nd4j.shade.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownFieldProto3(
input, unknownFields, extensionRegistry, tag)) {
done = true;
}
break;
}
case 8: {
producer_ = input.readInt32();
break;
}
case 16: {
minConsumer_ = input.readInt32();
break;
}
case 24: {
if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
badConsumers_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000004;
}
badConsumers_.add(input.readInt32());
break;
}
case 26: {
int length = input.readRawVarint32();
int limit = input.pushLimit(length);
if (!((mutable_bitField0_ & 0x00000004) == 0x00000004) && input.getBytesUntilLimit() > 0) {
badConsumers_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000004;
}
while (input.getBytesUntilLimit() > 0) {
badConsumers_.add(input.readInt32());
}
input.popLimit(limit);
break;
}
}
}
} catch (org.nd4j.shade.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new org.nd4j.shade.protobuf.InvalidProtocolBufferException(
e).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
badConsumers_ = java.util.Collections.unmodifiableList(badConsumers_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final org.nd4j.shade.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.tensorflow.framework.VersionsProtos.internal_static_tensorflow_VersionDef_descriptor;
}
protected org.nd4j.shade.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.tensorflow.framework.VersionsProtos.internal_static_tensorflow_VersionDef_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.tensorflow.framework.VersionDef.class, org.tensorflow.framework.VersionDef.Builder.class);
}
private int bitField0_;
public static final int PRODUCER_FIELD_NUMBER = 1;
private int producer_;
/**
*
* The version of the code that produced this data.
*
*
* int32 producer = 1;
*/
public int getProducer() {
return producer_;
}
public static final int MIN_CONSUMER_FIELD_NUMBER = 2;
private int minConsumer_;
/**
*
* Any consumer below this version is not allowed to consume this data.
*
*
* int32 min_consumer = 2;
*/
public int getMinConsumer() {
return minConsumer_;
}
public static final int BAD_CONSUMERS_FIELD_NUMBER = 3;
private java.util.List badConsumers_;
/**
*
* Specific consumer versions which are disallowed (e.g. due to bugs).
*
* Specific consumer versions which are disallowed (e.g. due to bugs).
*
*
* repeated int32 bad_consumers = 3;
*/
public int getBadConsumersCount() {
return badConsumers_.size();
}
/**
*
* Specific consumer versions which are disallowed (e.g. due to bugs).
*
*
* repeated int32 bad_consumers = 3;
*/
public int getBadConsumers(int index) {
return badConsumers_.get(index);
}
private int badConsumersMemoizedSerializedSize = -1;
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(org.nd4j.shade.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (producer_ != 0) {
output.writeInt32(1, producer_);
}
if (minConsumer_ != 0) {
output.writeInt32(2, minConsumer_);
}
if (getBadConsumersList().size() > 0) {
output.writeUInt32NoTag(26);
output.writeUInt32NoTag(badConsumersMemoizedSerializedSize);
}
for (int i = 0; i < badConsumers_.size(); i++) {
output.writeInt32NoTag(badConsumers_.get(i));
}
unknownFields.writeTo(output);
}
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (producer_ != 0) {
size += org.nd4j.shade.protobuf.CodedOutputStream
.computeInt32Size(1, producer_);
}
if (minConsumer_ != 0) {
size += org.nd4j.shade.protobuf.CodedOutputStream
.computeInt32Size(2, minConsumer_);
}
{
int dataSize = 0;
for (int i = 0; i < badConsumers_.size(); i++) {
dataSize += org.nd4j.shade.protobuf.CodedOutputStream
.computeInt32SizeNoTag(badConsumers_.get(i));
}
size += dataSize;
if (!getBadConsumersList().isEmpty()) {
size += 1;
size += org.nd4j.shade.protobuf.CodedOutputStream
.computeInt32SizeNoTag(dataSize);
}
badConsumersMemoizedSerializedSize = dataSize;
}
size += unknownFields.getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.tensorflow.framework.VersionDef)) {
return super.equals(obj);
}
org.tensorflow.framework.VersionDef other = (org.tensorflow.framework.VersionDef) obj;
boolean result = true;
result = result && (getProducer()
== other.getProducer());
result = result && (getMinConsumer()
== other.getMinConsumer());
result = result && getBadConsumersList()
.equals(other.getBadConsumersList());
result = result && unknownFields.equals(other.unknownFields);
return result;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (37 * hash) + PRODUCER_FIELD_NUMBER;
hash = (53 * hash) + getProducer();
hash = (37 * hash) + MIN_CONSUMER_FIELD_NUMBER;
hash = (53 * hash) + getMinConsumer();
if (getBadConsumersCount() > 0) {
hash = (37 * hash) + BAD_CONSUMERS_FIELD_NUMBER;
hash = (53 * hash) + getBadConsumersList().hashCode();
}
hash = (29 * hash) + unknownFields.hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.tensorflow.framework.VersionDef parseFrom(
java.nio.ByteBuffer data)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.tensorflow.framework.VersionDef parseFrom(
java.nio.ByteBuffer data,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.tensorflow.framework.VersionDef parseFrom(
org.nd4j.shade.protobuf.ByteString data)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.tensorflow.framework.VersionDef parseFrom(
org.nd4j.shade.protobuf.ByteString data,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.tensorflow.framework.VersionDef parseFrom(byte[] data)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.tensorflow.framework.VersionDef parseFrom(
byte[] data,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.nd4j.shade.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.tensorflow.framework.VersionDef parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.nd4j.shade.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.tensorflow.framework.VersionDef parseFrom(
java.io.InputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.nd4j.shade.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.tensorflow.framework.VersionDef parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.nd4j.shade.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.tensorflow.framework.VersionDef parseDelimitedFrom(
java.io.InputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.nd4j.shade.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.tensorflow.framework.VersionDef parseFrom(
org.nd4j.shade.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.nd4j.shade.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.tensorflow.framework.VersionDef parseFrom(
org.nd4j.shade.protobuf.CodedInputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.nd4j.shade.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.tensorflow.framework.VersionDef prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.nd4j.shade.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
*
* Version information for a piece of serialized data
* There are different types of versions for each type of data
* (GraphDef, etc.), but they all have the same common shape
* described here.
* Each consumer has "consumer" and "min_producer" versions (specified
* elsewhere). A consumer is allowed to consume this data if
* producer >= min_producer
* consumer >= min_consumer
* consumer not in bad_consumers
*
*
* Protobuf type {@code tensorflow.VersionDef}
*/
public static final class Builder extends
org.nd4j.shade.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:tensorflow.VersionDef)
org.tensorflow.framework.VersionDefOrBuilder {
public static final org.nd4j.shade.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.tensorflow.framework.VersionsProtos.internal_static_tensorflow_VersionDef_descriptor;
}
protected org.nd4j.shade.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.tensorflow.framework.VersionsProtos.internal_static_tensorflow_VersionDef_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.tensorflow.framework.VersionDef.class, org.tensorflow.framework.VersionDef.Builder.class);
}
// Construct using org.tensorflow.framework.VersionDef.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.nd4j.shade.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.nd4j.shade.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
}
}
public Builder clear() {
super.clear();
producer_ = 0;
minConsumer_ = 0;
badConsumers_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000004);
return this;
}
public org.nd4j.shade.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.tensorflow.framework.VersionsProtos.internal_static_tensorflow_VersionDef_descriptor;
}
public org.tensorflow.framework.VersionDef getDefaultInstanceForType() {
return org.tensorflow.framework.VersionDef.getDefaultInstance();
}
public org.tensorflow.framework.VersionDef build() {
org.tensorflow.framework.VersionDef result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.tensorflow.framework.VersionDef buildPartial() {
org.tensorflow.framework.VersionDef result = new org.tensorflow.framework.VersionDef(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
result.producer_ = producer_;
result.minConsumer_ = minConsumer_;
if (((bitField0_ & 0x00000004) == 0x00000004)) {
badConsumers_ = java.util.Collections.unmodifiableList(badConsumers_);
bitField0_ = (bitField0_ & ~0x00000004);
}
result.badConsumers_ = badConsumers_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder clone() {
return (Builder) super.clone();
}
public Builder setField(
org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return (Builder) super.setField(field, value);
}
public Builder clearField(
org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field) {
return (Builder) super.clearField(field);
}
public Builder clearOneof(
org.nd4j.shade.protobuf.Descriptors.OneofDescriptor oneof) {
return (Builder) super.clearOneof(oneof);
}
public Builder setRepeatedField(
org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return (Builder) super.setRepeatedField(field, index, value);
}
public Builder addRepeatedField(
org.nd4j.shade.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return (Builder) super.addRepeatedField(field, value);
}
public Builder mergeFrom(org.nd4j.shade.protobuf.Message other) {
if (other instanceof org.tensorflow.framework.VersionDef) {
return mergeFrom((org.tensorflow.framework.VersionDef)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.tensorflow.framework.VersionDef other) {
if (other == org.tensorflow.framework.VersionDef.getDefaultInstance()) return this;
if (other.getProducer() != 0) {
setProducer(other.getProducer());
}
if (other.getMinConsumer() != 0) {
setMinConsumer(other.getMinConsumer());
}
if (!other.badConsumers_.isEmpty()) {
if (badConsumers_.isEmpty()) {
badConsumers_ = other.badConsumers_;
bitField0_ = (bitField0_ & ~0x00000004);
} else {
ensureBadConsumersIsMutable();
badConsumers_.addAll(other.badConsumers_);
}
onChanged();
}
this.mergeUnknownFields(other.unknownFields);
onChanged();
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
org.nd4j.shade.protobuf.CodedInputStream input,
org.nd4j.shade.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.tensorflow.framework.VersionDef parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (org.nd4j.shade.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.tensorflow.framework.VersionDef) e.getUnfinishedMessage();
throw e.unwrapIOException();
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
private int producer_ ;
/**
*
* The version of the code that produced this data.
*
*
* int32 producer = 1;
*/
public int getProducer() {
return producer_;
}
/**
*
* The version of the code that produced this data.
*