org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos Maven / Gradle / Ivy
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: ProtocolInfo.proto
package org.apache.hadoop.ipc.protobuf;
public final class ProtocolInfoProtos {
private ProtocolInfoProtos() {}
public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
}
public interface GetProtocolVersionsRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string protocol = 1;
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
boolean hasProtocol();
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
java.lang.String getProtocol();
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
com.google.protobuf.ByteString
getProtocolBytes();
}
/**
* Protobuf type {@code hadoop.common.GetProtocolVersionsRequestProto}
*
*
**
* Request to get protocol versions for all supported rpc kinds.
*
*/
public static final class GetProtocolVersionsRequestProto extends
com.google.protobuf.GeneratedMessage
implements GetProtocolVersionsRequestProtoOrBuilder {
// Use GetProtocolVersionsRequestProto.newBuilder() to construct.
private GetProtocolVersionsRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private GetProtocolVersionsRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final GetProtocolVersionsRequestProto defaultInstance;
public static GetProtocolVersionsRequestProto getDefaultInstance() {
return defaultInstance;
}
public GetProtocolVersionsRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetProtocolVersionsRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
protocol_ = input.readBytes();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public GetProtocolVersionsRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new GetProtocolVersionsRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string protocol = 1;
public static final int PROTOCOL_FIELD_NUMBER = 1;
private java.lang.Object protocol_;
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public boolean hasProtocol() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public java.lang.String getProtocol() {
java.lang.Object ref = protocol_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
protocol_ = s;
}
return s;
}
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public com.google.protobuf.ByteString
getProtocolBytes() {
java.lang.Object ref = protocol_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
protocol_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
private void initFields() {
protocol_ = "";
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasProtocol()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getProtocolBytes());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getProtocolBytes());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto other = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto) obj;
boolean result = true;
result = result && (hasProtocol() == other.hasProtocol());
if (hasProtocol()) {
result = result && getProtocol()
.equals(other.getProtocol());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasProtocol()) {
hash = (37 * hash) + PROTOCOL_FIELD_NUMBER;
hash = (53 * hash) + getProtocol().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.common.GetProtocolVersionsRequestProto}
*
*
**
* Request to get protocol versions for all supported rpc kinds.
*
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
protocol_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsRequestProto_descriptor;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.getDefaultInstance();
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto build() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto buildPartial() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto result = new org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.protocol_ = protocol_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto) {
return mergeFrom((org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto other) {
if (other == org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.getDefaultInstance()) return this;
if (other.hasProtocol()) {
bitField0_ |= 0x00000001;
protocol_ = other.protocol_;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasProtocol()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string protocol = 1;
private java.lang.Object protocol_ = "";
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public boolean hasProtocol() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public java.lang.String getProtocol() {
java.lang.Object ref = protocol_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
protocol_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public com.google.protobuf.ByteString
getProtocolBytes() {
java.lang.Object ref = protocol_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
protocol_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public Builder setProtocol(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
protocol_ = value;
onChanged();
return this;
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public Builder clearProtocol() {
bitField0_ = (bitField0_ & ~0x00000001);
protocol_ = getDefaultInstance().getProtocol();
onChanged();
return this;
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public Builder setProtocolBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
protocol_ = value;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.common.GetProtocolVersionsRequestProto)
}
static {
defaultInstance = new GetProtocolVersionsRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.common.GetProtocolVersionsRequestProto)
}
public interface ProtocolVersionProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string rpcKind = 1;
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
boolean hasRpcKind();
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
java.lang.String getRpcKind();
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
com.google.protobuf.ByteString
getRpcKindBytes();
// repeated uint64 versions = 2;
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
java.util.List getVersionsList();
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
int getVersionsCount();
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
long getVersions(int index);
}
/**
* Protobuf type {@code hadoop.common.ProtocolVersionProto}
*
*
**
* Protocol version with corresponding rpc kind.
*
*/
public static final class ProtocolVersionProto extends
com.google.protobuf.GeneratedMessage
implements ProtocolVersionProtoOrBuilder {
// Use ProtocolVersionProto.newBuilder() to construct.
private ProtocolVersionProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private ProtocolVersionProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final ProtocolVersionProto defaultInstance;
public static ProtocolVersionProto getDefaultInstance() {
return defaultInstance;
}
public ProtocolVersionProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ProtocolVersionProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
rpcKind_ = input.readBytes();
break;
}
case 16: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
versions_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
versions_.add(input.readUInt64());
break;
}
case 18: {
int length = input.readRawVarint32();
int limit = input.pushLimit(length);
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) {
versions_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
while (input.getBytesUntilLimit() > 0) {
versions_.add(input.readUInt64());
}
input.popLimit(limit);
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
versions_ = java.util.Collections.unmodifiableList(versions_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolVersionProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolVersionProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public ProtocolVersionProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new ProtocolVersionProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string rpcKind = 1;
public static final int RPCKIND_FIELD_NUMBER = 1;
private java.lang.Object rpcKind_;
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public boolean hasRpcKind() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public java.lang.String getRpcKind() {
java.lang.Object ref = rpcKind_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
rpcKind_ = s;
}
return s;
}
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public com.google.protobuf.ByteString
getRpcKindBytes() {
java.lang.Object ref = rpcKind_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
rpcKind_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// repeated uint64 versions = 2;
public static final int VERSIONS_FIELD_NUMBER = 2;
private java.util.List versions_;
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public java.util.List
getVersionsList() {
return versions_;
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public int getVersionsCount() {
return versions_.size();
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public long getVersions(int index) {
return versions_.get(index);
}
private void initFields() {
rpcKind_ = "";
versions_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasRpcKind()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getRpcKindBytes());
}
for (int i = 0; i < versions_.size(); i++) {
output.writeUInt64(2, versions_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getRpcKindBytes());
}
{
int dataSize = 0;
for (int i = 0; i < versions_.size(); i++) {
dataSize += com.google.protobuf.CodedOutputStream
.computeUInt64SizeNoTag(versions_.get(i));
}
size += dataSize;
size += 1 * getVersionsList().size();
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto)) {
return super.equals(obj);
}
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto other = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto) obj;
boolean result = true;
result = result && (hasRpcKind() == other.hasRpcKind());
if (hasRpcKind()) {
result = result && getRpcKind()
.equals(other.getRpcKind());
}
result = result && getVersionsList()
.equals(other.getVersionsList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasRpcKind()) {
hash = (37 * hash) + RPCKIND_FIELD_NUMBER;
hash = (53 * hash) + getRpcKind().hashCode();
}
if (getVersionsCount() > 0) {
hash = (37 * hash) + VERSIONS_FIELD_NUMBER;
hash = (53 * hash) + getVersionsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.common.ProtocolVersionProto}
*
*
**
* Protocol version with corresponding rpc kind.
*
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolVersionProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolVersionProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder.class);
}
// Construct using org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
rpcKind_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
versions_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolVersionProto_descriptor;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto getDefaultInstanceForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.getDefaultInstance();
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto build() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto buildPartial() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto result = new org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.rpcKind_ = rpcKind_;
if (((bitField0_ & 0x00000002) == 0x00000002)) {
versions_ = java.util.Collections.unmodifiableList(versions_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.versions_ = versions_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto) {
return mergeFrom((org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto other) {
if (other == org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.getDefaultInstance()) return this;
if (other.hasRpcKind()) {
bitField0_ |= 0x00000001;
rpcKind_ = other.rpcKind_;
onChanged();
}
if (!other.versions_.isEmpty()) {
if (versions_.isEmpty()) {
versions_ = other.versions_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureVersionsIsMutable();
versions_.addAll(other.versions_);
}
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasRpcKind()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string rpcKind = 1;
private java.lang.Object rpcKind_ = "";
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public boolean hasRpcKind() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public java.lang.String getRpcKind() {
java.lang.Object ref = rpcKind_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
rpcKind_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public com.google.protobuf.ByteString
getRpcKindBytes() {
java.lang.Object ref = rpcKind_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
rpcKind_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public Builder setRpcKind(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
rpcKind_ = value;
onChanged();
return this;
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public Builder clearRpcKind() {
bitField0_ = (bitField0_ & ~0x00000001);
rpcKind_ = getDefaultInstance().getRpcKind();
onChanged();
return this;
}
/**
* required string rpcKind = 1;
*
*
*RPC kind
*
*/
public Builder setRpcKindBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
rpcKind_ = value;
onChanged();
return this;
}
// repeated uint64 versions = 2;
private java.util.List versions_ = java.util.Collections.emptyList();
private void ensureVersionsIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
versions_ = new java.util.ArrayList(versions_);
bitField0_ |= 0x00000002;
}
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public java.util.List
getVersionsList() {
return java.util.Collections.unmodifiableList(versions_);
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public int getVersionsCount() {
return versions_.size();
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public long getVersions(int index) {
return versions_.get(index);
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public Builder setVersions(
int index, long value) {
ensureVersionsIsMutable();
versions_.set(index, value);
onChanged();
return this;
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public Builder addVersions(long value) {
ensureVersionsIsMutable();
versions_.add(value);
onChanged();
return this;
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public Builder addAllVersions(
java.lang.Iterable extends java.lang.Long> values) {
ensureVersionsIsMutable();
super.addAll(values, versions_);
onChanged();
return this;
}
/**
* repeated uint64 versions = 2;
*
*
*Protocol version corresponding to the rpc kind.
*
*/
public Builder clearVersions() {
versions_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.common.ProtocolVersionProto)
}
static {
defaultInstance = new ProtocolVersionProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.common.ProtocolVersionProto)
}
public interface GetProtocolVersionsResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
java.util.List
getProtocolVersionsList();
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto getProtocolVersions(int index);
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
int getProtocolVersionsCount();
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
java.util.List extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder>
getProtocolVersionsOrBuilderList();
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder getProtocolVersionsOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.common.GetProtocolVersionsResponseProto}
*
*
**
* Get protocol version response.
*
*/
public static final class GetProtocolVersionsResponseProto extends
com.google.protobuf.GeneratedMessage
implements GetProtocolVersionsResponseProtoOrBuilder {
// Use GetProtocolVersionsResponseProto.newBuilder() to construct.
private GetProtocolVersionsResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private GetProtocolVersionsResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final GetProtocolVersionsResponseProto defaultInstance;
public static GetProtocolVersionsResponseProto getDefaultInstance() {
return defaultInstance;
}
public GetProtocolVersionsResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetProtocolVersionsResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
protocolVersions_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000001;
}
protocolVersions_.add(input.readMessage(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.PARSER, extensionRegistry));
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
protocolVersions_ = java.util.Collections.unmodifiableList(protocolVersions_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public GetProtocolVersionsResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new GetProtocolVersionsResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
// repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
public static final int PROTOCOLVERSIONS_FIELD_NUMBER = 1;
private java.util.List protocolVersions_;
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public java.util.List getProtocolVersionsList() {
return protocolVersions_;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public java.util.List extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder>
getProtocolVersionsOrBuilderList() {
return protocolVersions_;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public int getProtocolVersionsCount() {
return protocolVersions_.size();
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto getProtocolVersions(int index) {
return protocolVersions_.get(index);
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder getProtocolVersionsOrBuilder(
int index) {
return protocolVersions_.get(index);
}
private void initFields() {
protocolVersions_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
for (int i = 0; i < getProtocolVersionsCount(); i++) {
if (!getProtocolVersions(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
for (int i = 0; i < protocolVersions_.size(); i++) {
output.writeMessage(1, protocolVersions_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < protocolVersions_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(1, protocolVersions_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto other = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto) obj;
boolean result = true;
result = result && getProtocolVersionsList()
.equals(other.getProtocolVersionsList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getProtocolVersionsCount() > 0) {
hash = (37 * hash) + PROTOCOLVERSIONS_FIELD_NUMBER;
hash = (53 * hash) + getProtocolVersionsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.common.GetProtocolVersionsResponseProto}
*
*
**
* Get protocol version response.
*
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getProtocolVersionsFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
if (protocolVersionsBuilder_ == null) {
protocolVersions_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
} else {
protocolVersionsBuilder_.clear();
}
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolVersionsResponseProto_descriptor;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance();
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto build() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto buildPartial() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto result = new org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto(this);
int from_bitField0_ = bitField0_;
if (protocolVersionsBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
protocolVersions_ = java.util.Collections.unmodifiableList(protocolVersions_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.protocolVersions_ = protocolVersions_;
} else {
result.protocolVersions_ = protocolVersionsBuilder_.build();
}
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto) {
return mergeFrom((org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto other) {
if (other == org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance()) return this;
if (protocolVersionsBuilder_ == null) {
if (!other.protocolVersions_.isEmpty()) {
if (protocolVersions_.isEmpty()) {
protocolVersions_ = other.protocolVersions_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureProtocolVersionsIsMutable();
protocolVersions_.addAll(other.protocolVersions_);
}
onChanged();
}
} else {
if (!other.protocolVersions_.isEmpty()) {
if (protocolVersionsBuilder_.isEmpty()) {
protocolVersionsBuilder_.dispose();
protocolVersionsBuilder_ = null;
protocolVersions_ = other.protocolVersions_;
bitField0_ = (bitField0_ & ~0x00000001);
protocolVersionsBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getProtocolVersionsFieldBuilder() : null;
} else {
protocolVersionsBuilder_.addAllMessages(other.protocolVersions_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
for (int i = 0; i < getProtocolVersionsCount(); i++) {
if (!getProtocolVersions(i).isInitialized()) {
return false;
}
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
private java.util.List protocolVersions_ =
java.util.Collections.emptyList();
private void ensureProtocolVersionsIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
protocolVersions_ = new java.util.ArrayList(protocolVersions_);
bitField0_ |= 0x00000001;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder> protocolVersionsBuilder_;
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public java.util.List getProtocolVersionsList() {
if (protocolVersionsBuilder_ == null) {
return java.util.Collections.unmodifiableList(protocolVersions_);
} else {
return protocolVersionsBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public int getProtocolVersionsCount() {
if (protocolVersionsBuilder_ == null) {
return protocolVersions_.size();
} else {
return protocolVersionsBuilder_.getCount();
}
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto getProtocolVersions(int index) {
if (protocolVersionsBuilder_ == null) {
return protocolVersions_.get(index);
} else {
return protocolVersionsBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder setProtocolVersions(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto value) {
if (protocolVersionsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureProtocolVersionsIsMutable();
protocolVersions_.set(index, value);
onChanged();
} else {
protocolVersionsBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder setProtocolVersions(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder builderForValue) {
if (protocolVersionsBuilder_ == null) {
ensureProtocolVersionsIsMutable();
protocolVersions_.set(index, builderForValue.build());
onChanged();
} else {
protocolVersionsBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder addProtocolVersions(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto value) {
if (protocolVersionsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureProtocolVersionsIsMutable();
protocolVersions_.add(value);
onChanged();
} else {
protocolVersionsBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder addProtocolVersions(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto value) {
if (protocolVersionsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureProtocolVersionsIsMutable();
protocolVersions_.add(index, value);
onChanged();
} else {
protocolVersionsBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder addProtocolVersions(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder builderForValue) {
if (protocolVersionsBuilder_ == null) {
ensureProtocolVersionsIsMutable();
protocolVersions_.add(builderForValue.build());
onChanged();
} else {
protocolVersionsBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder addProtocolVersions(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder builderForValue) {
if (protocolVersionsBuilder_ == null) {
ensureProtocolVersionsIsMutable();
protocolVersions_.add(index, builderForValue.build());
onChanged();
} else {
protocolVersionsBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder addAllProtocolVersions(
java.lang.Iterable extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto> values) {
if (protocolVersionsBuilder_ == null) {
ensureProtocolVersionsIsMutable();
super.addAll(values, protocolVersions_);
onChanged();
} else {
protocolVersionsBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder clearProtocolVersions() {
if (protocolVersionsBuilder_ == null) {
protocolVersions_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
protocolVersionsBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public Builder removeProtocolVersions(int index) {
if (protocolVersionsBuilder_ == null) {
ensureProtocolVersionsIsMutable();
protocolVersions_.remove(index);
onChanged();
} else {
protocolVersionsBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder getProtocolVersionsBuilder(
int index) {
return getProtocolVersionsFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder getProtocolVersionsOrBuilder(
int index) {
if (protocolVersionsBuilder_ == null) {
return protocolVersions_.get(index); } else {
return protocolVersionsBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public java.util.List extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder>
getProtocolVersionsOrBuilderList() {
if (protocolVersionsBuilder_ != null) {
return protocolVersionsBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(protocolVersions_);
}
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder addProtocolVersionsBuilder() {
return getProtocolVersionsFieldBuilder().addBuilder(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.getDefaultInstance());
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder addProtocolVersionsBuilder(
int index) {
return getProtocolVersionsFieldBuilder().addBuilder(
index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.getDefaultInstance());
}
/**
* repeated .hadoop.common.ProtocolVersionProto protocolVersions = 1;
*/
public java.util.List
getProtocolVersionsBuilderList() {
return getProtocolVersionsFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder>
getProtocolVersionsFieldBuilder() {
if (protocolVersionsBuilder_ == null) {
protocolVersionsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProto.Builder, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolVersionProtoOrBuilder>(
protocolVersions_,
((bitField0_ & 0x00000001) == 0x00000001),
getParentForChildren(),
isClean());
protocolVersions_ = null;
}
return protocolVersionsBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.common.GetProtocolVersionsResponseProto)
}
static {
defaultInstance = new GetProtocolVersionsResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.common.GetProtocolVersionsResponseProto)
}
public interface GetProtocolSignatureRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string protocol = 1;
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
boolean hasProtocol();
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
java.lang.String getProtocol();
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
com.google.protobuf.ByteString
getProtocolBytes();
// required string rpcKind = 2;
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
boolean hasRpcKind();
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
java.lang.String getRpcKind();
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
com.google.protobuf.ByteString
getRpcKindBytes();
}
/**
* Protobuf type {@code hadoop.common.GetProtocolSignatureRequestProto}
*
*
**
* Get protocol signature request.
*
*/
public static final class GetProtocolSignatureRequestProto extends
com.google.protobuf.GeneratedMessage
implements GetProtocolSignatureRequestProtoOrBuilder {
// Use GetProtocolSignatureRequestProto.newBuilder() to construct.
private GetProtocolSignatureRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private GetProtocolSignatureRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final GetProtocolSignatureRequestProto defaultInstance;
public static GetProtocolSignatureRequestProto getDefaultInstance() {
return defaultInstance;
}
public GetProtocolSignatureRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetProtocolSignatureRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
protocol_ = input.readBytes();
break;
}
case 18: {
bitField0_ |= 0x00000002;
rpcKind_ = input.readBytes();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public GetProtocolSignatureRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new GetProtocolSignatureRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string protocol = 1;
public static final int PROTOCOL_FIELD_NUMBER = 1;
private java.lang.Object protocol_;
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public boolean hasProtocol() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public java.lang.String getProtocol() {
java.lang.Object ref = protocol_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
protocol_ = s;
}
return s;
}
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public com.google.protobuf.ByteString
getProtocolBytes() {
java.lang.Object ref = protocol_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
protocol_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// required string rpcKind = 2;
public static final int RPCKIND_FIELD_NUMBER = 2;
private java.lang.Object rpcKind_;
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public boolean hasRpcKind() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public java.lang.String getRpcKind() {
java.lang.Object ref = rpcKind_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
rpcKind_ = s;
}
return s;
}
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public com.google.protobuf.ByteString
getRpcKindBytes() {
java.lang.Object ref = rpcKind_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
rpcKind_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
private void initFields() {
protocol_ = "";
rpcKind_ = "";
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasProtocol()) {
memoizedIsInitialized = 0;
return false;
}
if (!hasRpcKind()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getProtocolBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, getRpcKindBytes());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getProtocolBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(2, getRpcKindBytes());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto other = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto) obj;
boolean result = true;
result = result && (hasProtocol() == other.hasProtocol());
if (hasProtocol()) {
result = result && getProtocol()
.equals(other.getProtocol());
}
result = result && (hasRpcKind() == other.hasRpcKind());
if (hasRpcKind()) {
result = result && getRpcKind()
.equals(other.getRpcKind());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasProtocol()) {
hash = (37 * hash) + PROTOCOL_FIELD_NUMBER;
hash = (53 * hash) + getProtocol().hashCode();
}
if (hasRpcKind()) {
hash = (37 * hash) + RPCKIND_FIELD_NUMBER;
hash = (53 * hash) + getRpcKind().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.common.GetProtocolSignatureRequestProto}
*
*
**
* Get protocol signature request.
*
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
protocol_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
rpcKind_ = "";
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureRequestProto_descriptor;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.getDefaultInstance();
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto build() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto buildPartial() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto result = new org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.protocol_ = protocol_;
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
to_bitField0_ |= 0x00000002;
}
result.rpcKind_ = rpcKind_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto) {
return mergeFrom((org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto other) {
if (other == org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.getDefaultInstance()) return this;
if (other.hasProtocol()) {
bitField0_ |= 0x00000001;
protocol_ = other.protocol_;
onChanged();
}
if (other.hasRpcKind()) {
bitField0_ |= 0x00000002;
rpcKind_ = other.rpcKind_;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasProtocol()) {
return false;
}
if (!hasRpcKind()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string protocol = 1;
private java.lang.Object protocol_ = "";
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public boolean hasProtocol() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public java.lang.String getProtocol() {
java.lang.Object ref = protocol_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
protocol_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public com.google.protobuf.ByteString
getProtocolBytes() {
java.lang.Object ref = protocol_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
protocol_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public Builder setProtocol(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
protocol_ = value;
onChanged();
return this;
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public Builder clearProtocol() {
bitField0_ = (bitField0_ & ~0x00000001);
protocol_ = getDefaultInstance().getProtocol();
onChanged();
return this;
}
/**
* required string protocol = 1;
*
*
* Protocol name
*
*/
public Builder setProtocolBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
protocol_ = value;
onChanged();
return this;
}
// required string rpcKind = 2;
private java.lang.Object rpcKind_ = "";
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public boolean hasRpcKind() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public java.lang.String getRpcKind() {
java.lang.Object ref = rpcKind_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
rpcKind_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public com.google.protobuf.ByteString
getRpcKindBytes() {
java.lang.Object ref = rpcKind_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
rpcKind_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public Builder setRpcKind(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
rpcKind_ = value;
onChanged();
return this;
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public Builder clearRpcKind() {
bitField0_ = (bitField0_ & ~0x00000002);
rpcKind_ = getDefaultInstance().getRpcKind();
onChanged();
return this;
}
/**
* required string rpcKind = 2;
*
*
* RPC kind
*
*/
public Builder setRpcKindBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
rpcKind_ = value;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.common.GetProtocolSignatureRequestProto)
}
static {
defaultInstance = new GetProtocolSignatureRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.common.GetProtocolSignatureRequestProto)
}
public interface GetProtocolSignatureResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
java.util.List
getProtocolSignatureList();
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto getProtocolSignature(int index);
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
int getProtocolSignatureCount();
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
java.util.List extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder>
getProtocolSignatureOrBuilderList();
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder getProtocolSignatureOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.common.GetProtocolSignatureResponseProto}
*
*
**
* Get protocol signature response.
*
*/
public static final class GetProtocolSignatureResponseProto extends
com.google.protobuf.GeneratedMessage
implements GetProtocolSignatureResponseProtoOrBuilder {
// Use GetProtocolSignatureResponseProto.newBuilder() to construct.
private GetProtocolSignatureResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private GetProtocolSignatureResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final GetProtocolSignatureResponseProto defaultInstance;
public static GetProtocolSignatureResponseProto getDefaultInstance() {
return defaultInstance;
}
public GetProtocolSignatureResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetProtocolSignatureResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
protocolSignature_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000001;
}
protocolSignature_.add(input.readMessage(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.PARSER, extensionRegistry));
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
protocolSignature_ = java.util.Collections.unmodifiableList(protocolSignature_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public GetProtocolSignatureResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new GetProtocolSignatureResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
// repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
public static final int PROTOCOLSIGNATURE_FIELD_NUMBER = 1;
private java.util.List protocolSignature_;
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public java.util.List getProtocolSignatureList() {
return protocolSignature_;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public java.util.List extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder>
getProtocolSignatureOrBuilderList() {
return protocolSignature_;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public int getProtocolSignatureCount() {
return protocolSignature_.size();
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto getProtocolSignature(int index) {
return protocolSignature_.get(index);
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder getProtocolSignatureOrBuilder(
int index) {
return protocolSignature_.get(index);
}
private void initFields() {
protocolSignature_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
for (int i = 0; i < getProtocolSignatureCount(); i++) {
if (!getProtocolSignature(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
for (int i = 0; i < protocolSignature_.size(); i++) {
output.writeMessage(1, protocolSignature_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < protocolSignature_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(1, protocolSignature_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto other = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto) obj;
boolean result = true;
result = result && getProtocolSignatureList()
.equals(other.getProtocolSignatureList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (getProtocolSignatureCount() > 0) {
hash = (37 * hash) + PROTOCOLSIGNATURE_FIELD_NUMBER;
hash = (53 * hash) + getProtocolSignatureList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.common.GetProtocolSignatureResponseProto}
*
*
**
* Get protocol signature response.
*
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getProtocolSignatureFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
if (protocolSignatureBuilder_ == null) {
protocolSignature_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
} else {
protocolSignatureBuilder_.clear();
}
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_GetProtocolSignatureResponseProto_descriptor;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance();
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto build() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto buildPartial() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto result = new org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto(this);
int from_bitField0_ = bitField0_;
if (protocolSignatureBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001)) {
protocolSignature_ = java.util.Collections.unmodifiableList(protocolSignature_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.protocolSignature_ = protocolSignature_;
} else {
result.protocolSignature_ = protocolSignatureBuilder_.build();
}
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto) {
return mergeFrom((org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto other) {
if (other == org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance()) return this;
if (protocolSignatureBuilder_ == null) {
if (!other.protocolSignature_.isEmpty()) {
if (protocolSignature_.isEmpty()) {
protocolSignature_ = other.protocolSignature_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureProtocolSignatureIsMutable();
protocolSignature_.addAll(other.protocolSignature_);
}
onChanged();
}
} else {
if (!other.protocolSignature_.isEmpty()) {
if (protocolSignatureBuilder_.isEmpty()) {
protocolSignatureBuilder_.dispose();
protocolSignatureBuilder_ = null;
protocolSignature_ = other.protocolSignature_;
bitField0_ = (bitField0_ & ~0x00000001);
protocolSignatureBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getProtocolSignatureFieldBuilder() : null;
} else {
protocolSignatureBuilder_.addAllMessages(other.protocolSignature_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
for (int i = 0; i < getProtocolSignatureCount(); i++) {
if (!getProtocolSignature(i).isInitialized()) {
return false;
}
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
private java.util.List protocolSignature_ =
java.util.Collections.emptyList();
private void ensureProtocolSignatureIsMutable() {
if (!((bitField0_ & 0x00000001) == 0x00000001)) {
protocolSignature_ = new java.util.ArrayList(protocolSignature_);
bitField0_ |= 0x00000001;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder> protocolSignatureBuilder_;
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public java.util.List getProtocolSignatureList() {
if (protocolSignatureBuilder_ == null) {
return java.util.Collections.unmodifiableList(protocolSignature_);
} else {
return protocolSignatureBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public int getProtocolSignatureCount() {
if (protocolSignatureBuilder_ == null) {
return protocolSignature_.size();
} else {
return protocolSignatureBuilder_.getCount();
}
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto getProtocolSignature(int index) {
if (protocolSignatureBuilder_ == null) {
return protocolSignature_.get(index);
} else {
return protocolSignatureBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder setProtocolSignature(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto value) {
if (protocolSignatureBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureProtocolSignatureIsMutable();
protocolSignature_.set(index, value);
onChanged();
} else {
protocolSignatureBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder setProtocolSignature(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder builderForValue) {
if (protocolSignatureBuilder_ == null) {
ensureProtocolSignatureIsMutable();
protocolSignature_.set(index, builderForValue.build());
onChanged();
} else {
protocolSignatureBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder addProtocolSignature(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto value) {
if (protocolSignatureBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureProtocolSignatureIsMutable();
protocolSignature_.add(value);
onChanged();
} else {
protocolSignatureBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder addProtocolSignature(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto value) {
if (protocolSignatureBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureProtocolSignatureIsMutable();
protocolSignature_.add(index, value);
onChanged();
} else {
protocolSignatureBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder addProtocolSignature(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder builderForValue) {
if (protocolSignatureBuilder_ == null) {
ensureProtocolSignatureIsMutable();
protocolSignature_.add(builderForValue.build());
onChanged();
} else {
protocolSignatureBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder addProtocolSignature(
int index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder builderForValue) {
if (protocolSignatureBuilder_ == null) {
ensureProtocolSignatureIsMutable();
protocolSignature_.add(index, builderForValue.build());
onChanged();
} else {
protocolSignatureBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder addAllProtocolSignature(
java.lang.Iterable extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto> values) {
if (protocolSignatureBuilder_ == null) {
ensureProtocolSignatureIsMutable();
super.addAll(values, protocolSignature_);
onChanged();
} else {
protocolSignatureBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder clearProtocolSignature() {
if (protocolSignatureBuilder_ == null) {
protocolSignature_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
protocolSignatureBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public Builder removeProtocolSignature(int index) {
if (protocolSignatureBuilder_ == null) {
ensureProtocolSignatureIsMutable();
protocolSignature_.remove(index);
onChanged();
} else {
protocolSignatureBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder getProtocolSignatureBuilder(
int index) {
return getProtocolSignatureFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder getProtocolSignatureOrBuilder(
int index) {
if (protocolSignatureBuilder_ == null) {
return protocolSignature_.get(index); } else {
return protocolSignatureBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public java.util.List extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder>
getProtocolSignatureOrBuilderList() {
if (protocolSignatureBuilder_ != null) {
return protocolSignatureBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(protocolSignature_);
}
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder addProtocolSignatureBuilder() {
return getProtocolSignatureFieldBuilder().addBuilder(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.getDefaultInstance());
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder addProtocolSignatureBuilder(
int index) {
return getProtocolSignatureFieldBuilder().addBuilder(
index, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.getDefaultInstance());
}
/**
* repeated .hadoop.common.ProtocolSignatureProto protocolSignature = 1;
*/
public java.util.List
getProtocolSignatureBuilderList() {
return getProtocolSignatureFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder>
getProtocolSignatureFieldBuilder() {
if (protocolSignatureBuilder_ == null) {
protocolSignatureBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder>(
protocolSignature_,
((bitField0_ & 0x00000001) == 0x00000001),
getParentForChildren(),
isClean());
protocolSignature_ = null;
}
return protocolSignatureBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.common.GetProtocolSignatureResponseProto)
}
static {
defaultInstance = new GetProtocolSignatureResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.common.GetProtocolSignatureResponseProto)
}
public interface ProtocolSignatureProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint64 version = 1;
/**
* required uint64 version = 1;
*/
boolean hasVersion();
/**
* required uint64 version = 1;
*/
long getVersion();
// repeated uint32 methods = 2;
/**
* repeated uint32 methods = 2;
*/
java.util.List getMethodsList();
/**
* repeated uint32 methods = 2;
*/
int getMethodsCount();
/**
* repeated uint32 methods = 2;
*/
int getMethods(int index);
}
/**
* Protobuf type {@code hadoop.common.ProtocolSignatureProto}
*/
public static final class ProtocolSignatureProto extends
com.google.protobuf.GeneratedMessage
implements ProtocolSignatureProtoOrBuilder {
// Use ProtocolSignatureProto.newBuilder() to construct.
private ProtocolSignatureProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private ProtocolSignatureProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final ProtocolSignatureProto defaultInstance;
public static ProtocolSignatureProto getDefaultInstance() {
return defaultInstance;
}
public ProtocolSignatureProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ProtocolSignatureProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 8: {
bitField0_ |= 0x00000001;
version_ = input.readUInt64();
break;
}
case 16: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
methods_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
methods_.add(input.readUInt32());
break;
}
case 18: {
int length = input.readRawVarint32();
int limit = input.pushLimit(length);
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002) && input.getBytesUntilLimit() > 0) {
methods_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
while (input.getBytesUntilLimit() > 0) {
methods_.add(input.readUInt32());
}
input.popLimit(limit);
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
methods_ = java.util.Collections.unmodifiableList(methods_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolSignatureProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolSignatureProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public ProtocolSignatureProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new ProtocolSignatureProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required uint64 version = 1;
public static final int VERSION_FIELD_NUMBER = 1;
private long version_;
/**
* required uint64 version = 1;
*/
public boolean hasVersion() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required uint64 version = 1;
*/
public long getVersion() {
return version_;
}
// repeated uint32 methods = 2;
public static final int METHODS_FIELD_NUMBER = 2;
private java.util.List methods_;
/**
* repeated uint32 methods = 2;
*/
public java.util.List
getMethodsList() {
return methods_;
}
/**
* repeated uint32 methods = 2;
*/
public int getMethodsCount() {
return methods_.size();
}
/**
* repeated uint32 methods = 2;
*/
public int getMethods(int index) {
return methods_.get(index);
}
private void initFields() {
version_ = 0L;
methods_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasVersion()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt64(1, version_);
}
for (int i = 0; i < methods_.size(); i++) {
output.writeUInt32(2, methods_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeUInt64Size(1, version_);
}
{
int dataSize = 0;
for (int i = 0; i < methods_.size(); i++) {
dataSize += com.google.protobuf.CodedOutputStream
.computeUInt32SizeNoTag(methods_.get(i));
}
size += dataSize;
size += 1 * getMethodsList().size();
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto)) {
return super.equals(obj);
}
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto other = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto) obj;
boolean result = true;
result = result && (hasVersion() == other.hasVersion());
if (hasVersion()) {
result = result && (getVersion()
== other.getVersion());
}
result = result && getMethodsList()
.equals(other.getMethodsList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasVersion()) {
hash = (37 * hash) + VERSION_FIELD_NUMBER;
hash = (53 * hash) + hashLong(getVersion());
}
if (getMethodsCount() > 0) {
hash = (37 * hash) + METHODS_FIELD_NUMBER;
hash = (53 * hash) + getMethodsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.common.ProtocolSignatureProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolSignatureProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolSignatureProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.class, org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.Builder.class);
}
// Construct using org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
version_ = 0L;
bitField0_ = (bitField0_ & ~0x00000001);
methods_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.internal_static_hadoop_common_ProtocolSignatureProto_descriptor;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto getDefaultInstanceForType() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.getDefaultInstance();
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto build() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto buildPartial() {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto result = new org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.version_ = version_;
if (((bitField0_ & 0x00000002) == 0x00000002)) {
methods_ = java.util.Collections.unmodifiableList(methods_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.methods_ = methods_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto) {
return mergeFrom((org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto other) {
if (other == org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto.getDefaultInstance()) return this;
if (other.hasVersion()) {
setVersion(other.getVersion());
}
if (!other.methods_.isEmpty()) {
if (methods_.isEmpty()) {
methods_ = other.methods_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureMethodsIsMutable();
methods_.addAll(other.methods_);
}
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasVersion()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolSignatureProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required uint64 version = 1;
private long version_ ;
/**
* required uint64 version = 1;
*/
public boolean hasVersion() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required uint64 version = 1;
*/
public long getVersion() {
return version_;
}
/**
* required uint64 version = 1;
*/
public Builder setVersion(long value) {
bitField0_ |= 0x00000001;
version_ = value;
onChanged();
return this;
}
/**
* required uint64 version = 1;
*/
public Builder clearVersion() {
bitField0_ = (bitField0_ & ~0x00000001);
version_ = 0L;
onChanged();
return this;
}
// repeated uint32 methods = 2;
private java.util.List methods_ = java.util.Collections.emptyList();
private void ensureMethodsIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
methods_ = new java.util.ArrayList(methods_);
bitField0_ |= 0x00000002;
}
}
/**
* repeated uint32 methods = 2;
*/
public java.util.List
getMethodsList() {
return java.util.Collections.unmodifiableList(methods_);
}
/**
* repeated uint32 methods = 2;
*/
public int getMethodsCount() {
return methods_.size();
}
/**
* repeated uint32 methods = 2;
*/
public int getMethods(int index) {
return methods_.get(index);
}
/**
* repeated uint32 methods = 2;
*/
public Builder setMethods(
int index, int value) {
ensureMethodsIsMutable();
methods_.set(index, value);
onChanged();
return this;
}
/**
* repeated uint32 methods = 2;
*/
public Builder addMethods(int value) {
ensureMethodsIsMutable();
methods_.add(value);
onChanged();
return this;
}
/**
* repeated uint32 methods = 2;
*/
public Builder addAllMethods(
java.lang.Iterable extends java.lang.Integer> values) {
ensureMethodsIsMutable();
super.addAll(values, methods_);
onChanged();
return this;
}
/**
* repeated uint32 methods = 2;
*/
public Builder clearMethods() {
methods_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.common.ProtocolSignatureProto)
}
static {
defaultInstance = new ProtocolSignatureProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.common.ProtocolSignatureProto)
}
/**
* Protobuf service {@code hadoop.common.ProtocolInfoService}
*
*
**
* Protocol to get information about protocols.
*
*/
public static abstract class ProtocolInfoService
implements com.google.protobuf.Service {
protected ProtocolInfoService() {}
public interface Interface {
/**
* rpc getProtocolVersions(.hadoop.common.GetProtocolVersionsRequestProto) returns (.hadoop.common.GetProtocolVersionsResponseProto);
*
*
**
* Return protocol version corresponding to protocol interface for each
* supported rpc kind.
*
*/
public abstract void getProtocolVersions(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto request,
com.google.protobuf.RpcCallback done);
/**
* rpc getProtocolSignature(.hadoop.common.GetProtocolSignatureRequestProto) returns (.hadoop.common.GetProtocolSignatureResponseProto);
*
*
**
* Return protocol version corresponding to protocol interface.
*
*/
public abstract void getProtocolSignature(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto request,
com.google.protobuf.RpcCallback done);
}
public static com.google.protobuf.Service newReflectiveService(
final Interface impl) {
return new ProtocolInfoService() {
@java.lang.Override
public void getProtocolVersions(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto request,
com.google.protobuf.RpcCallback done) {
impl.getProtocolVersions(controller, request, done);
}
@java.lang.Override
public void getProtocolSignature(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto request,
com.google.protobuf.RpcCallback done) {
impl.getProtocolSignature(controller, request, done);
}
};
}
public static com.google.protobuf.BlockingService
newReflectiveBlockingService(final BlockingInterface impl) {
return new com.google.protobuf.BlockingService() {
public final com.google.protobuf.Descriptors.ServiceDescriptor
getDescriptorForType() {
return getDescriptor();
}
public final com.google.protobuf.Message callBlockingMethod(
com.google.protobuf.Descriptors.MethodDescriptor method,
com.google.protobuf.RpcController controller,
com.google.protobuf.Message request)
throws com.google.protobuf.ServiceException {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.callBlockingMethod() given method descriptor for " +
"wrong service type.");
}
switch(method.getIndex()) {
case 0:
return impl.getProtocolVersions(controller, (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto)request);
case 1:
return impl.getProtocolSignature(controller, (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto)request);
default:
throw new java.lang.AssertionError("Can't get here.");
}
}
public final com.google.protobuf.Message
getRequestPrototype(
com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getRequestPrototype() given method " +
"descriptor for wrong service type.");
}
switch(method.getIndex()) {
case 0:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.getDefaultInstance();
case 1:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
}
}
public final com.google.protobuf.Message
getResponsePrototype(
com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getResponsePrototype() given method " +
"descriptor for wrong service type.");
}
switch(method.getIndex()) {
case 0:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance();
case 1:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
}
}
};
}
/**
* rpc getProtocolVersions(.hadoop.common.GetProtocolVersionsRequestProto) returns (.hadoop.common.GetProtocolVersionsResponseProto);
*
*
**
* Return protocol version corresponding to protocol interface for each
* supported rpc kind.
*
*/
public abstract void getProtocolVersions(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto request,
com.google.protobuf.RpcCallback done);
/**
* rpc getProtocolSignature(.hadoop.common.GetProtocolSignatureRequestProto) returns (.hadoop.common.GetProtocolSignatureResponseProto);
*
*
**
* Return protocol version corresponding to protocol interface.
*
*/
public abstract void getProtocolSignature(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto request,
com.google.protobuf.RpcCallback done);
public static final
com.google.protobuf.Descriptors.ServiceDescriptor
getDescriptor() {
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.getDescriptor().getServices().get(0);
}
public final com.google.protobuf.Descriptors.ServiceDescriptor
getDescriptorForType() {
return getDescriptor();
}
public final void callMethod(
com.google.protobuf.Descriptors.MethodDescriptor method,
com.google.protobuf.RpcController controller,
com.google.protobuf.Message request,
com.google.protobuf.RpcCallback<
com.google.protobuf.Message> done) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.callMethod() given method descriptor for wrong " +
"service type.");
}
switch(method.getIndex()) {
case 0:
this.getProtocolVersions(controller, (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto)request,
com.google.protobuf.RpcUtil.specializeCallback(
done));
return;
case 1:
this.getProtocolSignature(controller, (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto)request,
com.google.protobuf.RpcUtil.specializeCallback(
done));
return;
default:
throw new java.lang.AssertionError("Can't get here.");
}
}
public final com.google.protobuf.Message
getRequestPrototype(
com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getRequestPrototype() given method " +
"descriptor for wrong service type.");
}
switch(method.getIndex()) {
case 0:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto.getDefaultInstance();
case 1:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
}
}
public final com.google.protobuf.Message
getResponsePrototype(
com.google.protobuf.Descriptors.MethodDescriptor method) {
if (method.getService() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"Service.getResponsePrototype() given method " +
"descriptor for wrong service type.");
}
switch(method.getIndex()) {
case 0:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance();
case 1:
return org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance();
default:
throw new java.lang.AssertionError("Can't get here.");
}
}
public static Stub newStub(
com.google.protobuf.RpcChannel channel) {
return new Stub(channel);
}
public static final class Stub extends org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.ProtocolInfoService implements Interface {
private Stub(com.google.protobuf.RpcChannel channel) {
this.channel = channel;
}
private final com.google.protobuf.RpcChannel channel;
public com.google.protobuf.RpcChannel getChannel() {
return channel;
}
public void getProtocolVersions(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto request,
com.google.protobuf.RpcCallback done) {
channel.callMethod(
getDescriptor().getMethods().get(0),
controller,
request,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance(),
com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.class,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance()));
}
public void getProtocolSignature(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto request,
com.google.protobuf.RpcCallback done) {
channel.callMethod(
getDescriptor().getMethods().get(1),
controller,
request,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance(),
com.google.protobuf.RpcUtil.generalizeCallback(
done,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.class,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance()));
}
}
public static BlockingInterface newBlockingStub(
com.google.protobuf.BlockingRpcChannel channel) {
return new BlockingStub(channel);
}
public interface BlockingInterface {
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto getProtocolVersions(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto request)
throws com.google.protobuf.ServiceException;
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto getProtocolSignature(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto request)
throws com.google.protobuf.ServiceException;
}
private static final class BlockingStub implements BlockingInterface {
private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
this.channel = channel;
}
private final com.google.protobuf.BlockingRpcChannel channel;
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto getProtocolVersions(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsRequestProto request)
throws com.google.protobuf.ServiceException {
return (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(0),
controller,
request,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolVersionsResponseProto.getDefaultInstance());
}
public org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto getProtocolSignature(
com.google.protobuf.RpcController controller,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureRequestProto request)
throws com.google.protobuf.ServiceException {
return (org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto) channel.callBlockingMethod(
getDescriptor().getMethods().get(1),
controller,
request,
org.apache.hadoop.ipc.protobuf.ProtocolInfoProtos.GetProtocolSignatureResponseProto.getDefaultInstance());
}
}
// @@protoc_insertion_point(class_scope:hadoop.common.ProtocolInfoService)
}
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_common_GetProtocolVersionsRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_common_GetProtocolVersionsRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_common_ProtocolVersionProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_common_ProtocolVersionProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_common_GetProtocolVersionsResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_common_GetProtocolVersionsResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_common_GetProtocolSignatureRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_common_GetProtocolSignatureRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_common_GetProtocolSignatureResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_common_GetProtocolSignatureResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_common_ProtocolSignatureProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_common_ProtocolSignatureProto_fieldAccessorTable;
public static com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
private static com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\022ProtocolInfo.proto\022\rhadoop.common\"3\n\037G" +
"etProtocolVersionsRequestProto\022\020\n\010protoc" +
"ol\030\001 \002(\t\"9\n\024ProtocolVersionProto\022\017\n\007rpcK" +
"ind\030\001 \002(\t\022\020\n\010versions\030\002 \003(\004\"a\n GetProtoc" +
"olVersionsResponseProto\022=\n\020protocolVersi" +
"ons\030\001 \003(\0132#.hadoop.common.ProtocolVersio" +
"nProto\"E\n GetProtocolSignatureRequestPro" +
"to\022\020\n\010protocol\030\001 \002(\t\022\017\n\007rpcKind\030\002 \002(\t\"e\n" +
"!GetProtocolSignatureResponseProto\022@\n\021pr" +
"otocolSignature\030\001 \003(\0132%.hadoop.common.Pr",
"otocolSignatureProto\":\n\026ProtocolSignatur" +
"eProto\022\017\n\007version\030\001 \002(\004\022\017\n\007methods\030\002 \003(\r" +
"2\210\002\n\023ProtocolInfoService\022v\n\023getProtocolV" +
"ersions\022..hadoop.common.GetProtocolVersi" +
"onsRequestProto\032/.hadoop.common.GetProto" +
"colVersionsResponseProto\022y\n\024getProtocolS" +
"ignature\022/.hadoop.common.GetProtocolSign" +
"atureRequestProto\0320.hadoop.common.GetPro" +
"tocolSignatureResponseProtoB:\n\036org.apach" +
"e.hadoop.ipc.protobufB\022ProtocolInfoProto",
"s\210\001\001\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
public com.google.protobuf.ExtensionRegistry assignDescriptors(
com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
internal_static_hadoop_common_GetProtocolVersionsRequestProto_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hadoop_common_GetProtocolVersionsRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_common_GetProtocolVersionsRequestProto_descriptor,
new java.lang.String[] { "Protocol", });
internal_static_hadoop_common_ProtocolVersionProto_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hadoop_common_ProtocolVersionProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_common_ProtocolVersionProto_descriptor,
new java.lang.String[] { "RpcKind", "Versions", });
internal_static_hadoop_common_GetProtocolVersionsResponseProto_descriptor =
getDescriptor().getMessageTypes().get(2);
internal_static_hadoop_common_GetProtocolVersionsResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_common_GetProtocolVersionsResponseProto_descriptor,
new java.lang.String[] { "ProtocolVersions", });
internal_static_hadoop_common_GetProtocolSignatureRequestProto_descriptor =
getDescriptor().getMessageTypes().get(3);
internal_static_hadoop_common_GetProtocolSignatureRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_common_GetProtocolSignatureRequestProto_descriptor,
new java.lang.String[] { "Protocol", "RpcKind", });
internal_static_hadoop_common_GetProtocolSignatureResponseProto_descriptor =
getDescriptor().getMessageTypes().get(4);
internal_static_hadoop_common_GetProtocolSignatureResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_common_GetProtocolSignatureResponseProto_descriptor,
new java.lang.String[] { "ProtocolSignature", });
internal_static_hadoop_common_ProtocolSignatureProto_descriptor =
getDescriptor().getMessageTypes().get(5);
internal_static_hadoop_common_ProtocolSignatureProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_common_ProtocolSignatureProto_descriptor,
new java.lang.String[] { "Version", "Methods", });
return null;
}
};
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
}
// @@protoc_insertion_point(outer_class_scope)
}
© 2015 - 2024 Weber Informatics LLC | Privacy Policy