All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos Maven / Gradle / Ivy

There is a newer version: 3.4.0
Show newest version
// Generated by the protocol buffer compiler.  DO NOT EDIT!
// source: DatanodeProtocol.proto

package org.apache.hadoop.hdfs.protocol.proto;

public final class DatanodeProtocolProtos {
  private DatanodeProtocolProtos() {}
  public static void registerAllExtensions(
      org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite registry) {
  }

  public static void registerAllExtensions(
      org.apache.hadoop.thirdparty.protobuf.ExtensionRegistry registry) {
    registerAllExtensions(
        (org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite) registry);
  }
  public interface DatanodeRegistrationProtoOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.DatanodeRegistrationProto)
      org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {

    /**
     * 
     * Datanode information
     * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ boolean hasDatanodeID(); /** *
     * Datanode information
     * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto getDatanodeID(); /** *
     * Datanode information
     * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder getDatanodeIDOrBuilder(); /** *
     * Node information
     * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ boolean hasStorageInfo(); /** *
     * Node information
     * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getStorageInfo(); /** *
     * Node information
     * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getStorageInfoOrBuilder(); /** *
     * Block keys
     * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ boolean hasKeys(); /** *
     * Block keys
     * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto getKeys(); /** *
     * Block keys
     * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder getKeysOrBuilder(); /** *
     * Software version of the DN, e.g. "2.0.0"
     * 
* * required string softwareVersion = 4; */ boolean hasSoftwareVersion(); /** *
     * Software version of the DN, e.g. "2.0.0"
     * 
* * required string softwareVersion = 4; */ java.lang.String getSoftwareVersion(); /** *
     * Software version of the DN, e.g. "2.0.0"
     * 
* * required string softwareVersion = 4; */ org.apache.hadoop.thirdparty.protobuf.ByteString getSoftwareVersionBytes(); } /** *
   **
   * Information to identify a datanode to a namenode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.DatanodeRegistrationProto} */ public static final class DatanodeRegistrationProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.DatanodeRegistrationProto) DatanodeRegistrationProtoOrBuilder { private static final long serialVersionUID = 0L; // Use DatanodeRegistrationProto.newBuilder() to construct. private DatanodeRegistrationProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DatanodeRegistrationProto() { softwareVersion_ = ""; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DatanodeRegistrationProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = datanodeID_.toBuilder(); } datanodeID_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(datanodeID_); datanodeID_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = storageInfo_.toBuilder(); } storageInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(storageInfo_); storageInfo_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 26: { org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder subBuilder = null; if (((bitField0_ & 0x00000004) != 0)) { subBuilder = keys_.toBuilder(); } keys_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(keys_); keys_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } case 34: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000008; softwareVersion_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder.class); } private int bitField0_; public static final int DATANODEID_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto datanodeID_; /** *
     * Datanode information
     * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public boolean hasDatanodeID() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Datanode information
     * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto getDatanodeID() { return datanodeID_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance() : datanodeID_; } /** *
     * Datanode information
     * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder getDatanodeIDOrBuilder() { return datanodeID_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance() : datanodeID_; } public static final int STORAGEINFO_FIELD_NUMBER = 2; private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto storageInfo_; /** *
     * Node information
     * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public boolean hasStorageInfo() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * Node information
     * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getStorageInfo() { return storageInfo_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance() : storageInfo_; } /** *
     * Node information
     * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getStorageInfoOrBuilder() { return storageInfo_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance() : storageInfo_; } public static final int KEYS_FIELD_NUMBER = 3; private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto keys_; /** *
     * Block keys
     * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public boolean hasKeys() { return ((bitField0_ & 0x00000004) != 0); } /** *
     * Block keys
     * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto getKeys() { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } /** *
     * Block keys
     * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder getKeysOrBuilder() { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } public static final int SOFTWAREVERSION_FIELD_NUMBER = 4; private volatile java.lang.Object softwareVersion_; /** *
     * Software version of the DN, e.g. "2.0.0"
     * 
* * required string softwareVersion = 4; */ public boolean hasSoftwareVersion() { return ((bitField0_ & 0x00000008) != 0); } /** *
     * Software version of the DN, e.g. "2.0.0"
     * 
* * required string softwareVersion = 4; */ public java.lang.String getSoftwareVersion() { java.lang.Object ref = softwareVersion_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { softwareVersion_ = s; } return s; } } /** *
     * Software version of the DN, e.g. "2.0.0"
     * 
* * required string softwareVersion = 4; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getSoftwareVersionBytes() { java.lang.Object ref = softwareVersion_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); softwareVersion_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasDatanodeID()) { memoizedIsInitialized = 0; return false; } if (!hasStorageInfo()) { memoizedIsInitialized = 0; return false; } if (!hasKeys()) { memoizedIsInitialized = 0; return false; } if (!hasSoftwareVersion()) { memoizedIsInitialized = 0; return false; } if (!getDatanodeID().isInitialized()) { memoizedIsInitialized = 0; return false; } if (!getStorageInfo().isInitialized()) { memoizedIsInitialized = 0; return false; } if (!getKeys().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getDatanodeID()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getStorageInfo()); } if (((bitField0_ & 0x00000004) != 0)) { output.writeMessage(3, getKeys()); } if (((bitField0_ & 0x00000008) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 4, softwareVersion_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getDatanodeID()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(2, getStorageInfo()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, getKeys()); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(4, softwareVersion_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto) obj; if (hasDatanodeID() != other.hasDatanodeID()) return false; if (hasDatanodeID()) { if (!getDatanodeID() .equals(other.getDatanodeID())) return false; } if (hasStorageInfo() != other.hasStorageInfo()) return false; if (hasStorageInfo()) { if (!getStorageInfo() .equals(other.getStorageInfo())) return false; } if (hasKeys() != other.hasKeys()) return false; if (hasKeys()) { if (!getKeys() .equals(other.getKeys())) return false; } if (hasSoftwareVersion() != other.hasSoftwareVersion()) return false; if (hasSoftwareVersion()) { if (!getSoftwareVersion() .equals(other.getSoftwareVersion())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasDatanodeID()) { hash = (37 * hash) + DATANODEID_FIELD_NUMBER; hash = (53 * hash) + getDatanodeID().hashCode(); } if (hasStorageInfo()) { hash = (37 * hash) + STORAGEINFO_FIELD_NUMBER; hash = (53 * hash) + getStorageInfo().hashCode(); } if (hasKeys()) { hash = (37 * hash) + KEYS_FIELD_NUMBER; hash = (53 * hash) + getKeys().hashCode(); } if (hasSoftwareVersion()) { hash = (37 * hash) + SOFTWAREVERSION_FIELD_NUMBER; hash = (53 * hash) + getSoftwareVersion().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Information to identify a datanode to a namenode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.DatanodeRegistrationProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.DatanodeRegistrationProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getDatanodeIDFieldBuilder(); getStorageInfoFieldBuilder(); getKeysFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (datanodeIDBuilder_ == null) { datanodeID_ = null; } else { datanodeIDBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); if (storageInfoBuilder_ == null) { storageInfo_ = null; } else { storageInfoBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); if (keysBuilder_ == null) { keys_ = null; } else { keysBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); softwareVersion_ = ""; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (datanodeIDBuilder_ == null) { result.datanodeID_ = datanodeID_; } else { result.datanodeID_ = datanodeIDBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { if (storageInfoBuilder_ == null) { result.storageInfo_ = storageInfo_; } else { result.storageInfo_ = storageInfoBuilder_.build(); } to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { if (keysBuilder_ == null) { result.keys_ = keys_; } else { result.keys_ = keysBuilder_.build(); } to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { to_bitField0_ |= 0x00000008; } result.softwareVersion_ = softwareVersion_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) return this; if (other.hasDatanodeID()) { mergeDatanodeID(other.getDatanodeID()); } if (other.hasStorageInfo()) { mergeStorageInfo(other.getStorageInfo()); } if (other.hasKeys()) { mergeKeys(other.getKeys()); } if (other.hasSoftwareVersion()) { bitField0_ |= 0x00000008; softwareVersion_ = other.softwareVersion_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasDatanodeID()) { return false; } if (!hasStorageInfo()) { return false; } if (!hasKeys()) { return false; } if (!hasSoftwareVersion()) { return false; } if (!getDatanodeID().isInitialized()) { return false; } if (!getStorageInfo().isInitialized()) { return false; } if (!getKeys().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto datanodeID_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder> datanodeIDBuilder_; /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public boolean hasDatanodeID() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto getDatanodeID() { if (datanodeIDBuilder_ == null) { return datanodeID_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance() : datanodeID_; } else { return datanodeIDBuilder_.getMessage(); } } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public Builder setDatanodeID(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto value) { if (datanodeIDBuilder_ == null) { if (value == null) { throw new NullPointerException(); } datanodeID_ = value; onChanged(); } else { datanodeIDBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public Builder setDatanodeID( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder builderForValue) { if (datanodeIDBuilder_ == null) { datanodeID_ = builderForValue.build(); onChanged(); } else { datanodeIDBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public Builder mergeDatanodeID(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto value) { if (datanodeIDBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && datanodeID_ != null && datanodeID_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance()) { datanodeID_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.newBuilder(datanodeID_).mergeFrom(value).buildPartial(); } else { datanodeID_ = value; } onChanged(); } else { datanodeIDBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public Builder clearDatanodeID() { if (datanodeIDBuilder_ == null) { datanodeID_ = null; onChanged(); } else { datanodeIDBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder getDatanodeIDBuilder() { bitField0_ |= 0x00000001; onChanged(); return getDatanodeIDFieldBuilder().getBuilder(); } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder getDatanodeIDOrBuilder() { if (datanodeIDBuilder_ != null) { return datanodeIDBuilder_.getMessageOrBuilder(); } else { return datanodeID_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance() : datanodeID_; } } /** *
       * Datanode information
       * 
* * required .hadoop.hdfs.DatanodeIDProto datanodeID = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder> getDatanodeIDFieldBuilder() { if (datanodeIDBuilder_ == null) { datanodeIDBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder>( getDatanodeID(), getParentForChildren(), isClean()); datanodeID_ = null; } return datanodeIDBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto storageInfo_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> storageInfoBuilder_; /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public boolean hasStorageInfo() { return ((bitField0_ & 0x00000002) != 0); } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto getStorageInfo() { if (storageInfoBuilder_ == null) { return storageInfo_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance() : storageInfo_; } else { return storageInfoBuilder_.getMessage(); } } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public Builder setStorageInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) { if (storageInfoBuilder_ == null) { if (value == null) { throw new NullPointerException(); } storageInfo_ = value; onChanged(); } else { storageInfoBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public Builder setStorageInfo( org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder builderForValue) { if (storageInfoBuilder_ == null) { storageInfo_ = builderForValue.build(); onChanged(); } else { storageInfoBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public Builder mergeStorageInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto value) { if (storageInfoBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && storageInfo_ != null && storageInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance()) { storageInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.newBuilder(storageInfo_).mergeFrom(value).buildPartial(); } else { storageInfo_ = value; } onChanged(); } else { storageInfoBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public Builder clearStorageInfo() { if (storageInfoBuilder_ == null) { storageInfo_ = null; onChanged(); } else { storageInfoBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder getStorageInfoBuilder() { bitField0_ |= 0x00000002; onChanged(); return getStorageInfoFieldBuilder().getBuilder(); } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder getStorageInfoOrBuilder() { if (storageInfoBuilder_ != null) { return storageInfoBuilder_.getMessageOrBuilder(); } else { return storageInfo_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.getDefaultInstance() : storageInfo_; } } /** *
       * Node information
       * 
* * required .hadoop.hdfs.StorageInfoProto storageInfo = 2; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder> getStorageInfoFieldBuilder() { if (storageInfoBuilder_ == null) { storageInfoBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.StorageInfoProtoOrBuilder>( getStorageInfo(), getParentForChildren(), isClean()); storageInfo_ = null; } return storageInfoBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto keys_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder> keysBuilder_; /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public boolean hasKeys() { return ((bitField0_ & 0x00000004) != 0); } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto getKeys() { if (keysBuilder_ == null) { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } else { return keysBuilder_.getMessage(); } } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public Builder setKeys(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto value) { if (keysBuilder_ == null) { if (value == null) { throw new NullPointerException(); } keys_ = value; onChanged(); } else { keysBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public Builder setKeys( org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder builderForValue) { if (keysBuilder_ == null) { keys_ = builderForValue.build(); onChanged(); } else { keysBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public Builder mergeKeys(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto value) { if (keysBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0) && keys_ != null && keys_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance()) { keys_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.newBuilder(keys_).mergeFrom(value).buildPartial(); } else { keys_ = value; } onChanged(); } else { keysBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public Builder clearKeys() { if (keysBuilder_ == null) { keys_ = null; onChanged(); } else { keysBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder getKeysBuilder() { bitField0_ |= 0x00000004; onChanged(); return getKeysFieldBuilder().getBuilder(); } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder getKeysOrBuilder() { if (keysBuilder_ != null) { return keysBuilder_.getMessageOrBuilder(); } else { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } } /** *
       * Block keys
       * 
* * required .hadoop.hdfs.ExportedBlockKeysProto keys = 3; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder> getKeysFieldBuilder() { if (keysBuilder_ == null) { keysBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder>( getKeys(), getParentForChildren(), isClean()); keys_ = null; } return keysBuilder_; } private java.lang.Object softwareVersion_ = ""; /** *
       * Software version of the DN, e.g. "2.0.0"
       * 
* * required string softwareVersion = 4; */ public boolean hasSoftwareVersion() { return ((bitField0_ & 0x00000008) != 0); } /** *
       * Software version of the DN, e.g. "2.0.0"
       * 
* * required string softwareVersion = 4; */ public java.lang.String getSoftwareVersion() { java.lang.Object ref = softwareVersion_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { softwareVersion_ = s; } return s; } else { return (java.lang.String) ref; } } /** *
       * Software version of the DN, e.g. "2.0.0"
       * 
* * required string softwareVersion = 4; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getSoftwareVersionBytes() { java.lang.Object ref = softwareVersion_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); softwareVersion_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** *
       * Software version of the DN, e.g. "2.0.0"
       * 
* * required string softwareVersion = 4; */ public Builder setSoftwareVersion( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; softwareVersion_ = value; onChanged(); return this; } /** *
       * Software version of the DN, e.g. "2.0.0"
       * 
* * required string softwareVersion = 4; */ public Builder clearSoftwareVersion() { bitField0_ = (bitField0_ & ~0x00000008); softwareVersion_ = getDefaultInstance().getSoftwareVersion(); onChanged(); return this; } /** *
       * Software version of the DN, e.g. "2.0.0"
       * 
* * required string softwareVersion = 4; */ public Builder setSoftwareVersionBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; softwareVersion_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.DatanodeRegistrationProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.DatanodeRegistrationProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public DatanodeRegistrationProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new DatanodeRegistrationProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DatanodeCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.DatanodeCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Type of the command
     * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ boolean hasCmdType(); /** *
     * Type of the command
     * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type getCmdType(); /** *
     * One of the following command is available when the corresponding
     * cmdType is set
     * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ boolean hasBalancerCmd(); /** *
     * One of the following command is available when the corresponding
     * cmdType is set
     * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto getBalancerCmd(); /** *
     * One of the following command is available when the corresponding
     * cmdType is set
     * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder getBalancerCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ boolean hasBlkCmd(); /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto getBlkCmd(); /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder getBlkCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ boolean hasRecoveryCmd(); /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto getRecoveryCmd(); /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder getRecoveryCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ boolean hasFinalizeCmd(); /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getFinalizeCmd(); /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder getFinalizeCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ boolean hasKeyUpdateCmd(); /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto getKeyUpdateCmd(); /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder getKeyUpdateCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ boolean hasRegisterCmd(); /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto getRegisterCmd(); /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder getRegisterCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ boolean hasBlkIdCmd(); /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto getBlkIdCmd(); /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder getBlkIdCmdOrBuilder(); /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ boolean hasBlkECReconstructionCmd(); /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto getBlkECReconstructionCmd(); /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder getBlkECReconstructionCmdOrBuilder(); } /** *
   **
   * Commands sent from namenode to the datanodes
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.DatanodeCommandProto} */ public static final class DatanodeCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.DatanodeCommandProto) DatanodeCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use DatanodeCommandProto.newBuilder() to construct. private DatanodeCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DatanodeCommandProto() { cmdType_ = 0; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DatanodeCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type value = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; cmdType_ = rawValue; } break; } case 18: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = balancerCmd_.toBuilder(); } balancerCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(balancerCmd_); balancerCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 26: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000004) != 0)) { subBuilder = blkCmd_.toBuilder(); } blkCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(blkCmd_); blkCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } case 34: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000008) != 0)) { subBuilder = recoveryCmd_.toBuilder(); } recoveryCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(recoveryCmd_); recoveryCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000008; break; } case 42: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000010) != 0)) { subBuilder = finalizeCmd_.toBuilder(); } finalizeCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(finalizeCmd_); finalizeCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000010; break; } case 50: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000020) != 0)) { subBuilder = keyUpdateCmd_.toBuilder(); } keyUpdateCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(keyUpdateCmd_); keyUpdateCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000020; break; } case 58: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000040) != 0)) { subBuilder = registerCmd_.toBuilder(); } registerCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registerCmd_); registerCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000040; break; } case 66: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000080) != 0)) { subBuilder = blkIdCmd_.toBuilder(); } blkIdCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(blkIdCmd_); blkIdCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000080; break; } case 74: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000100) != 0)) { subBuilder = blkECReconstructionCmd_.toBuilder(); } blkECReconstructionCmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(blkECReconstructionCmd_); blkECReconstructionCmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000100; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder.class); } /** * Protobuf enum {@code hadoop.hdfs.datanode.DatanodeCommandProto.Type} */ public enum Type implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum { /** * BalancerBandwidthCommand = 0; */ BalancerBandwidthCommand(0), /** * BlockCommand = 1; */ BlockCommand(1), /** * BlockRecoveryCommand = 2; */ BlockRecoveryCommand(2), /** * FinalizeCommand = 3; */ FinalizeCommand(3), /** * KeyUpdateCommand = 4; */ KeyUpdateCommand(4), /** * RegisterCommand = 5; */ RegisterCommand(5), /** * UnusedUpgradeCommand = 6; */ UnusedUpgradeCommand(6), /** * NullDatanodeCommand = 7; */ NullDatanodeCommand(7), /** * BlockIdCommand = 8; */ BlockIdCommand(8), /** * BlockECReconstructionCommand = 9; */ BlockECReconstructionCommand(9), ; /** * BalancerBandwidthCommand = 0; */ public static final int BalancerBandwidthCommand_VALUE = 0; /** * BlockCommand = 1; */ public static final int BlockCommand_VALUE = 1; /** * BlockRecoveryCommand = 2; */ public static final int BlockRecoveryCommand_VALUE = 2; /** * FinalizeCommand = 3; */ public static final int FinalizeCommand_VALUE = 3; /** * KeyUpdateCommand = 4; */ public static final int KeyUpdateCommand_VALUE = 4; /** * RegisterCommand = 5; */ public static final int RegisterCommand_VALUE = 5; /** * UnusedUpgradeCommand = 6; */ public static final int UnusedUpgradeCommand_VALUE = 6; /** * NullDatanodeCommand = 7; */ public static final int NullDatanodeCommand_VALUE = 7; /** * BlockIdCommand = 8; */ public static final int BlockIdCommand_VALUE = 8; /** * BlockECReconstructionCommand = 9; */ public static final int BlockECReconstructionCommand_VALUE = 9; public final int getNumber() { return value; } /** * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static Type valueOf(int value) { return forNumber(value); } public static Type forNumber(int value) { switch (value) { case 0: return BalancerBandwidthCommand; case 1: return BlockCommand; case 2: return BlockRecoveryCommand; case 3: return FinalizeCommand; case 4: return KeyUpdateCommand; case 5: return RegisterCommand; case 6: return UnusedUpgradeCommand; case 7: return NullDatanodeCommand; case 8: return BlockIdCommand; case 9: return BlockECReconstructionCommand; default: return null; } } public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap< Type> internalValueMap = new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() { public Type findValueByNumber(int number) { return Type.forNumber(number); } }; public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDescriptor().getEnumTypes().get(0); } private static final Type[] VALUES = values(); public static Type valueOf( org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private Type(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hadoop.hdfs.datanode.DatanodeCommandProto.Type) } private int bitField0_; public static final int CMDTYPE_FIELD_NUMBER = 1; private int cmdType_; /** *
     * Type of the command
     * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ public boolean hasCmdType() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Type of the command
     * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type getCmdType() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.valueOf(cmdType_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.BalancerBandwidthCommand : result; } public static final int BALANCERCMD_FIELD_NUMBER = 2; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto balancerCmd_; /** *
     * One of the following command is available when the corresponding
     * cmdType is set
     * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public boolean hasBalancerCmd() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * One of the following command is available when the corresponding
     * cmdType is set
     * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto getBalancerCmd() { return balancerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance() : balancerCmd_; } /** *
     * One of the following command is available when the corresponding
     * cmdType is set
     * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder getBalancerCmdOrBuilder() { return balancerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance() : balancerCmd_; } public static final int BLKCMD_FIELD_NUMBER = 3; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto blkCmd_; /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public boolean hasBlkCmd() { return ((bitField0_ & 0x00000004) != 0); } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto getBlkCmd() { return blkCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance() : blkCmd_; } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder getBlkCmdOrBuilder() { return blkCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance() : blkCmd_; } public static final int RECOVERYCMD_FIELD_NUMBER = 4; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto recoveryCmd_; /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public boolean hasRecoveryCmd() { return ((bitField0_ & 0x00000008) != 0); } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto getRecoveryCmd() { return recoveryCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance() : recoveryCmd_; } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder getRecoveryCmdOrBuilder() { return recoveryCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance() : recoveryCmd_; } public static final int FINALIZECMD_FIELD_NUMBER = 5; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto finalizeCmd_; /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public boolean hasFinalizeCmd() { return ((bitField0_ & 0x00000010) != 0); } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getFinalizeCmd() { return finalizeCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance() : finalizeCmd_; } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder getFinalizeCmdOrBuilder() { return finalizeCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance() : finalizeCmd_; } public static final int KEYUPDATECMD_FIELD_NUMBER = 6; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto keyUpdateCmd_; /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public boolean hasKeyUpdateCmd() { return ((bitField0_ & 0x00000020) != 0); } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto getKeyUpdateCmd() { return keyUpdateCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance() : keyUpdateCmd_; } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder getKeyUpdateCmdOrBuilder() { return keyUpdateCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance() : keyUpdateCmd_; } public static final int REGISTERCMD_FIELD_NUMBER = 7; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto registerCmd_; /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public boolean hasRegisterCmd() { return ((bitField0_ & 0x00000040) != 0); } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto getRegisterCmd() { return registerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance() : registerCmd_; } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder getRegisterCmdOrBuilder() { return registerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance() : registerCmd_; } public static final int BLKIDCMD_FIELD_NUMBER = 8; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto blkIdCmd_; /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public boolean hasBlkIdCmd() { return ((bitField0_ & 0x00000080) != 0); } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto getBlkIdCmd() { return blkIdCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance() : blkIdCmd_; } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder getBlkIdCmdOrBuilder() { return blkIdCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance() : blkIdCmd_; } public static final int BLKECRECONSTRUCTIONCMD_FIELD_NUMBER = 9; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto blkECReconstructionCmd_; /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public boolean hasBlkECReconstructionCmd() { return ((bitField0_ & 0x00000100) != 0); } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto getBlkECReconstructionCmd() { return blkECReconstructionCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance() : blkECReconstructionCmd_; } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder getBlkECReconstructionCmdOrBuilder() { return blkECReconstructionCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance() : blkECReconstructionCmd_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasCmdType()) { memoizedIsInitialized = 0; return false; } if (hasBalancerCmd()) { if (!getBalancerCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasBlkCmd()) { if (!getBlkCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasRecoveryCmd()) { if (!getRecoveryCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasFinalizeCmd()) { if (!getFinalizeCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasKeyUpdateCmd()) { if (!getKeyUpdateCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasBlkIdCmd()) { if (!getBlkIdCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasBlkECReconstructionCmd()) { if (!getBlkECReconstructionCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeEnum(1, cmdType_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getBalancerCmd()); } if (((bitField0_ & 0x00000004) != 0)) { output.writeMessage(3, getBlkCmd()); } if (((bitField0_ & 0x00000008) != 0)) { output.writeMessage(4, getRecoveryCmd()); } if (((bitField0_ & 0x00000010) != 0)) { output.writeMessage(5, getFinalizeCmd()); } if (((bitField0_ & 0x00000020) != 0)) { output.writeMessage(6, getKeyUpdateCmd()); } if (((bitField0_ & 0x00000040) != 0)) { output.writeMessage(7, getRegisterCmd()); } if (((bitField0_ & 0x00000080) != 0)) { output.writeMessage(8, getBlkIdCmd()); } if (((bitField0_ & 0x00000100) != 0)) { output.writeMessage(9, getBlkECReconstructionCmd()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeEnumSize(1, cmdType_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(2, getBalancerCmd()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, getBlkCmd()); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(4, getRecoveryCmd()); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(5, getFinalizeCmd()); } if (((bitField0_ & 0x00000020) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(6, getKeyUpdateCmd()); } if (((bitField0_ & 0x00000040) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(7, getRegisterCmd()); } if (((bitField0_ & 0x00000080) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(8, getBlkIdCmd()); } if (((bitField0_ & 0x00000100) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(9, getBlkECReconstructionCmd()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto) obj; if (hasCmdType() != other.hasCmdType()) return false; if (hasCmdType()) { if (cmdType_ != other.cmdType_) return false; } if (hasBalancerCmd() != other.hasBalancerCmd()) return false; if (hasBalancerCmd()) { if (!getBalancerCmd() .equals(other.getBalancerCmd())) return false; } if (hasBlkCmd() != other.hasBlkCmd()) return false; if (hasBlkCmd()) { if (!getBlkCmd() .equals(other.getBlkCmd())) return false; } if (hasRecoveryCmd() != other.hasRecoveryCmd()) return false; if (hasRecoveryCmd()) { if (!getRecoveryCmd() .equals(other.getRecoveryCmd())) return false; } if (hasFinalizeCmd() != other.hasFinalizeCmd()) return false; if (hasFinalizeCmd()) { if (!getFinalizeCmd() .equals(other.getFinalizeCmd())) return false; } if (hasKeyUpdateCmd() != other.hasKeyUpdateCmd()) return false; if (hasKeyUpdateCmd()) { if (!getKeyUpdateCmd() .equals(other.getKeyUpdateCmd())) return false; } if (hasRegisterCmd() != other.hasRegisterCmd()) return false; if (hasRegisterCmd()) { if (!getRegisterCmd() .equals(other.getRegisterCmd())) return false; } if (hasBlkIdCmd() != other.hasBlkIdCmd()) return false; if (hasBlkIdCmd()) { if (!getBlkIdCmd() .equals(other.getBlkIdCmd())) return false; } if (hasBlkECReconstructionCmd() != other.hasBlkECReconstructionCmd()) return false; if (hasBlkECReconstructionCmd()) { if (!getBlkECReconstructionCmd() .equals(other.getBlkECReconstructionCmd())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasCmdType()) { hash = (37 * hash) + CMDTYPE_FIELD_NUMBER; hash = (53 * hash) + cmdType_; } if (hasBalancerCmd()) { hash = (37 * hash) + BALANCERCMD_FIELD_NUMBER; hash = (53 * hash) + getBalancerCmd().hashCode(); } if (hasBlkCmd()) { hash = (37 * hash) + BLKCMD_FIELD_NUMBER; hash = (53 * hash) + getBlkCmd().hashCode(); } if (hasRecoveryCmd()) { hash = (37 * hash) + RECOVERYCMD_FIELD_NUMBER; hash = (53 * hash) + getRecoveryCmd().hashCode(); } if (hasFinalizeCmd()) { hash = (37 * hash) + FINALIZECMD_FIELD_NUMBER; hash = (53 * hash) + getFinalizeCmd().hashCode(); } if (hasKeyUpdateCmd()) { hash = (37 * hash) + KEYUPDATECMD_FIELD_NUMBER; hash = (53 * hash) + getKeyUpdateCmd().hashCode(); } if (hasRegisterCmd()) { hash = (37 * hash) + REGISTERCMD_FIELD_NUMBER; hash = (53 * hash) + getRegisterCmd().hashCode(); } if (hasBlkIdCmd()) { hash = (37 * hash) + BLKIDCMD_FIELD_NUMBER; hash = (53 * hash) + getBlkIdCmd().hashCode(); } if (hasBlkECReconstructionCmd()) { hash = (37 * hash) + BLKECRECONSTRUCTIONCMD_FIELD_NUMBER; hash = (53 * hash) + getBlkECReconstructionCmd().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Commands sent from namenode to the datanodes
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.DatanodeCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.DatanodeCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBalancerCmdFieldBuilder(); getBlkCmdFieldBuilder(); getRecoveryCmdFieldBuilder(); getFinalizeCmdFieldBuilder(); getKeyUpdateCmdFieldBuilder(); getRegisterCmdFieldBuilder(); getBlkIdCmdFieldBuilder(); getBlkECReconstructionCmdFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); cmdType_ = 0; bitField0_ = (bitField0_ & ~0x00000001); if (balancerCmdBuilder_ == null) { balancerCmd_ = null; } else { balancerCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); if (blkCmdBuilder_ == null) { blkCmd_ = null; } else { blkCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); if (recoveryCmdBuilder_ == null) { recoveryCmd_ = null; } else { recoveryCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); if (finalizeCmdBuilder_ == null) { finalizeCmd_ = null; } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); if (keyUpdateCmdBuilder_ == null) { keyUpdateCmd_ = null; } else { keyUpdateCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000020); if (registerCmdBuilder_ == null) { registerCmd_ = null; } else { registerCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000040); if (blkIdCmdBuilder_ == null) { blkIdCmd_ = null; } else { blkIdCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000080); if (blkECReconstructionCmdBuilder_ == null) { blkECReconstructionCmd_ = null; } else { blkECReconstructionCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000100); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.cmdType_ = cmdType_; if (((from_bitField0_ & 0x00000002) != 0)) { if (balancerCmdBuilder_ == null) { result.balancerCmd_ = balancerCmd_; } else { result.balancerCmd_ = balancerCmdBuilder_.build(); } to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { if (blkCmdBuilder_ == null) { result.blkCmd_ = blkCmd_; } else { result.blkCmd_ = blkCmdBuilder_.build(); } to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { if (recoveryCmdBuilder_ == null) { result.recoveryCmd_ = recoveryCmd_; } else { result.recoveryCmd_ = recoveryCmdBuilder_.build(); } to_bitField0_ |= 0x00000008; } if (((from_bitField0_ & 0x00000010) != 0)) { if (finalizeCmdBuilder_ == null) { result.finalizeCmd_ = finalizeCmd_; } else { result.finalizeCmd_ = finalizeCmdBuilder_.build(); } to_bitField0_ |= 0x00000010; } if (((from_bitField0_ & 0x00000020) != 0)) { if (keyUpdateCmdBuilder_ == null) { result.keyUpdateCmd_ = keyUpdateCmd_; } else { result.keyUpdateCmd_ = keyUpdateCmdBuilder_.build(); } to_bitField0_ |= 0x00000020; } if (((from_bitField0_ & 0x00000040) != 0)) { if (registerCmdBuilder_ == null) { result.registerCmd_ = registerCmd_; } else { result.registerCmd_ = registerCmdBuilder_.build(); } to_bitField0_ |= 0x00000040; } if (((from_bitField0_ & 0x00000080) != 0)) { if (blkIdCmdBuilder_ == null) { result.blkIdCmd_ = blkIdCmd_; } else { result.blkIdCmd_ = blkIdCmdBuilder_.build(); } to_bitField0_ |= 0x00000080; } if (((from_bitField0_ & 0x00000100) != 0)) { if (blkECReconstructionCmdBuilder_ == null) { result.blkECReconstructionCmd_ = blkECReconstructionCmd_; } else { result.blkECReconstructionCmd_ = blkECReconstructionCmdBuilder_.build(); } to_bitField0_ |= 0x00000100; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance()) return this; if (other.hasCmdType()) { setCmdType(other.getCmdType()); } if (other.hasBalancerCmd()) { mergeBalancerCmd(other.getBalancerCmd()); } if (other.hasBlkCmd()) { mergeBlkCmd(other.getBlkCmd()); } if (other.hasRecoveryCmd()) { mergeRecoveryCmd(other.getRecoveryCmd()); } if (other.hasFinalizeCmd()) { mergeFinalizeCmd(other.getFinalizeCmd()); } if (other.hasKeyUpdateCmd()) { mergeKeyUpdateCmd(other.getKeyUpdateCmd()); } if (other.hasRegisterCmd()) { mergeRegisterCmd(other.getRegisterCmd()); } if (other.hasBlkIdCmd()) { mergeBlkIdCmd(other.getBlkIdCmd()); } if (other.hasBlkECReconstructionCmd()) { mergeBlkECReconstructionCmd(other.getBlkECReconstructionCmd()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasCmdType()) { return false; } if (hasBalancerCmd()) { if (!getBalancerCmd().isInitialized()) { return false; } } if (hasBlkCmd()) { if (!getBlkCmd().isInitialized()) { return false; } } if (hasRecoveryCmd()) { if (!getRecoveryCmd().isInitialized()) { return false; } } if (hasFinalizeCmd()) { if (!getFinalizeCmd().isInitialized()) { return false; } } if (hasKeyUpdateCmd()) { if (!getKeyUpdateCmd().isInitialized()) { return false; } } if (hasBlkIdCmd()) { if (!getBlkIdCmd().isInitialized()) { return false; } } if (hasBlkECReconstructionCmd()) { if (!getBlkECReconstructionCmd().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int cmdType_ = 0; /** *
       * Type of the command
       * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ public boolean hasCmdType() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Type of the command
       * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type getCmdType() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.valueOf(cmdType_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type.BalancerBandwidthCommand : result; } /** *
       * Type of the command
       * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ public Builder setCmdType(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Type value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; cmdType_ = value.getNumber(); onChanged(); return this; } /** *
       * Type of the command
       * 
* * required .hadoop.hdfs.datanode.DatanodeCommandProto.Type cmdType = 1; */ public Builder clearCmdType() { bitField0_ = (bitField0_ & ~0x00000001); cmdType_ = 0; onChanged(); return this; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto balancerCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder> balancerCmdBuilder_; /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public boolean hasBalancerCmd() { return ((bitField0_ & 0x00000002) != 0); } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto getBalancerCmd() { if (balancerCmdBuilder_ == null) { return balancerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance() : balancerCmd_; } else { return balancerCmdBuilder_.getMessage(); } } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public Builder setBalancerCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto value) { if (balancerCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } balancerCmd_ = value; onChanged(); } else { balancerCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public Builder setBalancerCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder builderForValue) { if (balancerCmdBuilder_ == null) { balancerCmd_ = builderForValue.build(); onChanged(); } else { balancerCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public Builder mergeBalancerCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto value) { if (balancerCmdBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && balancerCmd_ != null && balancerCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance()) { balancerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.newBuilder(balancerCmd_).mergeFrom(value).buildPartial(); } else { balancerCmd_ = value; } onChanged(); } else { balancerCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public Builder clearBalancerCmd() { if (balancerCmdBuilder_ == null) { balancerCmd_ = null; onChanged(); } else { balancerCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder getBalancerCmdBuilder() { bitField0_ |= 0x00000002; onChanged(); return getBalancerCmdFieldBuilder().getBuilder(); } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder getBalancerCmdOrBuilder() { if (balancerCmdBuilder_ != null) { return balancerCmdBuilder_.getMessageOrBuilder(); } else { return balancerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance() : balancerCmd_; } } /** *
       * One of the following command is available when the corresponding
       * cmdType is set
       * 
* * optional .hadoop.hdfs.datanode.BalancerBandwidthCommandProto balancerCmd = 2; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder> getBalancerCmdFieldBuilder() { if (balancerCmdBuilder_ == null) { balancerCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder>( getBalancerCmd(), getParentForChildren(), isClean()); balancerCmd_ = null; } return balancerCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto blkCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder> blkCmdBuilder_; /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public boolean hasBlkCmd() { return ((bitField0_ & 0x00000004) != 0); } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto getBlkCmd() { if (blkCmdBuilder_ == null) { return blkCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance() : blkCmd_; } else { return blkCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public Builder setBlkCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto value) { if (blkCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } blkCmd_ = value; onChanged(); } else { blkCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public Builder setBlkCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder builderForValue) { if (blkCmdBuilder_ == null) { blkCmd_ = builderForValue.build(); onChanged(); } else { blkCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public Builder mergeBlkCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto value) { if (blkCmdBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0) && blkCmd_ != null && blkCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance()) { blkCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.newBuilder(blkCmd_).mergeFrom(value).buildPartial(); } else { blkCmd_ = value; } onChanged(); } else { blkCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public Builder clearBlkCmd() { if (blkCmdBuilder_ == null) { blkCmd_ = null; onChanged(); } else { blkCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder getBlkCmdBuilder() { bitField0_ |= 0x00000004; onChanged(); return getBlkCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder getBlkCmdOrBuilder() { if (blkCmdBuilder_ != null) { return blkCmdBuilder_.getMessageOrBuilder(); } else { return blkCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance() : blkCmd_; } } /** * optional .hadoop.hdfs.datanode.BlockCommandProto blkCmd = 3; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder> getBlkCmdFieldBuilder() { if (blkCmdBuilder_ == null) { blkCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder>( getBlkCmd(), getParentForChildren(), isClean()); blkCmd_ = null; } return blkCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto recoveryCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder> recoveryCmdBuilder_; /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public boolean hasRecoveryCmd() { return ((bitField0_ & 0x00000008) != 0); } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto getRecoveryCmd() { if (recoveryCmdBuilder_ == null) { return recoveryCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance() : recoveryCmd_; } else { return recoveryCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public Builder setRecoveryCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto value) { if (recoveryCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } recoveryCmd_ = value; onChanged(); } else { recoveryCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public Builder setRecoveryCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder builderForValue) { if (recoveryCmdBuilder_ == null) { recoveryCmd_ = builderForValue.build(); onChanged(); } else { recoveryCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public Builder mergeRecoveryCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto value) { if (recoveryCmdBuilder_ == null) { if (((bitField0_ & 0x00000008) != 0) && recoveryCmd_ != null && recoveryCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance()) { recoveryCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.newBuilder(recoveryCmd_).mergeFrom(value).buildPartial(); } else { recoveryCmd_ = value; } onChanged(); } else { recoveryCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public Builder clearRecoveryCmd() { if (recoveryCmdBuilder_ == null) { recoveryCmd_ = null; onChanged(); } else { recoveryCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); return this; } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder getRecoveryCmdBuilder() { bitField0_ |= 0x00000008; onChanged(); return getRecoveryCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder getRecoveryCmdOrBuilder() { if (recoveryCmdBuilder_ != null) { return recoveryCmdBuilder_.getMessageOrBuilder(); } else { return recoveryCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance() : recoveryCmd_; } } /** * optional .hadoop.hdfs.datanode.BlockRecoveryCommandProto recoveryCmd = 4; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder> getRecoveryCmdFieldBuilder() { if (recoveryCmdBuilder_ == null) { recoveryCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder>( getRecoveryCmd(), getParentForChildren(), isClean()); recoveryCmd_ = null; } return recoveryCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto finalizeCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder> finalizeCmdBuilder_; /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public boolean hasFinalizeCmd() { return ((bitField0_ & 0x00000010) != 0); } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getFinalizeCmd() { if (finalizeCmdBuilder_ == null) { return finalizeCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance() : finalizeCmd_; } else { return finalizeCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public Builder setFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } finalizeCmd_ = value; onChanged(); } else { finalizeCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000010; return this; } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public Builder setFinalizeCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder builderForValue) { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = builderForValue.build(); onChanged(); } else { finalizeCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000010; return this; } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public Builder mergeFinalizeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto value) { if (finalizeCmdBuilder_ == null) { if (((bitField0_ & 0x00000010) != 0) && finalizeCmd_ != null && finalizeCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) { finalizeCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder(finalizeCmd_).mergeFrom(value).buildPartial(); } else { finalizeCmd_ = value; } onChanged(); } else { finalizeCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public Builder clearFinalizeCmd() { if (finalizeCmdBuilder_ == null) { finalizeCmd_ = null; onChanged(); } else { finalizeCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder getFinalizeCmdBuilder() { bitField0_ |= 0x00000010; onChanged(); return getFinalizeCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder getFinalizeCmdOrBuilder() { if (finalizeCmdBuilder_ != null) { return finalizeCmdBuilder_.getMessageOrBuilder(); } else { return finalizeCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance() : finalizeCmd_; } } /** * optional .hadoop.hdfs.datanode.FinalizeCommandProto finalizeCmd = 5; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder> getFinalizeCmdFieldBuilder() { if (finalizeCmdBuilder_ == null) { finalizeCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder>( getFinalizeCmd(), getParentForChildren(), isClean()); finalizeCmd_ = null; } return finalizeCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto keyUpdateCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder> keyUpdateCmdBuilder_; /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public boolean hasKeyUpdateCmd() { return ((bitField0_ & 0x00000020) != 0); } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto getKeyUpdateCmd() { if (keyUpdateCmdBuilder_ == null) { return keyUpdateCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance() : keyUpdateCmd_; } else { return keyUpdateCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public Builder setKeyUpdateCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto value) { if (keyUpdateCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } keyUpdateCmd_ = value; onChanged(); } else { keyUpdateCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000020; return this; } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public Builder setKeyUpdateCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder builderForValue) { if (keyUpdateCmdBuilder_ == null) { keyUpdateCmd_ = builderForValue.build(); onChanged(); } else { keyUpdateCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000020; return this; } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public Builder mergeKeyUpdateCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto value) { if (keyUpdateCmdBuilder_ == null) { if (((bitField0_ & 0x00000020) != 0) && keyUpdateCmd_ != null && keyUpdateCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance()) { keyUpdateCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.newBuilder(keyUpdateCmd_).mergeFrom(value).buildPartial(); } else { keyUpdateCmd_ = value; } onChanged(); } else { keyUpdateCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000020; return this; } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public Builder clearKeyUpdateCmd() { if (keyUpdateCmdBuilder_ == null) { keyUpdateCmd_ = null; onChanged(); } else { keyUpdateCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000020); return this; } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder getKeyUpdateCmdBuilder() { bitField0_ |= 0x00000020; onChanged(); return getKeyUpdateCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder getKeyUpdateCmdOrBuilder() { if (keyUpdateCmdBuilder_ != null) { return keyUpdateCmdBuilder_.getMessageOrBuilder(); } else { return keyUpdateCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance() : keyUpdateCmd_; } } /** * optional .hadoop.hdfs.datanode.KeyUpdateCommandProto keyUpdateCmd = 6; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder> getKeyUpdateCmdFieldBuilder() { if (keyUpdateCmdBuilder_ == null) { keyUpdateCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder>( getKeyUpdateCmd(), getParentForChildren(), isClean()); keyUpdateCmd_ = null; } return keyUpdateCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto registerCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder> registerCmdBuilder_; /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public boolean hasRegisterCmd() { return ((bitField0_ & 0x00000040) != 0); } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto getRegisterCmd() { if (registerCmdBuilder_ == null) { return registerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance() : registerCmd_; } else { return registerCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public Builder setRegisterCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto value) { if (registerCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registerCmd_ = value; onChanged(); } else { registerCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000040; return this; } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public Builder setRegisterCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder builderForValue) { if (registerCmdBuilder_ == null) { registerCmd_ = builderForValue.build(); onChanged(); } else { registerCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000040; return this; } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public Builder mergeRegisterCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto value) { if (registerCmdBuilder_ == null) { if (((bitField0_ & 0x00000040) != 0) && registerCmd_ != null && registerCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance()) { registerCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.newBuilder(registerCmd_).mergeFrom(value).buildPartial(); } else { registerCmd_ = value; } onChanged(); } else { registerCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000040; return this; } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public Builder clearRegisterCmd() { if (registerCmdBuilder_ == null) { registerCmd_ = null; onChanged(); } else { registerCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000040); return this; } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder getRegisterCmdBuilder() { bitField0_ |= 0x00000040; onChanged(); return getRegisterCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder getRegisterCmdOrBuilder() { if (registerCmdBuilder_ != null) { return registerCmdBuilder_.getMessageOrBuilder(); } else { return registerCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance() : registerCmd_; } } /** * optional .hadoop.hdfs.datanode.RegisterCommandProto registerCmd = 7; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder> getRegisterCmdFieldBuilder() { if (registerCmdBuilder_ == null) { registerCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder>( getRegisterCmd(), getParentForChildren(), isClean()); registerCmd_ = null; } return registerCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto blkIdCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder> blkIdCmdBuilder_; /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public boolean hasBlkIdCmd() { return ((bitField0_ & 0x00000080) != 0); } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto getBlkIdCmd() { if (blkIdCmdBuilder_ == null) { return blkIdCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance() : blkIdCmd_; } else { return blkIdCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public Builder setBlkIdCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto value) { if (blkIdCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } blkIdCmd_ = value; onChanged(); } else { blkIdCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000080; return this; } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public Builder setBlkIdCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder builderForValue) { if (blkIdCmdBuilder_ == null) { blkIdCmd_ = builderForValue.build(); onChanged(); } else { blkIdCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000080; return this; } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public Builder mergeBlkIdCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto value) { if (blkIdCmdBuilder_ == null) { if (((bitField0_ & 0x00000080) != 0) && blkIdCmd_ != null && blkIdCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance()) { blkIdCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.newBuilder(blkIdCmd_).mergeFrom(value).buildPartial(); } else { blkIdCmd_ = value; } onChanged(); } else { blkIdCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000080; return this; } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public Builder clearBlkIdCmd() { if (blkIdCmdBuilder_ == null) { blkIdCmd_ = null; onChanged(); } else { blkIdCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000080); return this; } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder getBlkIdCmdBuilder() { bitField0_ |= 0x00000080; onChanged(); return getBlkIdCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder getBlkIdCmdOrBuilder() { if (blkIdCmdBuilder_ != null) { return blkIdCmdBuilder_.getMessageOrBuilder(); } else { return blkIdCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance() : blkIdCmd_; } } /** * optional .hadoop.hdfs.datanode.BlockIdCommandProto blkIdCmd = 8; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder> getBlkIdCmdFieldBuilder() { if (blkIdCmdBuilder_ == null) { blkIdCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder>( getBlkIdCmd(), getParentForChildren(), isClean()); blkIdCmd_ = null; } return blkIdCmdBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto blkECReconstructionCmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder> blkECReconstructionCmdBuilder_; /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public boolean hasBlkECReconstructionCmd() { return ((bitField0_ & 0x00000100) != 0); } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto getBlkECReconstructionCmd() { if (blkECReconstructionCmdBuilder_ == null) { return blkECReconstructionCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance() : blkECReconstructionCmd_; } else { return blkECReconstructionCmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public Builder setBlkECReconstructionCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto value) { if (blkECReconstructionCmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } blkECReconstructionCmd_ = value; onChanged(); } else { blkECReconstructionCmdBuilder_.setMessage(value); } bitField0_ |= 0x00000100; return this; } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public Builder setBlkECReconstructionCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder builderForValue) { if (blkECReconstructionCmdBuilder_ == null) { blkECReconstructionCmd_ = builderForValue.build(); onChanged(); } else { blkECReconstructionCmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000100; return this; } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public Builder mergeBlkECReconstructionCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto value) { if (blkECReconstructionCmdBuilder_ == null) { if (((bitField0_ & 0x00000100) != 0) && blkECReconstructionCmd_ != null && blkECReconstructionCmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance()) { blkECReconstructionCmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.newBuilder(blkECReconstructionCmd_).mergeFrom(value).buildPartial(); } else { blkECReconstructionCmd_ = value; } onChanged(); } else { blkECReconstructionCmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000100; return this; } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public Builder clearBlkECReconstructionCmd() { if (blkECReconstructionCmdBuilder_ == null) { blkECReconstructionCmd_ = null; onChanged(); } else { blkECReconstructionCmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000100); return this; } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder getBlkECReconstructionCmdBuilder() { bitField0_ |= 0x00000100; onChanged(); return getBlkECReconstructionCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder getBlkECReconstructionCmdOrBuilder() { if (blkECReconstructionCmdBuilder_ != null) { return blkECReconstructionCmdBuilder_.getMessageOrBuilder(); } else { return blkECReconstructionCmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance() : blkECReconstructionCmd_; } } /** * optional .hadoop.hdfs.datanode.BlockECReconstructionCommandProto blkECReconstructionCmd = 9; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder> getBlkECReconstructionCmdFieldBuilder() { if (blkECReconstructionCmdBuilder_ == null) { blkECReconstructionCmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder>( getBlkECReconstructionCmd(), getParentForChildren(), isClean()); blkECReconstructionCmd_ = null; } return blkECReconstructionCmdBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.DatanodeCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.DatanodeCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public DatanodeCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new DatanodeCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalancerBandwidthCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BalancerBandwidthCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Maximum bandwidth to be used by datanode for balancing
     * 
* * required uint64 bandwidth = 1; */ boolean hasBandwidth(); /** *
     * Maximum bandwidth to be used by datanode for balancing
     * 
* * required uint64 bandwidth = 1; */ long getBandwidth(); } /** *
   **
   * Command sent from namenode to datanode to set the
   * maximum bandwidth to be used for balancing.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BalancerBandwidthCommandProto} */ public static final class BalancerBandwidthCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BalancerBandwidthCommandProto) BalancerBandwidthCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BalancerBandwidthCommandProto.newBuilder() to construct. private BalancerBandwidthCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalancerBandwidthCommandProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalancerBandwidthCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; bandwidth_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder.class); } private int bitField0_; public static final int BANDWIDTH_FIELD_NUMBER = 1; private long bandwidth_; /** *
     * Maximum bandwidth to be used by datanode for balancing
     * 
* * required uint64 bandwidth = 1; */ public boolean hasBandwidth() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Maximum bandwidth to be used by datanode for balancing
     * 
* * required uint64 bandwidth = 1; */ public long getBandwidth() { return bandwidth_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasBandwidth()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, bandwidth_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(1, bandwidth_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto) obj; if (hasBandwidth() != other.hasBandwidth()) return false; if (hasBandwidth()) { if (getBandwidth() != other.getBandwidth()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasBandwidth()) { hash = (37 * hash) + BANDWIDTH_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getBandwidth()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Command sent from namenode to datanode to set the
     * maximum bandwidth to be used for balancing.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BalancerBandwidthCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BalancerBandwidthCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); bandwidth_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.bandwidth_ = bandwidth_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto.getDefaultInstance()) return this; if (other.hasBandwidth()) { setBandwidth(other.getBandwidth()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasBandwidth()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long bandwidth_ ; /** *
       * Maximum bandwidth to be used by datanode for balancing
       * 
* * required uint64 bandwidth = 1; */ public boolean hasBandwidth() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Maximum bandwidth to be used by datanode for balancing
       * 
* * required uint64 bandwidth = 1; */ public long getBandwidth() { return bandwidth_; } /** *
       * Maximum bandwidth to be used by datanode for balancing
       * 
* * required uint64 bandwidth = 1; */ public Builder setBandwidth(long value) { bitField0_ |= 0x00000001; bandwidth_ = value; onChanged(); return this; } /** *
       * Maximum bandwidth to be used by datanode for balancing
       * 
* * required uint64 bandwidth = 1; */ public Builder clearBandwidth() { bitField0_ = (bitField0_ & ~0x00000001); bandwidth_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BalancerBandwidthCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BalancerBandwidthCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BalancerBandwidthCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BalancerBandwidthCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BalancerBandwidthCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ boolean hasAction(); /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action getAction(); /** * required string blockPoolId = 2; */ boolean hasBlockPoolId(); /** * required string blockPoolId = 2; */ java.lang.String getBlockPoolId(); /** * required string blockPoolId = 2; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes(); /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ java.util.List getBlocksList(); /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index); /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ int getBlocksCount(); /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ java.util.List getBlocksOrBuilderList(); /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( int index); /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ java.util.List getTargetsList(); /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto getTargets(int index); /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ int getTargetsCount(); /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ java.util.List getTargetsOrBuilderList(); /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProtoOrBuilder getTargetsOrBuilder( int index); /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ java.util.List getTargetStorageUuidsList(); /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto getTargetStorageUuids(int index); /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ int getTargetStorageUuidsCount(); /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ java.util.List getTargetStorageUuidsOrBuilderList(); /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProtoOrBuilder getTargetStorageUuidsOrBuilder( int index); /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ java.util.List getTargetStorageTypesList(); /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto getTargetStorageTypes(int index); /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ int getTargetStorageTypesCount(); /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ java.util.List getTargetStorageTypesOrBuilderList(); /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProtoOrBuilder getTargetStorageTypesOrBuilder( int index); } /** *
   **
   * Command to instruct datanodes to perform certain action
   * on the given set of blocks.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockCommandProto} */ public static final class BlockCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockCommandProto) BlockCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockCommandProto.newBuilder() to construct. private BlockCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockCommandProto() { action_ = 1; blockPoolId_ = ""; blocks_ = java.util.Collections.emptyList(); targets_ = java.util.Collections.emptyList(); targetStorageUuids_ = java.util.Collections.emptyList(); targetStorageTypes_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action value = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; action_ = rawValue; } break; } case 18: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; blockPoolId_ = bs; break; } case 26: { if (!((mutable_bitField0_ & 0x00000004) != 0)) { blocks_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000004; } blocks_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.PARSER, extensionRegistry)); break; } case 34: { if (!((mutable_bitField0_ & 0x00000008) != 0)) { targets_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000008; } targets_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.PARSER, extensionRegistry)); break; } case 42: { if (!((mutable_bitField0_ & 0x00000010) != 0)) { targetStorageUuids_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000010; } targetStorageUuids_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.PARSER, extensionRegistry)); break; } case 50: { if (!((mutable_bitField0_ & 0x00000020) != 0)) { targetStorageTypes_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000020; } targetStorageTypes_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); } if (((mutable_bitField0_ & 0x00000008) != 0)) { targets_ = java.util.Collections.unmodifiableList(targets_); } if (((mutable_bitField0_ & 0x00000010) != 0)) { targetStorageUuids_ = java.util.Collections.unmodifiableList(targetStorageUuids_); } if (((mutable_bitField0_ & 0x00000020) != 0)) { targetStorageTypes_ = java.util.Collections.unmodifiableList(targetStorageTypes_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder.class); } /** * Protobuf enum {@code hadoop.hdfs.datanode.BlockCommandProto.Action} */ public enum Action implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum { /** *
       * Transfer blocks to another datanode
       * 
* * TRANSFER = 1; */ TRANSFER(1), /** *
       * Invalidate blocks
       * 
* * INVALIDATE = 2; */ INVALIDATE(2), /** *
       * Shutdown the datanode
       * 
* * SHUTDOWN = 3; */ SHUTDOWN(3), ; /** *
       * Transfer blocks to another datanode
       * 
* * TRANSFER = 1; */ public static final int TRANSFER_VALUE = 1; /** *
       * Invalidate blocks
       * 
* * INVALIDATE = 2; */ public static final int INVALIDATE_VALUE = 2; /** *
       * Shutdown the datanode
       * 
* * SHUTDOWN = 3; */ public static final int SHUTDOWN_VALUE = 3; public final int getNumber() { return value; } /** * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static Action valueOf(int value) { return forNumber(value); } public static Action forNumber(int value) { switch (value) { case 1: return TRANSFER; case 2: return INVALIDATE; case 3: return SHUTDOWN; default: return null; } } public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap< Action> internalValueMap = new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() { public Action findValueByNumber(int number) { return Action.forNumber(number); } }; public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDescriptor().getEnumTypes().get(0); } private static final Action[] VALUES = values(); public static Action valueOf( org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private Action(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hadoop.hdfs.datanode.BlockCommandProto.Action) } private int bitField0_; public static final int ACTION_FIELD_NUMBER = 1; private int action_; /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ public boolean hasAction() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action getAction() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action.valueOf(action_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action.TRANSFER : result; } public static final int BLOCKPOOLID_FIELD_NUMBER = 2; private volatile java.lang.Object blockPoolId_; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int BLOCKS_FIELD_NUMBER = 3; private java.util.List blocks_; /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public java.util.List getBlocksOrBuilderList() { return blocks_; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index) { return blocks_.get(index); } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( int index) { return blocks_.get(index); } public static final int TARGETS_FIELD_NUMBER = 4; private java.util.List targets_; /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public java.util.List getTargetsList() { return targets_; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public java.util.List getTargetsOrBuilderList() { return targets_; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public int getTargetsCount() { return targets_.size(); } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto getTargets(int index) { return targets_.get(index); } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProtoOrBuilder getTargetsOrBuilder( int index) { return targets_.get(index); } public static final int TARGETSTORAGEUUIDS_FIELD_NUMBER = 5; private java.util.List targetStorageUuids_; /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public java.util.List getTargetStorageUuidsList() { return targetStorageUuids_; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public java.util.List getTargetStorageUuidsOrBuilderList() { return targetStorageUuids_; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public int getTargetStorageUuidsCount() { return targetStorageUuids_.size(); } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto getTargetStorageUuids(int index) { return targetStorageUuids_.get(index); } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProtoOrBuilder getTargetStorageUuidsOrBuilder( int index) { return targetStorageUuids_.get(index); } public static final int TARGETSTORAGETYPES_FIELD_NUMBER = 6; private java.util.List targetStorageTypes_; /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public java.util.List getTargetStorageTypesList() { return targetStorageTypes_; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public java.util.List getTargetStorageTypesOrBuilderList() { return targetStorageTypes_; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public int getTargetStorageTypesCount() { return targetStorageTypes_.size(); } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto getTargetStorageTypes(int index) { return targetStorageTypes_.get(index); } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProtoOrBuilder getTargetStorageTypesOrBuilder( int index) { return targetStorageTypes_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasAction()) { memoizedIsInitialized = 0; return false; } if (!hasBlockPoolId()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } for (int i = 0; i < getTargetsCount(); i++) { if (!getTargets(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeEnum(1, action_); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, blockPoolId_); } for (int i = 0; i < blocks_.size(); i++) { output.writeMessage(3, blocks_.get(i)); } for (int i = 0; i < targets_.size(); i++) { output.writeMessage(4, targets_.get(i)); } for (int i = 0; i < targetStorageUuids_.size(); i++) { output.writeMessage(5, targetStorageUuids_.get(i)); } for (int i = 0; i < targetStorageTypes_.size(); i++) { output.writeMessage(6, targetStorageTypes_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeEnumSize(1, action_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, blockPoolId_); } for (int i = 0; i < blocks_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, blocks_.get(i)); } for (int i = 0; i < targets_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(4, targets_.get(i)); } for (int i = 0; i < targetStorageUuids_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(5, targetStorageUuids_.get(i)); } for (int i = 0; i < targetStorageTypes_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(6, targetStorageTypes_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto) obj; if (hasAction() != other.hasAction()) return false; if (hasAction()) { if (action_ != other.action_) return false; } if (hasBlockPoolId() != other.hasBlockPoolId()) return false; if (hasBlockPoolId()) { if (!getBlockPoolId() .equals(other.getBlockPoolId())) return false; } if (!getBlocksList() .equals(other.getBlocksList())) return false; if (!getTargetsList() .equals(other.getTargetsList())) return false; if (!getTargetStorageUuidsList() .equals(other.getTargetStorageUuidsList())) return false; if (!getTargetStorageTypesList() .equals(other.getTargetStorageTypesList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasAction()) { hash = (37 * hash) + ACTION_FIELD_NUMBER; hash = (53 * hash) + action_; } if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } if (getTargetsCount() > 0) { hash = (37 * hash) + TARGETS_FIELD_NUMBER; hash = (53 * hash) + getTargetsList().hashCode(); } if (getTargetStorageUuidsCount() > 0) { hash = (37 * hash) + TARGETSTORAGEUUIDS_FIELD_NUMBER; hash = (53 * hash) + getTargetStorageUuidsList().hashCode(); } if (getTargetStorageTypesCount() > 0) { hash = (37 * hash) + TARGETSTORAGETYPES_FIELD_NUMBER; hash = (53 * hash) + getTargetStorageTypesList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Command to instruct datanodes to perform certain action
     * on the given set of blocks.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlocksFieldBuilder(); getTargetsFieldBuilder(); getTargetStorageUuidsFieldBuilder(); getTargetStorageTypesFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); action_ = 1; bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); } else { blocksBuilder_.clear(); } if (targetsBuilder_ == null) { targets_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); } else { targetsBuilder_.clear(); } if (targetStorageUuidsBuilder_ == null) { targetStorageUuids_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000010); } else { targetStorageUuidsBuilder_.clear(); } if (targetStorageTypesBuilder_ == null) { targetStorageTypes_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); } else { targetStorageTypesBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.action_ = action_; if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.blockPoolId_ = blockPoolId_; if (blocksBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); bitField0_ = (bitField0_ & ~0x00000004); } result.blocks_ = blocks_; } else { result.blocks_ = blocksBuilder_.build(); } if (targetsBuilder_ == null) { if (((bitField0_ & 0x00000008) != 0)) { targets_ = java.util.Collections.unmodifiableList(targets_); bitField0_ = (bitField0_ & ~0x00000008); } result.targets_ = targets_; } else { result.targets_ = targetsBuilder_.build(); } if (targetStorageUuidsBuilder_ == null) { if (((bitField0_ & 0x00000010) != 0)) { targetStorageUuids_ = java.util.Collections.unmodifiableList(targetStorageUuids_); bitField0_ = (bitField0_ & ~0x00000010); } result.targetStorageUuids_ = targetStorageUuids_; } else { result.targetStorageUuids_ = targetStorageUuidsBuilder_.build(); } if (targetStorageTypesBuilder_ == null) { if (((bitField0_ & 0x00000020) != 0)) { targetStorageTypes_ = java.util.Collections.unmodifiableList(targetStorageTypes_); bitField0_ = (bitField0_ & ~0x00000020); } result.targetStorageTypes_ = targetStorageTypes_; } else { result.targetStorageTypes_ = targetStorageTypesBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.getDefaultInstance()) return this; if (other.hasAction()) { setAction(other.getAction()); } if (other.hasBlockPoolId()) { bitField0_ |= 0x00000002; blockPoolId_ = other.blockPoolId_; onChanged(); } if (blocksBuilder_ == null) { if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } } else { if (!other.blocks_.isEmpty()) { if (blocksBuilder_.isEmpty()) { blocksBuilder_.dispose(); blocksBuilder_ = null; blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000004); blocksBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBlocksFieldBuilder() : null; } else { blocksBuilder_.addAllMessages(other.blocks_); } } } if (targetsBuilder_ == null) { if (!other.targets_.isEmpty()) { if (targets_.isEmpty()) { targets_ = other.targets_; bitField0_ = (bitField0_ & ~0x00000008); } else { ensureTargetsIsMutable(); targets_.addAll(other.targets_); } onChanged(); } } else { if (!other.targets_.isEmpty()) { if (targetsBuilder_.isEmpty()) { targetsBuilder_.dispose(); targetsBuilder_ = null; targets_ = other.targets_; bitField0_ = (bitField0_ & ~0x00000008); targetsBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTargetsFieldBuilder() : null; } else { targetsBuilder_.addAllMessages(other.targets_); } } } if (targetStorageUuidsBuilder_ == null) { if (!other.targetStorageUuids_.isEmpty()) { if (targetStorageUuids_.isEmpty()) { targetStorageUuids_ = other.targetStorageUuids_; bitField0_ = (bitField0_ & ~0x00000010); } else { ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.addAll(other.targetStorageUuids_); } onChanged(); } } else { if (!other.targetStorageUuids_.isEmpty()) { if (targetStorageUuidsBuilder_.isEmpty()) { targetStorageUuidsBuilder_.dispose(); targetStorageUuidsBuilder_ = null; targetStorageUuids_ = other.targetStorageUuids_; bitField0_ = (bitField0_ & ~0x00000010); targetStorageUuidsBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTargetStorageUuidsFieldBuilder() : null; } else { targetStorageUuidsBuilder_.addAllMessages(other.targetStorageUuids_); } } } if (targetStorageTypesBuilder_ == null) { if (!other.targetStorageTypes_.isEmpty()) { if (targetStorageTypes_.isEmpty()) { targetStorageTypes_ = other.targetStorageTypes_; bitField0_ = (bitField0_ & ~0x00000020); } else { ensureTargetStorageTypesIsMutable(); targetStorageTypes_.addAll(other.targetStorageTypes_); } onChanged(); } } else { if (!other.targetStorageTypes_.isEmpty()) { if (targetStorageTypesBuilder_.isEmpty()) { targetStorageTypesBuilder_.dispose(); targetStorageTypesBuilder_ = null; targetStorageTypes_ = other.targetStorageTypes_; bitField0_ = (bitField0_ & ~0x00000020); targetStorageTypesBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTargetStorageTypesFieldBuilder() : null; } else { targetStorageTypesBuilder_.addAllMessages(other.targetStorageTypes_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasAction()) { return false; } if (!hasBlockPoolId()) { return false; } for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { return false; } } for (int i = 0; i < getTargetsCount(); i++) { if (!getTargets(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int action_ = 1; /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ public boolean hasAction() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action getAction() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action.valueOf(action_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action.TRANSFER : result; } /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ public Builder setAction(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto.Action value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; action_ = value.getNumber(); onChanged(); return this; } /** * required .hadoop.hdfs.datanode.BlockCommandProto.Action action = 1; */ public Builder clearAction() { bitField0_ = (bitField0_ & ~0x00000001); action_ = 1; onChanged(); return this; } private java.lang.Object blockPoolId_ = ""; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * required string blockPoolId = 2; */ public Builder setBlockPoolId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000002); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder setBlockPoolIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } private java.util.List blocks_ = java.util.Collections.emptyList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { blocks_ = new java.util.ArrayList(blocks_); bitField0_ |= 0x00000004; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> blocksBuilder_; /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public java.util.List getBlocksList() { if (blocksBuilder_ == null) { return java.util.Collections.unmodifiableList(blocks_); } else { return blocksBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public int getBlocksCount() { if (blocksBuilder_ == null) { return blocks_.size(); } else { return blocksBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlocks(int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.set(index, value); onChanged(); } else { blocksBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.set(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(value); onChanged(); } else { blocksBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(index, value); onChanged(); } else { blocksBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder addBlocks( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder addAllBlocks( java.lang.Iterable values) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); } else { blocksBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder clearBlocks() { if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); } else { blocksBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public Builder removeBlocks(int index) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.remove(index); onChanged(); } else { blocksBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder getBlocksBuilder( int index) { return getBlocksFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlocksOrBuilder( int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public java.util.List getBlocksOrBuilderList() { if (blocksBuilder_ != null) { return blocksBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(blocks_); } } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.BlockProto blocks = 3; */ public java.util.List getBlocksBuilderList() { return getBlocksFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> getBlocksFieldBuilder() { if (blocksBuilder_ == null) { blocksBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder>( blocks_, ((bitField0_ & 0x00000004) != 0), getParentForChildren(), isClean()); blocks_ = null; } return blocksBuilder_; } private java.util.List targets_ = java.util.Collections.emptyList(); private void ensureTargetsIsMutable() { if (!((bitField0_ & 0x00000008) != 0)) { targets_ = new java.util.ArrayList(targets_); bitField0_ |= 0x00000008; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProtoOrBuilder> targetsBuilder_; /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public java.util.List getTargetsList() { if (targetsBuilder_ == null) { return java.util.Collections.unmodifiableList(targets_); } else { return targetsBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public int getTargetsCount() { if (targetsBuilder_ == null) { return targets_.size(); } else { return targetsBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto getTargets(int index) { if (targetsBuilder_ == null) { return targets_.get(index); } else { return targetsBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder setTargets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto value) { if (targetsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetsIsMutable(); targets_.set(index, value); onChanged(); } else { targetsBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder setTargets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder builderForValue) { if (targetsBuilder_ == null) { ensureTargetsIsMutable(); targets_.set(index, builderForValue.build()); onChanged(); } else { targetsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder addTargets(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto value) { if (targetsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetsIsMutable(); targets_.add(value); onChanged(); } else { targetsBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder addTargets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto value) { if (targetsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetsIsMutable(); targets_.add(index, value); onChanged(); } else { targetsBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder addTargets( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder builderForValue) { if (targetsBuilder_ == null) { ensureTargetsIsMutable(); targets_.add(builderForValue.build()); onChanged(); } else { targetsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder addTargets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder builderForValue) { if (targetsBuilder_ == null) { ensureTargetsIsMutable(); targets_.add(index, builderForValue.build()); onChanged(); } else { targetsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder addAllTargets( java.lang.Iterable values) { if (targetsBuilder_ == null) { ensureTargetsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, targets_); onChanged(); } else { targetsBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder clearTargets() { if (targetsBuilder_ == null) { targets_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); onChanged(); } else { targetsBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public Builder removeTargets(int index) { if (targetsBuilder_ == null) { ensureTargetsIsMutable(); targets_.remove(index); onChanged(); } else { targetsBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder getTargetsBuilder( int index) { return getTargetsFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProtoOrBuilder getTargetsOrBuilder( int index) { if (targetsBuilder_ == null) { return targets_.get(index); } else { return targetsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public java.util.List getTargetsOrBuilderList() { if (targetsBuilder_ != null) { return targetsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(targets_); } } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder addTargetsBuilder() { return getTargetsFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder addTargetsBuilder( int index) { return getTargetsFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.DatanodeInfosProto targets = 4; */ public java.util.List getTargetsBuilderList() { return getTargetsFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProtoOrBuilder> getTargetsFieldBuilder() { if (targetsBuilder_ == null) { targetsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeInfosProtoOrBuilder>( targets_, ((bitField0_ & 0x00000008) != 0), getParentForChildren(), isClean()); targets_ = null; } return targetsBuilder_; } private java.util.List targetStorageUuids_ = java.util.Collections.emptyList(); private void ensureTargetStorageUuidsIsMutable() { if (!((bitField0_ & 0x00000010) != 0)) { targetStorageUuids_ = new java.util.ArrayList(targetStorageUuids_); bitField0_ |= 0x00000010; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProtoOrBuilder> targetStorageUuidsBuilder_; /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public java.util.List getTargetStorageUuidsList() { if (targetStorageUuidsBuilder_ == null) { return java.util.Collections.unmodifiableList(targetStorageUuids_); } else { return targetStorageUuidsBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public int getTargetStorageUuidsCount() { if (targetStorageUuidsBuilder_ == null) { return targetStorageUuids_.size(); } else { return targetStorageUuidsBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto getTargetStorageUuids(int index) { if (targetStorageUuidsBuilder_ == null) { return targetStorageUuids_.get(index); } else { return targetStorageUuidsBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder setTargetStorageUuids( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto value) { if (targetStorageUuidsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.set(index, value); onChanged(); } else { targetStorageUuidsBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder setTargetStorageUuids( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder builderForValue) { if (targetStorageUuidsBuilder_ == null) { ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.set(index, builderForValue.build()); onChanged(); } else { targetStorageUuidsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder addTargetStorageUuids(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto value) { if (targetStorageUuidsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.add(value); onChanged(); } else { targetStorageUuidsBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder addTargetStorageUuids( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto value) { if (targetStorageUuidsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.add(index, value); onChanged(); } else { targetStorageUuidsBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder addTargetStorageUuids( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder builderForValue) { if (targetStorageUuidsBuilder_ == null) { ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.add(builderForValue.build()); onChanged(); } else { targetStorageUuidsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder addTargetStorageUuids( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder builderForValue) { if (targetStorageUuidsBuilder_ == null) { ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.add(index, builderForValue.build()); onChanged(); } else { targetStorageUuidsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder addAllTargetStorageUuids( java.lang.Iterable values) { if (targetStorageUuidsBuilder_ == null) { ensureTargetStorageUuidsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, targetStorageUuids_); onChanged(); } else { targetStorageUuidsBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder clearTargetStorageUuids() { if (targetStorageUuidsBuilder_ == null) { targetStorageUuids_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000010); onChanged(); } else { targetStorageUuidsBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public Builder removeTargetStorageUuids(int index) { if (targetStorageUuidsBuilder_ == null) { ensureTargetStorageUuidsIsMutable(); targetStorageUuids_.remove(index); onChanged(); } else { targetStorageUuidsBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder getTargetStorageUuidsBuilder( int index) { return getTargetStorageUuidsFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProtoOrBuilder getTargetStorageUuidsOrBuilder( int index) { if (targetStorageUuidsBuilder_ == null) { return targetStorageUuids_.get(index); } else { return targetStorageUuidsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public java.util.List getTargetStorageUuidsOrBuilderList() { if (targetStorageUuidsBuilder_ != null) { return targetStorageUuidsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(targetStorageUuids_); } } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder addTargetStorageUuidsBuilder() { return getTargetStorageUuidsFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder addTargetStorageUuidsBuilder( int index) { return getTargetStorageUuidsFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.StorageUuidsProto targetStorageUuids = 5; */ public java.util.List getTargetStorageUuidsBuilderList() { return getTargetStorageUuidsFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProtoOrBuilder> getTargetStorageUuidsFieldBuilder() { if (targetStorageUuidsBuilder_ == null) { targetStorageUuidsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageUuidsProtoOrBuilder>( targetStorageUuids_, ((bitField0_ & 0x00000010) != 0), getParentForChildren(), isClean()); targetStorageUuids_ = null; } return targetStorageUuidsBuilder_; } private java.util.List targetStorageTypes_ = java.util.Collections.emptyList(); private void ensureTargetStorageTypesIsMutable() { if (!((bitField0_ & 0x00000020) != 0)) { targetStorageTypes_ = new java.util.ArrayList(targetStorageTypes_); bitField0_ |= 0x00000020; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProtoOrBuilder> targetStorageTypesBuilder_; /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public java.util.List getTargetStorageTypesList() { if (targetStorageTypesBuilder_ == null) { return java.util.Collections.unmodifiableList(targetStorageTypes_); } else { return targetStorageTypesBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public int getTargetStorageTypesCount() { if (targetStorageTypesBuilder_ == null) { return targetStorageTypes_.size(); } else { return targetStorageTypesBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto getTargetStorageTypes(int index) { if (targetStorageTypesBuilder_ == null) { return targetStorageTypes_.get(index); } else { return targetStorageTypesBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder setTargetStorageTypes( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto value) { if (targetStorageTypesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetStorageTypesIsMutable(); targetStorageTypes_.set(index, value); onChanged(); } else { targetStorageTypesBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder setTargetStorageTypes( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder builderForValue) { if (targetStorageTypesBuilder_ == null) { ensureTargetStorageTypesIsMutable(); targetStorageTypes_.set(index, builderForValue.build()); onChanged(); } else { targetStorageTypesBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder addTargetStorageTypes(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto value) { if (targetStorageTypesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetStorageTypesIsMutable(); targetStorageTypes_.add(value); onChanged(); } else { targetStorageTypesBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder addTargetStorageTypes( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto value) { if (targetStorageTypesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTargetStorageTypesIsMutable(); targetStorageTypes_.add(index, value); onChanged(); } else { targetStorageTypesBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder addTargetStorageTypes( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder builderForValue) { if (targetStorageTypesBuilder_ == null) { ensureTargetStorageTypesIsMutable(); targetStorageTypes_.add(builderForValue.build()); onChanged(); } else { targetStorageTypesBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder addTargetStorageTypes( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder builderForValue) { if (targetStorageTypesBuilder_ == null) { ensureTargetStorageTypesIsMutable(); targetStorageTypes_.add(index, builderForValue.build()); onChanged(); } else { targetStorageTypesBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder addAllTargetStorageTypes( java.lang.Iterable values) { if (targetStorageTypesBuilder_ == null) { ensureTargetStorageTypesIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, targetStorageTypes_); onChanged(); } else { targetStorageTypesBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder clearTargetStorageTypes() { if (targetStorageTypesBuilder_ == null) { targetStorageTypes_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); onChanged(); } else { targetStorageTypesBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public Builder removeTargetStorageTypes(int index) { if (targetStorageTypesBuilder_ == null) { ensureTargetStorageTypesIsMutable(); targetStorageTypes_.remove(index); onChanged(); } else { targetStorageTypesBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder getTargetStorageTypesBuilder( int index) { return getTargetStorageTypesFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProtoOrBuilder getTargetStorageTypesOrBuilder( int index) { if (targetStorageTypesBuilder_ == null) { return targetStorageTypes_.get(index); } else { return targetStorageTypesBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public java.util.List getTargetStorageTypesOrBuilderList() { if (targetStorageTypesBuilder_ != null) { return targetStorageTypesBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(targetStorageTypes_); } } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder addTargetStorageTypesBuilder() { return getTargetStorageTypesFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder addTargetStorageTypesBuilder( int index) { return getTargetStorageTypesFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.StorageTypesProto targetStorageTypes = 6; */ public java.util.List getTargetStorageTypesBuilderList() { return getTargetStorageTypesFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProtoOrBuilder> getTargetStorageTypesFieldBuilder() { if (targetStorageTypesBuilder_ == null) { targetStorageTypesBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageTypesProtoOrBuilder>( targetStorageTypes_, ((bitField0_ & 0x00000020) != 0), getParentForChildren(), isClean()); targetStorageTypes_ = null; } return targetStorageTypesBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockIdCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockIdCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ boolean hasAction(); /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action getAction(); /** * required string blockPoolId = 2; */ boolean hasBlockPoolId(); /** * required string blockPoolId = 2; */ java.lang.String getBlockPoolId(); /** * required string blockPoolId = 2; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes(); /** * repeated uint64 blockIds = 3 [packed = true]; */ java.util.List getBlockIdsList(); /** * repeated uint64 blockIds = 3 [packed = true]; */ int getBlockIdsCount(); /** * repeated uint64 blockIds = 3 [packed = true]; */ long getBlockIds(int index); } /** *
   **
   * Command to instruct datanodes to perform certain action
   * on the given set of block IDs.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockIdCommandProto} */ public static final class BlockIdCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockIdCommandProto) BlockIdCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockIdCommandProto.newBuilder() to construct. private BlockIdCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockIdCommandProto() { action_ = 1; blockPoolId_ = ""; blockIds_ = emptyLongList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockIdCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action value = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; action_ = rawValue; } break; } case 18: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; blockPoolId_ = bs; break; } case 24: { if (!((mutable_bitField0_ & 0x00000004) != 0)) { blockIds_ = newLongList(); mutable_bitField0_ |= 0x00000004; } blockIds_.addLong(input.readUInt64()); break; } case 26: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000004) != 0) && input.getBytesUntilLimit() > 0) { blockIds_ = newLongList(); mutable_bitField0_ |= 0x00000004; } while (input.getBytesUntilLimit() > 0) { blockIds_.addLong(input.readUInt64()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { blockIds_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder.class); } /** * Protobuf enum {@code hadoop.hdfs.datanode.BlockIdCommandProto.Action} */ public enum Action implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum { /** * CACHE = 1; */ CACHE(1), /** * UNCACHE = 2; */ UNCACHE(2), ; /** * CACHE = 1; */ public static final int CACHE_VALUE = 1; /** * UNCACHE = 2; */ public static final int UNCACHE_VALUE = 2; public final int getNumber() { return value; } /** * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static Action valueOf(int value) { return forNumber(value); } public static Action forNumber(int value) { switch (value) { case 1: return CACHE; case 2: return UNCACHE; default: return null; } } public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap< Action> internalValueMap = new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() { public Action findValueByNumber(int number) { return Action.forNumber(number); } }; public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDescriptor().getEnumTypes().get(0); } private static final Action[] VALUES = values(); public static Action valueOf( org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private Action(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hadoop.hdfs.datanode.BlockIdCommandProto.Action) } private int bitField0_; public static final int ACTION_FIELD_NUMBER = 1; private int action_; /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ public boolean hasAction() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action getAction() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action.valueOf(action_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action.CACHE : result; } public static final int BLOCKPOOLID_FIELD_NUMBER = 2; private volatile java.lang.Object blockPoolId_; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int BLOCKIDS_FIELD_NUMBER = 3; private org.apache.hadoop.thirdparty.protobuf.Internal.LongList blockIds_; /** * repeated uint64 blockIds = 3 [packed = true]; */ public java.util.List getBlockIdsList() { return blockIds_; } /** * repeated uint64 blockIds = 3 [packed = true]; */ public int getBlockIdsCount() { return blockIds_.size(); } /** * repeated uint64 blockIds = 3 [packed = true]; */ public long getBlockIds(int index) { return blockIds_.getLong(index); } private int blockIdsMemoizedSerializedSize = -1; private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasAction()) { memoizedIsInitialized = 0; return false; } if (!hasBlockPoolId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) != 0)) { output.writeEnum(1, action_); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, blockPoolId_); } if (getBlockIdsList().size() > 0) { output.writeUInt32NoTag(26); output.writeUInt32NoTag(blockIdsMemoizedSerializedSize); } for (int i = 0; i < blockIds_.size(); i++) { output.writeUInt64NoTag(blockIds_.getLong(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeEnumSize(1, action_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, blockPoolId_); } { int dataSize = 0; for (int i = 0; i < blockIds_.size(); i++) { dataSize += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64SizeNoTag(blockIds_.getLong(i)); } size += dataSize; if (!getBlockIdsList().isEmpty()) { size += 1; size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeInt32SizeNoTag(dataSize); } blockIdsMemoizedSerializedSize = dataSize; } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto) obj; if (hasAction() != other.hasAction()) return false; if (hasAction()) { if (action_ != other.action_) return false; } if (hasBlockPoolId() != other.hasBlockPoolId()) return false; if (hasBlockPoolId()) { if (!getBlockPoolId() .equals(other.getBlockPoolId())) return false; } if (!getBlockIdsList() .equals(other.getBlockIdsList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasAction()) { hash = (37 * hash) + ACTION_FIELD_NUMBER; hash = (53 * hash) + action_; } if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } if (getBlockIdsCount() > 0) { hash = (37 * hash) + BLOCKIDS_FIELD_NUMBER; hash = (53 * hash) + getBlockIdsList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Command to instruct datanodes to perform certain action
     * on the given set of block IDs.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockIdCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockIdCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); action_ = 1; bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); blockIds_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.action_ = action_; if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.blockPoolId_ = blockPoolId_; if (((bitField0_ & 0x00000004) != 0)) { blockIds_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000004); } result.blockIds_ = blockIds_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.getDefaultInstance()) return this; if (other.hasAction()) { setAction(other.getAction()); } if (other.hasBlockPoolId()) { bitField0_ |= 0x00000002; blockPoolId_ = other.blockPoolId_; onChanged(); } if (!other.blockIds_.isEmpty()) { if (blockIds_.isEmpty()) { blockIds_ = other.blockIds_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureBlockIdsIsMutable(); blockIds_.addAll(other.blockIds_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasAction()) { return false; } if (!hasBlockPoolId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int action_ = 1; /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ public boolean hasAction() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action getAction() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action.valueOf(action_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action.CACHE : result; } /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ public Builder setAction(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto.Action value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; action_ = value.getNumber(); onChanged(); return this; } /** * required .hadoop.hdfs.datanode.BlockIdCommandProto.Action action = 1; */ public Builder clearAction() { bitField0_ = (bitField0_ & ~0x00000001); action_ = 1; onChanged(); return this; } private java.lang.Object blockPoolId_ = ""; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * required string blockPoolId = 2; */ public Builder setBlockPoolId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000002); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder setBlockPoolIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } private org.apache.hadoop.thirdparty.protobuf.Internal.LongList blockIds_ = emptyLongList(); private void ensureBlockIdsIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { blockIds_ = mutableCopy(blockIds_); bitField0_ |= 0x00000004; } } /** * repeated uint64 blockIds = 3 [packed = true]; */ public java.util.List getBlockIdsList() { return ((bitField0_ & 0x00000004) != 0) ? java.util.Collections.unmodifiableList(blockIds_) : blockIds_; } /** * repeated uint64 blockIds = 3 [packed = true]; */ public int getBlockIdsCount() { return blockIds_.size(); } /** * repeated uint64 blockIds = 3 [packed = true]; */ public long getBlockIds(int index) { return blockIds_.getLong(index); } /** * repeated uint64 blockIds = 3 [packed = true]; */ public Builder setBlockIds( int index, long value) { ensureBlockIdsIsMutable(); blockIds_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 blockIds = 3 [packed = true]; */ public Builder addBlockIds(long value) { ensureBlockIdsIsMutable(); blockIds_.addLong(value); onChanged(); return this; } /** * repeated uint64 blockIds = 3 [packed = true]; */ public Builder addAllBlockIds( java.lang.Iterable values) { ensureBlockIdsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blockIds_); onChanged(); return this; } /** * repeated uint64 blockIds = 3 [packed = true]; */ public Builder clearBlockIds() { blockIds_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockIdCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockIdCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockIdCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockIdCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockIdCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockRecoveryCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockRecoveryCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ java.util.List getBlocksList(); /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto getBlocks(int index); /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ int getBlocksCount(); /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ java.util.List getBlocksOrBuilderList(); /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProtoOrBuilder getBlocksOrBuilder( int index); } /** *
   **
   * List of blocks to be recovered by the datanode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockRecoveryCommandProto} */ public static final class BlockRecoveryCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockRecoveryCommandProto) BlockRecoveryCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockRecoveryCommandProto.newBuilder() to construct. private BlockRecoveryCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockRecoveryCommandProto() { blocks_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockRecoveryCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { blocks_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } blocks_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder.class); } public static final int BLOCKS_FIELD_NUMBER = 1; private java.util.List blocks_; /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public java.util.List getBlocksOrBuilderList() { return blocks_; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto getBlocks(int index) { return blocks_.get(index); } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProtoOrBuilder getBlocksOrBuilder( int index) { return blocks_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < blocks_.size(); i++) { output.writeMessage(1, blocks_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < blocks_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, blocks_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto) obj; if (!getBlocksList() .equals(other.getBlocksList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * List of blocks to be recovered by the datanode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockRecoveryCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockRecoveryCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlocksFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { blocksBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto(this); int from_bitField0_ = bitField0_; if (blocksBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); bitField0_ = (bitField0_ & ~0x00000001); } result.blocks_ = blocks_; } else { result.blocks_ = blocksBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto.getDefaultInstance()) return this; if (blocksBuilder_ == null) { if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } } else { if (!other.blocks_.isEmpty()) { if (blocksBuilder_.isEmpty()) { blocksBuilder_.dispose(); blocksBuilder_ = null; blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000001); blocksBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBlocksFieldBuilder() : null; } else { blocksBuilder_.addAllMessages(other.blocks_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List blocks_ = java.util.Collections.emptyList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { blocks_ = new java.util.ArrayList(blocks_); bitField0_ |= 0x00000001; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProtoOrBuilder> blocksBuilder_; /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public java.util.List getBlocksList() { if (blocksBuilder_ == null) { return java.util.Collections.unmodifiableList(blocks_); } else { return blocksBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public int getBlocksCount() { if (blocksBuilder_ == null) { return blocks_.size(); } else { return blocksBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto getBlocks(int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.set(index, value); onChanged(); } else { blocksBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.set(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(value); onChanged(); } else { blocksBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(index, value); onChanged(); } else { blocksBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder addBlocks( org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder addAllBlocks( java.lang.Iterable values) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); } else { blocksBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder clearBlocks() { if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { blocksBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public Builder removeBlocks(int index) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.remove(index); onChanged(); } else { blocksBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder getBlocksBuilder( int index) { return getBlocksFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProtoOrBuilder getBlocksOrBuilder( int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public java.util.List getBlocksOrBuilderList() { if (blocksBuilder_ != null) { return blocksBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(blocks_); } } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.RecoveringBlockProto blocks = 1; */ public java.util.List getBlocksBuilderList() { return getBlocksFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProtoOrBuilder> getBlocksFieldBuilder() { if (blocksBuilder_ == null) { blocksBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.RecoveringBlockProtoOrBuilder>( blocks_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); blocks_ = null; } return blocksBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockRecoveryCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockRecoveryCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockRecoveryCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockRecoveryCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockRecoveryCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface FinalizeCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.FinalizeCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Block pool to be finalized
     * 
* * required string blockPoolId = 1; */ boolean hasBlockPoolId(); /** *
     * Block pool to be finalized
     * 
* * required string blockPoolId = 1; */ java.lang.String getBlockPoolId(); /** *
     * Block pool to be finalized
     * 
* * required string blockPoolId = 1; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes(); } /** *
   **
   * Finalize the upgrade at the datanode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.FinalizeCommandProto} */ public static final class FinalizeCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.FinalizeCommandProto) FinalizeCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use FinalizeCommandProto.newBuilder() to construct. private FinalizeCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private FinalizeCommandProto() { blockPoolId_ = ""; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private FinalizeCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; blockPoolId_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder.class); } private int bitField0_; public static final int BLOCKPOOLID_FIELD_NUMBER = 1; private volatile java.lang.Object blockPoolId_; /** *
     * Block pool to be finalized
     * 
* * required string blockPoolId = 1; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Block pool to be finalized
     * 
* * required string blockPoolId = 1; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } } /** *
     * Block pool to be finalized
     * 
* * required string blockPoolId = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasBlockPoolId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, blockPoolId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, blockPoolId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto) obj; if (hasBlockPoolId() != other.hasBlockPoolId()) return false; if (hasBlockPoolId()) { if (!getBlockPoolId() .equals(other.getBlockPoolId())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Finalize the upgrade at the datanode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.FinalizeCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.FinalizeCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); blockPoolId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.blockPoolId_ = blockPoolId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto.getDefaultInstance()) return this; if (other.hasBlockPoolId()) { bitField0_ |= 0x00000001; blockPoolId_ = other.blockPoolId_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasBlockPoolId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object blockPoolId_ = ""; /** *
       * Block pool to be finalized
       * 
* * required string blockPoolId = 1; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Block pool to be finalized
       * 
* * required string blockPoolId = 1; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } else { return (java.lang.String) ref; } } /** *
       * Block pool to be finalized
       * 
* * required string blockPoolId = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** *
       * Block pool to be finalized
       * 
* * required string blockPoolId = 1; */ public Builder setBlockPoolId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; blockPoolId_ = value; onChanged(); return this; } /** *
       * Block pool to be finalized
       * 
* * required string blockPoolId = 1; */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /** *
       * Block pool to be finalized
       * 
* * required string blockPoolId = 1; */ public Builder setBlockPoolIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; blockPoolId_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.FinalizeCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.FinalizeCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public FinalizeCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new FinalizeCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.FinalizeCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface KeyUpdateCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.KeyUpdateCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ boolean hasKeys(); /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto getKeys(); /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder getKeysOrBuilder(); } /** *
   **
   * Update the block keys at the datanode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.KeyUpdateCommandProto} */ public static final class KeyUpdateCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.KeyUpdateCommandProto) KeyUpdateCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use KeyUpdateCommandProto.newBuilder() to construct. private KeyUpdateCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private KeyUpdateCommandProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private KeyUpdateCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = keys_.toBuilder(); } keys_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(keys_); keys_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder.class); } private int bitField0_; public static final int KEYS_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto keys_; /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public boolean hasKeys() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto getKeys() { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder getKeysOrBuilder() { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasKeys()) { memoizedIsInitialized = 0; return false; } if (!getKeys().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getKeys()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getKeys()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto) obj; if (hasKeys() != other.hasKeys()) return false; if (hasKeys()) { if (!getKeys() .equals(other.getKeys())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasKeys()) { hash = (37 * hash) + KEYS_FIELD_NUMBER; hash = (53 * hash) + getKeys().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Update the block keys at the datanode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.KeyUpdateCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.KeyUpdateCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getKeysFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (keysBuilder_ == null) { keys_ = null; } else { keysBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (keysBuilder_ == null) { result.keys_ = keys_; } else { result.keys_ = keysBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto.getDefaultInstance()) return this; if (other.hasKeys()) { mergeKeys(other.getKeys()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasKeys()) { return false; } if (!getKeys().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto keys_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder> keysBuilder_; /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public boolean hasKeys() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto getKeys() { if (keysBuilder_ == null) { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } else { return keysBuilder_.getMessage(); } } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public Builder setKeys(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto value) { if (keysBuilder_ == null) { if (value == null) { throw new NullPointerException(); } keys_ = value; onChanged(); } else { keysBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public Builder setKeys( org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder builderForValue) { if (keysBuilder_ == null) { keys_ = builderForValue.build(); onChanged(); } else { keysBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public Builder mergeKeys(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto value) { if (keysBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && keys_ != null && keys_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance()) { keys_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.newBuilder(keys_).mergeFrom(value).buildPartial(); } else { keys_ = value; } onChanged(); } else { keysBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public Builder clearKeys() { if (keysBuilder_ == null) { keys_ = null; onChanged(); } else { keysBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder getKeysBuilder() { bitField0_ |= 0x00000001; onChanged(); return getKeysFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder getKeysOrBuilder() { if (keysBuilder_ != null) { return keysBuilder_.getMessageOrBuilder(); } else { return keys_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.getDefaultInstance() : keys_; } } /** * required .hadoop.hdfs.ExportedBlockKeysProto keys = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder> getKeysFieldBuilder() { if (keysBuilder_ == null) { keysBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.ExportedBlockKeysProtoOrBuilder>( getKeys(), getParentForChildren(), isClean()); keys_ = null; } return keysBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.KeyUpdateCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.KeyUpdateCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public KeyUpdateCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new KeyUpdateCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.KeyUpdateCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RegisterCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.RegisterCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { } /** *
   **
   * Instruct datanode to register with the namenode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.RegisterCommandProto} */ public static final class RegisterCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.RegisterCommandProto) RegisterCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use RegisterCommandProto.newBuilder() to construct. private RegisterCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RegisterCommandProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RegisterCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Instruct datanode to register with the namenode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.RegisterCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.RegisterCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.RegisterCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.RegisterCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public RegisterCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new RegisterCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockECReconstructionCommandProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockECReconstructionCommandProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ java.util.List getBlockECReconstructioninfoList(); /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto getBlockECReconstructioninfo(int index); /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ int getBlockECReconstructioninfoCount(); /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ java.util.List getBlockECReconstructioninfoOrBuilderList(); /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProtoOrBuilder getBlockECReconstructioninfoOrBuilder( int index); } /** *
   **
   * Block Erasure coding reconstruction command
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockECReconstructionCommandProto} */ public static final class BlockECReconstructionCommandProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockECReconstructionCommandProto) BlockECReconstructionCommandProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockECReconstructionCommandProto.newBuilder() to construct. private BlockECReconstructionCommandProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockECReconstructionCommandProto() { blockECReconstructioninfo_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockECReconstructionCommandProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { blockECReconstructioninfo_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } blockECReconstructioninfo_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { blockECReconstructioninfo_ = java.util.Collections.unmodifiableList(blockECReconstructioninfo_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder.class); } public static final int BLOCKECRECONSTRUCTIONINFO_FIELD_NUMBER = 1; private java.util.List blockECReconstructioninfo_; /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public java.util.List getBlockECReconstructioninfoList() { return blockECReconstructioninfo_; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public java.util.List getBlockECReconstructioninfoOrBuilderList() { return blockECReconstructioninfo_; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public int getBlockECReconstructioninfoCount() { return blockECReconstructioninfo_.size(); } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto getBlockECReconstructioninfo(int index) { return blockECReconstructioninfo_.get(index); } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProtoOrBuilder getBlockECReconstructioninfoOrBuilder( int index) { return blockECReconstructioninfo_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getBlockECReconstructioninfoCount(); i++) { if (!getBlockECReconstructioninfo(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < blockECReconstructioninfo_.size(); i++) { output.writeMessage(1, blockECReconstructioninfo_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < blockECReconstructioninfo_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, blockECReconstructioninfo_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto) obj; if (!getBlockECReconstructioninfoList() .equals(other.getBlockECReconstructioninfoList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getBlockECReconstructioninfoCount() > 0) { hash = (37 * hash) + BLOCKECRECONSTRUCTIONINFO_FIELD_NUMBER; hash = (53 * hash) + getBlockECReconstructioninfoList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Block Erasure coding reconstruction command
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockECReconstructionCommandProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockECReconstructionCommandProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlockECReconstructioninfoFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (blockECReconstructioninfoBuilder_ == null) { blockECReconstructioninfo_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { blockECReconstructioninfoBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto(this); int from_bitField0_ = bitField0_; if (blockECReconstructioninfoBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { blockECReconstructioninfo_ = java.util.Collections.unmodifiableList(blockECReconstructioninfo_); bitField0_ = (bitField0_ & ~0x00000001); } result.blockECReconstructioninfo_ = blockECReconstructioninfo_; } else { result.blockECReconstructioninfo_ = blockECReconstructioninfoBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto.getDefaultInstance()) return this; if (blockECReconstructioninfoBuilder_ == null) { if (!other.blockECReconstructioninfo_.isEmpty()) { if (blockECReconstructioninfo_.isEmpty()) { blockECReconstructioninfo_ = other.blockECReconstructioninfo_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.addAll(other.blockECReconstructioninfo_); } onChanged(); } } else { if (!other.blockECReconstructioninfo_.isEmpty()) { if (blockECReconstructioninfoBuilder_.isEmpty()) { blockECReconstructioninfoBuilder_.dispose(); blockECReconstructioninfoBuilder_ = null; blockECReconstructioninfo_ = other.blockECReconstructioninfo_; bitField0_ = (bitField0_ & ~0x00000001); blockECReconstructioninfoBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBlockECReconstructioninfoFieldBuilder() : null; } else { blockECReconstructioninfoBuilder_.addAllMessages(other.blockECReconstructioninfo_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getBlockECReconstructioninfoCount(); i++) { if (!getBlockECReconstructioninfo(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List blockECReconstructioninfo_ = java.util.Collections.emptyList(); private void ensureBlockECReconstructioninfoIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { blockECReconstructioninfo_ = new java.util.ArrayList(blockECReconstructioninfo_); bitField0_ |= 0x00000001; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProtoOrBuilder> blockECReconstructioninfoBuilder_; /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public java.util.List getBlockECReconstructioninfoList() { if (blockECReconstructioninfoBuilder_ == null) { return java.util.Collections.unmodifiableList(blockECReconstructioninfo_); } else { return blockECReconstructioninfoBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public int getBlockECReconstructioninfoCount() { if (blockECReconstructioninfoBuilder_ == null) { return blockECReconstructioninfo_.size(); } else { return blockECReconstructioninfoBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto getBlockECReconstructioninfo(int index) { if (blockECReconstructioninfoBuilder_ == null) { return blockECReconstructioninfo_.get(index); } else { return blockECReconstructioninfoBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder setBlockECReconstructioninfo( int index, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto value) { if (blockECReconstructioninfoBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.set(index, value); onChanged(); } else { blockECReconstructioninfoBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder setBlockECReconstructioninfo( int index, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder builderForValue) { if (blockECReconstructioninfoBuilder_ == null) { ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.set(index, builderForValue.build()); onChanged(); } else { blockECReconstructioninfoBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder addBlockECReconstructioninfo(org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto value) { if (blockECReconstructioninfoBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.add(value); onChanged(); } else { blockECReconstructioninfoBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder addBlockECReconstructioninfo( int index, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto value) { if (blockECReconstructioninfoBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.add(index, value); onChanged(); } else { blockECReconstructioninfoBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder addBlockECReconstructioninfo( org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder builderForValue) { if (blockECReconstructioninfoBuilder_ == null) { ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.add(builderForValue.build()); onChanged(); } else { blockECReconstructioninfoBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder addBlockECReconstructioninfo( int index, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder builderForValue) { if (blockECReconstructioninfoBuilder_ == null) { ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.add(index, builderForValue.build()); onChanged(); } else { blockECReconstructioninfoBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder addAllBlockECReconstructioninfo( java.lang.Iterable values) { if (blockECReconstructioninfoBuilder_ == null) { ensureBlockECReconstructioninfoIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blockECReconstructioninfo_); onChanged(); } else { blockECReconstructioninfoBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder clearBlockECReconstructioninfo() { if (blockECReconstructioninfoBuilder_ == null) { blockECReconstructioninfo_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { blockECReconstructioninfoBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public Builder removeBlockECReconstructioninfo(int index) { if (blockECReconstructioninfoBuilder_ == null) { ensureBlockECReconstructioninfoIsMutable(); blockECReconstructioninfo_.remove(index); onChanged(); } else { blockECReconstructioninfoBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder getBlockECReconstructioninfoBuilder( int index) { return getBlockECReconstructioninfoFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProtoOrBuilder getBlockECReconstructioninfoOrBuilder( int index) { if (blockECReconstructioninfoBuilder_ == null) { return blockECReconstructioninfo_.get(index); } else { return blockECReconstructioninfoBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public java.util.List getBlockECReconstructioninfoOrBuilderList() { if (blockECReconstructioninfoBuilder_ != null) { return blockECReconstructioninfoBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(blockECReconstructioninfo_); } } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder addBlockECReconstructioninfoBuilder() { return getBlockECReconstructioninfoFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder addBlockECReconstructioninfoBuilder( int index) { return getBlockECReconstructioninfoFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.BlockECReconstructionInfoProto blockECReconstructioninfo = 1; */ public java.util.List getBlockECReconstructioninfoBuilderList() { return getBlockECReconstructioninfoFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProtoOrBuilder> getBlockECReconstructioninfoFieldBuilder() { if (blockECReconstructioninfoBuilder_ == null) { blockECReconstructioninfoBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.BlockECReconstructionInfoProtoOrBuilder>( blockECReconstructioninfo_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); blockECReconstructioninfo_ = null; } return blockECReconstructioninfoBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockECReconstructionCommandProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockECReconstructionCommandProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockECReconstructionCommandProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockECReconstructionCommandProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockECReconstructionCommandProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RegisterDatanodeRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.RegisterDatanodeRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ boolean hasRegistration(); /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration(); /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder(); } /** *
   **
   * registration - Information of the datanode registering with the namenode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.RegisterDatanodeRequestProto} */ public static final class RegisterDatanodeRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.RegisterDatanodeRequestProto) RegisterDatanodeRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use RegisterDatanodeRequestProto.newBuilder() to construct. private RegisterDatanodeRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RegisterDatanodeRequestProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RegisterDatanodeRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registration_.toBuilder(); } registration_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registration_); registration_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.Builder.class); } private int bitField0_; public static final int REGISTRATION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistration()) { memoizedIsInitialized = 0; return false; } if (!getRegistration().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistration()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistration()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto) obj; if (hasRegistration() != other.hasRegistration()) return false; if (hasRegistration()) { if (!getRegistration() .equals(other.getRegistration())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistration()) { hash = (37 * hash) + REGISTRATION_FIELD_NUMBER; hash = (53 * hash) + getRegistration().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registration - Information of the datanode registering with the namenode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.RegisterDatanodeRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.RegisterDatanodeRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistrationFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registrationBuilder_ == null) { registration_ = null; } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registrationBuilder_ == null) { result.registration_ = registration_; } else { result.registration_ = registrationBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.getDefaultInstance()) return this; if (other.hasRegistration()) { mergeRegistration(other.getRegistration()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistration()) { return false; } if (!getRegistration().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registrationBuilder_; /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { if (registrationBuilder_ == null) { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } else { return registrationBuilder_.getMessage(); } } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registration_ = value; onChanged(); } else { registrationBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registrationBuilder_ == null) { registration_ = builderForValue.build(); onChanged(); } else { registrationBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder mergeRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registration_ != null && registration_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registration_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registration_).mergeFrom(value).buildPartial(); } else { registration_ = value; } onChanged(); } else { registrationBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder clearRegistration() { if (registrationBuilder_ == null) { registration_ = null; onChanged(); } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistrationBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistrationFieldBuilder().getBuilder(); } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { if (registrationBuilder_ != null) { return registrationBuilder_.getMessageOrBuilder(); } else { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistrationFieldBuilder() { if (registrationBuilder_ == null) { registrationBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistration(), getParentForChildren(), isClean()); registration_ = null; } return registrationBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.RegisterDatanodeRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.RegisterDatanodeRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public RegisterDatanodeRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new RegisterDatanodeRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RegisterDatanodeResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.RegisterDatanodeResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ boolean hasRegistration(); /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration(); /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder(); } /** *
   **
   * registration - Update registration of the datanode that successfully 
   *                registered. StorageInfo will be updated to include new 
   *                storage ID if the datanode did not have one in the request.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.RegisterDatanodeResponseProto} */ public static final class RegisterDatanodeResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.RegisterDatanodeResponseProto) RegisterDatanodeResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use RegisterDatanodeResponseProto.newBuilder() to construct. private RegisterDatanodeResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RegisterDatanodeResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RegisterDatanodeResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registration_.toBuilder(); } registration_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registration_); registration_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.Builder.class); } private int bitField0_; public static final int REGISTRATION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistration()) { memoizedIsInitialized = 0; return false; } if (!getRegistration().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistration()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistration()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto) obj; if (hasRegistration() != other.hasRegistration()) return false; if (hasRegistration()) { if (!getRegistration() .equals(other.getRegistration())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistration()) { hash = (37 * hash) + REGISTRATION_FIELD_NUMBER; hash = (53 * hash) + getRegistration().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registration - Update registration of the datanode that successfully 
     *                registered. StorageInfo will be updated to include new 
     *                storage ID if the datanode did not have one in the request.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.RegisterDatanodeResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.RegisterDatanodeResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistrationFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registrationBuilder_ == null) { registration_ = null; } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registrationBuilder_ == null) { result.registration_ = registration_; } else { result.registration_ = registrationBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance()) return this; if (other.hasRegistration()) { mergeRegistration(other.getRegistration()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistration()) { return false; } if (!getRegistration().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registrationBuilder_; /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { if (registrationBuilder_ == null) { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } else { return registrationBuilder_.getMessage(); } } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registration_ = value; onChanged(); } else { registrationBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registrationBuilder_ == null) { registration_ = builderForValue.build(); onChanged(); } else { registrationBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder mergeRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registration_ != null && registration_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registration_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registration_).mergeFrom(value).buildPartial(); } else { registration_ = value; } onChanged(); } else { registrationBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder clearRegistration() { if (registrationBuilder_ == null) { registration_ = null; onChanged(); } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistrationBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistrationFieldBuilder().getBuilder(); } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { if (registrationBuilder_ != null) { return registrationBuilder_.getMessageOrBuilder(); } else { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistrationFieldBuilder() { if (registrationBuilder_ == null) { registrationBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistration(), getParentForChildren(), isClean()); registration_ = null; } return registrationBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.RegisterDatanodeResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.RegisterDatanodeResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public RegisterDatanodeResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new RegisterDatanodeResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface VolumeFailureSummaryProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.VolumeFailureSummaryProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * repeated string failedStorageLocations = 1; */ java.util.List getFailedStorageLocationsList(); /** * repeated string failedStorageLocations = 1; */ int getFailedStorageLocationsCount(); /** * repeated string failedStorageLocations = 1; */ java.lang.String getFailedStorageLocations(int index); /** * repeated string failedStorageLocations = 1; */ org.apache.hadoop.thirdparty.protobuf.ByteString getFailedStorageLocationsBytes(int index); /** * required uint64 lastVolumeFailureDate = 2; */ boolean hasLastVolumeFailureDate(); /** * required uint64 lastVolumeFailureDate = 2; */ long getLastVolumeFailureDate(); /** * required uint64 estimatedCapacityLostTotal = 3; */ boolean hasEstimatedCapacityLostTotal(); /** * required uint64 estimatedCapacityLostTotal = 3; */ long getEstimatedCapacityLostTotal(); } /** *
   **
   * failedStorageLocations - storage locations that have failed
   * lastVolumeFailureDate - date/time of last volume failure
   * estimatedCapacityLost - estimate of total capacity lost due to volume failures
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.VolumeFailureSummaryProto} */ public static final class VolumeFailureSummaryProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.VolumeFailureSummaryProto) VolumeFailureSummaryProtoOrBuilder { private static final long serialVersionUID = 0L; // Use VolumeFailureSummaryProto.newBuilder() to construct. private VolumeFailureSummaryProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private VolumeFailureSummaryProto() { failedStorageLocations_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private VolumeFailureSummaryProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); if (!((mutable_bitField0_ & 0x00000001) != 0)) { failedStorageLocations_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000001; } failedStorageLocations_.add(bs); break; } case 16: { bitField0_ |= 0x00000001; lastVolumeFailureDate_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000002; estimatedCapacityLostTotal_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { failedStorageLocations_ = failedStorageLocations_.getUnmodifiableView(); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder.class); } private int bitField0_; public static final int FAILEDSTORAGELOCATIONS_FIELD_NUMBER = 1; private org.apache.hadoop.thirdparty.protobuf.LazyStringList failedStorageLocations_; /** * repeated string failedStorageLocations = 1; */ public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList getFailedStorageLocationsList() { return failedStorageLocations_; } /** * repeated string failedStorageLocations = 1; */ public int getFailedStorageLocationsCount() { return failedStorageLocations_.size(); } /** * repeated string failedStorageLocations = 1; */ public java.lang.String getFailedStorageLocations(int index) { return failedStorageLocations_.get(index); } /** * repeated string failedStorageLocations = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getFailedStorageLocationsBytes(int index) { return failedStorageLocations_.getByteString(index); } public static final int LASTVOLUMEFAILUREDATE_FIELD_NUMBER = 2; private long lastVolumeFailureDate_; /** * required uint64 lastVolumeFailureDate = 2; */ public boolean hasLastVolumeFailureDate() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 lastVolumeFailureDate = 2; */ public long getLastVolumeFailureDate() { return lastVolumeFailureDate_; } public static final int ESTIMATEDCAPACITYLOSTTOTAL_FIELD_NUMBER = 3; private long estimatedCapacityLostTotal_; /** * required uint64 estimatedCapacityLostTotal = 3; */ public boolean hasEstimatedCapacityLostTotal() { return ((bitField0_ & 0x00000002) != 0); } /** * required uint64 estimatedCapacityLostTotal = 3; */ public long getEstimatedCapacityLostTotal() { return estimatedCapacityLostTotal_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasLastVolumeFailureDate()) { memoizedIsInitialized = 0; return false; } if (!hasEstimatedCapacityLostTotal()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < failedStorageLocations_.size(); i++) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, failedStorageLocations_.getRaw(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(2, lastVolumeFailureDate_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(3, estimatedCapacityLostTotal_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < failedStorageLocations_.size(); i++) { dataSize += computeStringSizeNoTag(failedStorageLocations_.getRaw(i)); } size += dataSize; size += 1 * getFailedStorageLocationsList().size(); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(2, lastVolumeFailureDate_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(3, estimatedCapacityLostTotal_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto) obj; if (!getFailedStorageLocationsList() .equals(other.getFailedStorageLocationsList())) return false; if (hasLastVolumeFailureDate() != other.hasLastVolumeFailureDate()) return false; if (hasLastVolumeFailureDate()) { if (getLastVolumeFailureDate() != other.getLastVolumeFailureDate()) return false; } if (hasEstimatedCapacityLostTotal() != other.hasEstimatedCapacityLostTotal()) return false; if (hasEstimatedCapacityLostTotal()) { if (getEstimatedCapacityLostTotal() != other.getEstimatedCapacityLostTotal()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getFailedStorageLocationsCount() > 0) { hash = (37 * hash) + FAILEDSTORAGELOCATIONS_FIELD_NUMBER; hash = (53 * hash) + getFailedStorageLocationsList().hashCode(); } if (hasLastVolumeFailureDate()) { hash = (37 * hash) + LASTVOLUMEFAILUREDATE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getLastVolumeFailureDate()); } if (hasEstimatedCapacityLostTotal()) { hash = (37 * hash) + ESTIMATEDCAPACITYLOSTTOTAL_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getEstimatedCapacityLostTotal()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * failedStorageLocations - storage locations that have failed
     * lastVolumeFailureDate - date/time of last volume failure
     * estimatedCapacityLost - estimate of total capacity lost due to volume failures
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.VolumeFailureSummaryProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.VolumeFailureSummaryProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); failedStorageLocations_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); lastVolumeFailureDate_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); estimatedCapacityLostTotal_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((bitField0_ & 0x00000001) != 0)) { failedStorageLocations_ = failedStorageLocations_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000001); } result.failedStorageLocations_ = failedStorageLocations_; if (((from_bitField0_ & 0x00000002) != 0)) { result.lastVolumeFailureDate_ = lastVolumeFailureDate_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000004) != 0)) { result.estimatedCapacityLostTotal_ = estimatedCapacityLostTotal_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance()) return this; if (!other.failedStorageLocations_.isEmpty()) { if (failedStorageLocations_.isEmpty()) { failedStorageLocations_ = other.failedStorageLocations_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureFailedStorageLocationsIsMutable(); failedStorageLocations_.addAll(other.failedStorageLocations_); } onChanged(); } if (other.hasLastVolumeFailureDate()) { setLastVolumeFailureDate(other.getLastVolumeFailureDate()); } if (other.hasEstimatedCapacityLostTotal()) { setEstimatedCapacityLostTotal(other.getEstimatedCapacityLostTotal()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasLastVolumeFailureDate()) { return false; } if (!hasEstimatedCapacityLostTotal()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.thirdparty.protobuf.LazyStringList failedStorageLocations_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; private void ensureFailedStorageLocationsIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { failedStorageLocations_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(failedStorageLocations_); bitField0_ |= 0x00000001; } } /** * repeated string failedStorageLocations = 1; */ public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList getFailedStorageLocationsList() { return failedStorageLocations_.getUnmodifiableView(); } /** * repeated string failedStorageLocations = 1; */ public int getFailedStorageLocationsCount() { return failedStorageLocations_.size(); } /** * repeated string failedStorageLocations = 1; */ public java.lang.String getFailedStorageLocations(int index) { return failedStorageLocations_.get(index); } /** * repeated string failedStorageLocations = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getFailedStorageLocationsBytes(int index) { return failedStorageLocations_.getByteString(index); } /** * repeated string failedStorageLocations = 1; */ public Builder setFailedStorageLocations( int index, java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureFailedStorageLocationsIsMutable(); failedStorageLocations_.set(index, value); onChanged(); return this; } /** * repeated string failedStorageLocations = 1; */ public Builder addFailedStorageLocations( java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureFailedStorageLocationsIsMutable(); failedStorageLocations_.add(value); onChanged(); return this; } /** * repeated string failedStorageLocations = 1; */ public Builder addAllFailedStorageLocations( java.lang.Iterable values) { ensureFailedStorageLocationsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, failedStorageLocations_); onChanged(); return this; } /** * repeated string failedStorageLocations = 1; */ public Builder clearFailedStorageLocations() { failedStorageLocations_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } /** * repeated string failedStorageLocations = 1; */ public Builder addFailedStorageLocationsBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureFailedStorageLocationsIsMutable(); failedStorageLocations_.add(value); onChanged(); return this; } private long lastVolumeFailureDate_ ; /** * required uint64 lastVolumeFailureDate = 2; */ public boolean hasLastVolumeFailureDate() { return ((bitField0_ & 0x00000002) != 0); } /** * required uint64 lastVolumeFailureDate = 2; */ public long getLastVolumeFailureDate() { return lastVolumeFailureDate_; } /** * required uint64 lastVolumeFailureDate = 2; */ public Builder setLastVolumeFailureDate(long value) { bitField0_ |= 0x00000002; lastVolumeFailureDate_ = value; onChanged(); return this; } /** * required uint64 lastVolumeFailureDate = 2; */ public Builder clearLastVolumeFailureDate() { bitField0_ = (bitField0_ & ~0x00000002); lastVolumeFailureDate_ = 0L; onChanged(); return this; } private long estimatedCapacityLostTotal_ ; /** * required uint64 estimatedCapacityLostTotal = 3; */ public boolean hasEstimatedCapacityLostTotal() { return ((bitField0_ & 0x00000004) != 0); } /** * required uint64 estimatedCapacityLostTotal = 3; */ public long getEstimatedCapacityLostTotal() { return estimatedCapacityLostTotal_; } /** * required uint64 estimatedCapacityLostTotal = 3; */ public Builder setEstimatedCapacityLostTotal(long value) { bitField0_ |= 0x00000004; estimatedCapacityLostTotal_ = value; onChanged(); return this; } /** * required uint64 estimatedCapacityLostTotal = 3; */ public Builder clearEstimatedCapacityLostTotal() { bitField0_ = (bitField0_ & ~0x00000004); estimatedCapacityLostTotal_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.VolumeFailureSummaryProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.VolumeFailureSummaryProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public VolumeFailureSummaryProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new VolumeFailureSummaryProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface HeartbeatRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.HeartbeatRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ boolean hasRegistration(); /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration(); /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder(); /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ java.util.List getReportsList(); /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto getReports(int index); /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ int getReportsCount(); /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ java.util.List getReportsOrBuilderList(); /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProtoOrBuilder getReportsOrBuilder( int index); /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ boolean hasXmitsInProgress(); /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ int getXmitsInProgress(); /** * optional uint32 xceiverCount = 4 [default = 0]; */ boolean hasXceiverCount(); /** * optional uint32 xceiverCount = 4 [default = 0]; */ int getXceiverCount(); /** * optional uint32 failedVolumes = 5 [default = 0]; */ boolean hasFailedVolumes(); /** * optional uint32 failedVolumes = 5 [default = 0]; */ int getFailedVolumes(); /** * optional uint64 cacheCapacity = 6 [default = 0]; */ boolean hasCacheCapacity(); /** * optional uint64 cacheCapacity = 6 [default = 0]; */ long getCacheCapacity(); /** * optional uint64 cacheUsed = 7 [default = 0]; */ boolean hasCacheUsed(); /** * optional uint64 cacheUsed = 7 [default = 0]; */ long getCacheUsed(); /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ boolean hasVolumeFailureSummary(); /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto getVolumeFailureSummary(); /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder getVolumeFailureSummaryOrBuilder(); /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ boolean hasRequestFullBlockReportLease(); /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ boolean getRequestFullBlockReportLease(); /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ java.util.List getSlowPeersList(); /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto getSlowPeers(int index); /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ int getSlowPeersCount(); /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ java.util.List getSlowPeersOrBuilderList(); /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder getSlowPeersOrBuilder( int index); /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ java.util.List getSlowDisksList(); /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto getSlowDisks(int index); /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ int getSlowDisksCount(); /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ java.util.List getSlowDisksOrBuilderList(); /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder getSlowDisksOrBuilder( int index); } /** *
   **
   * registration - datanode registration information
   * capacity - total storage capacity available at the datanode
   * dfsUsed - storage used by HDFS
   * remaining - remaining storage available for HDFS
   * blockPoolUsed - storage used by the block pool
   * xmitsInProgress - number of transfers from this datanode to others
   * xceiverCount - number of active transceiver threads
   * failedVolumes - number of failed volumes.  This is redundant with the
   *     information included in volumeFailureSummary, but the field is retained
   *     for backwards compatibility.
   * cacheCapacity - total cache capacity available at the datanode
   * cacheUsed - amount of cache used
   * volumeFailureSummary - info about volume failures
   * slowPeers - info about peer DataNodes that are suspected to be slow.
   * slowDisks - info about DataNode disks that are suspected to be slow.
   * blksMovementResults - status of the scheduled blocks movements
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.HeartbeatRequestProto} */ public static final class HeartbeatRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.HeartbeatRequestProto) HeartbeatRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use HeartbeatRequestProto.newBuilder() to construct. private HeartbeatRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private HeartbeatRequestProto() { reports_ = java.util.Collections.emptyList(); slowPeers_ = java.util.Collections.emptyList(); slowDisks_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private HeartbeatRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registration_.toBuilder(); } registration_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registration_); registration_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { if (!((mutable_bitField0_ & 0x00000002) != 0)) { reports_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000002; } reports_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.PARSER, extensionRegistry)); break; } case 24: { bitField0_ |= 0x00000002; xmitsInProgress_ = input.readUInt32(); break; } case 32: { bitField0_ |= 0x00000004; xceiverCount_ = input.readUInt32(); break; } case 40: { bitField0_ |= 0x00000008; failedVolumes_ = input.readUInt32(); break; } case 48: { bitField0_ |= 0x00000010; cacheCapacity_ = input.readUInt64(); break; } case 56: { bitField0_ |= 0x00000020; cacheUsed_ = input.readUInt64(); break; } case 66: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder subBuilder = null; if (((bitField0_ & 0x00000040) != 0)) { subBuilder = volumeFailureSummary_.toBuilder(); } volumeFailureSummary_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(volumeFailureSummary_); volumeFailureSummary_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000040; break; } case 72: { bitField0_ |= 0x00000080; requestFullBlockReportLease_ = input.readBool(); break; } case 82: { if (!((mutable_bitField0_ & 0x00000200) != 0)) { slowPeers_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000200; } slowPeers_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.PARSER, extensionRegistry)); break; } case 90: { if (!((mutable_bitField0_ & 0x00000400) != 0)) { slowDisks_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000400; } slowDisks_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) != 0)) { reports_ = java.util.Collections.unmodifiableList(reports_); } if (((mutable_bitField0_ & 0x00000200) != 0)) { slowPeers_ = java.util.Collections.unmodifiableList(slowPeers_); } if (((mutable_bitField0_ & 0x00000400) != 0)) { slowDisks_ = java.util.Collections.unmodifiableList(slowDisks_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.Builder.class); } private int bitField0_; public static final int REGISTRATION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } /** *
     * Datanode info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } public static final int REPORTS_FIELD_NUMBER = 2; private java.util.List reports_; /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public java.util.List getReportsList() { return reports_; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public java.util.List getReportsOrBuilderList() { return reports_; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public int getReportsCount() { return reports_.size(); } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto getReports(int index) { return reports_.get(index); } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProtoOrBuilder getReportsOrBuilder( int index) { return reports_.get(index); } public static final int XMITSINPROGRESS_FIELD_NUMBER = 3; private int xmitsInProgress_; /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ public boolean hasXmitsInProgress() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ public int getXmitsInProgress() { return xmitsInProgress_; } public static final int XCEIVERCOUNT_FIELD_NUMBER = 4; private int xceiverCount_; /** * optional uint32 xceiverCount = 4 [default = 0]; */ public boolean hasXceiverCount() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint32 xceiverCount = 4 [default = 0]; */ public int getXceiverCount() { return xceiverCount_; } public static final int FAILEDVOLUMES_FIELD_NUMBER = 5; private int failedVolumes_; /** * optional uint32 failedVolumes = 5 [default = 0]; */ public boolean hasFailedVolumes() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint32 failedVolumes = 5 [default = 0]; */ public int getFailedVolumes() { return failedVolumes_; } public static final int CACHECAPACITY_FIELD_NUMBER = 6; private long cacheCapacity_; /** * optional uint64 cacheCapacity = 6 [default = 0]; */ public boolean hasCacheCapacity() { return ((bitField0_ & 0x00000010) != 0); } /** * optional uint64 cacheCapacity = 6 [default = 0]; */ public long getCacheCapacity() { return cacheCapacity_; } public static final int CACHEUSED_FIELD_NUMBER = 7; private long cacheUsed_; /** * optional uint64 cacheUsed = 7 [default = 0]; */ public boolean hasCacheUsed() { return ((bitField0_ & 0x00000020) != 0); } /** * optional uint64 cacheUsed = 7 [default = 0]; */ public long getCacheUsed() { return cacheUsed_; } public static final int VOLUMEFAILURESUMMARY_FIELD_NUMBER = 8; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto volumeFailureSummary_; /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public boolean hasVolumeFailureSummary() { return ((bitField0_ & 0x00000040) != 0); } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto getVolumeFailureSummary() { return volumeFailureSummary_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance() : volumeFailureSummary_; } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder getVolumeFailureSummaryOrBuilder() { return volumeFailureSummary_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance() : volumeFailureSummary_; } public static final int REQUESTFULLBLOCKREPORTLEASE_FIELD_NUMBER = 9; private boolean requestFullBlockReportLease_; /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ public boolean hasRequestFullBlockReportLease() { return ((bitField0_ & 0x00000080) != 0); } /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ public boolean getRequestFullBlockReportLease() { return requestFullBlockReportLease_; } public static final int SLOWPEERS_FIELD_NUMBER = 10; private java.util.List slowPeers_; /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public java.util.List getSlowPeersList() { return slowPeers_; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public java.util.List getSlowPeersOrBuilderList() { return slowPeers_; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public int getSlowPeersCount() { return slowPeers_.size(); } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto getSlowPeers(int index) { return slowPeers_.get(index); } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder getSlowPeersOrBuilder( int index) { return slowPeers_.get(index); } public static final int SLOWDISKS_FIELD_NUMBER = 11; private java.util.List slowDisks_; /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public java.util.List getSlowDisksList() { return slowDisks_; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public java.util.List getSlowDisksOrBuilderList() { return slowDisks_; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public int getSlowDisksCount() { return slowDisks_.size(); } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto getSlowDisks(int index) { return slowDisks_.get(index); } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder getSlowDisksOrBuilder( int index) { return slowDisks_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistration()) { memoizedIsInitialized = 0; return false; } if (!getRegistration().isInitialized()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getReportsCount(); i++) { if (!getReports(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasVolumeFailureSummary()) { if (!getVolumeFailureSummary().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistration()); } for (int i = 0; i < reports_.size(); i++) { output.writeMessage(2, reports_.get(i)); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt32(3, xmitsInProgress_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt32(4, xceiverCount_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeUInt32(5, failedVolumes_); } if (((bitField0_ & 0x00000010) != 0)) { output.writeUInt64(6, cacheCapacity_); } if (((bitField0_ & 0x00000020) != 0)) { output.writeUInt64(7, cacheUsed_); } if (((bitField0_ & 0x00000040) != 0)) { output.writeMessage(8, getVolumeFailureSummary()); } if (((bitField0_ & 0x00000080) != 0)) { output.writeBool(9, requestFullBlockReportLease_); } for (int i = 0; i < slowPeers_.size(); i++) { output.writeMessage(10, slowPeers_.get(i)); } for (int i = 0; i < slowDisks_.size(); i++) { output.writeMessage(11, slowDisks_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistration()); } for (int i = 0; i < reports_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(2, reports_.get(i)); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt32Size(3, xmitsInProgress_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt32Size(4, xceiverCount_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt32Size(5, failedVolumes_); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(6, cacheCapacity_); } if (((bitField0_ & 0x00000020) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(7, cacheUsed_); } if (((bitField0_ & 0x00000040) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(8, getVolumeFailureSummary()); } if (((bitField0_ & 0x00000080) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeBoolSize(9, requestFullBlockReportLease_); } for (int i = 0; i < slowPeers_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(10, slowPeers_.get(i)); } for (int i = 0; i < slowDisks_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(11, slowDisks_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto) obj; if (hasRegistration() != other.hasRegistration()) return false; if (hasRegistration()) { if (!getRegistration() .equals(other.getRegistration())) return false; } if (!getReportsList() .equals(other.getReportsList())) return false; if (hasXmitsInProgress() != other.hasXmitsInProgress()) return false; if (hasXmitsInProgress()) { if (getXmitsInProgress() != other.getXmitsInProgress()) return false; } if (hasXceiverCount() != other.hasXceiverCount()) return false; if (hasXceiverCount()) { if (getXceiverCount() != other.getXceiverCount()) return false; } if (hasFailedVolumes() != other.hasFailedVolumes()) return false; if (hasFailedVolumes()) { if (getFailedVolumes() != other.getFailedVolumes()) return false; } if (hasCacheCapacity() != other.hasCacheCapacity()) return false; if (hasCacheCapacity()) { if (getCacheCapacity() != other.getCacheCapacity()) return false; } if (hasCacheUsed() != other.hasCacheUsed()) return false; if (hasCacheUsed()) { if (getCacheUsed() != other.getCacheUsed()) return false; } if (hasVolumeFailureSummary() != other.hasVolumeFailureSummary()) return false; if (hasVolumeFailureSummary()) { if (!getVolumeFailureSummary() .equals(other.getVolumeFailureSummary())) return false; } if (hasRequestFullBlockReportLease() != other.hasRequestFullBlockReportLease()) return false; if (hasRequestFullBlockReportLease()) { if (getRequestFullBlockReportLease() != other.getRequestFullBlockReportLease()) return false; } if (!getSlowPeersList() .equals(other.getSlowPeersList())) return false; if (!getSlowDisksList() .equals(other.getSlowDisksList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistration()) { hash = (37 * hash) + REGISTRATION_FIELD_NUMBER; hash = (53 * hash) + getRegistration().hashCode(); } if (getReportsCount() > 0) { hash = (37 * hash) + REPORTS_FIELD_NUMBER; hash = (53 * hash) + getReportsList().hashCode(); } if (hasXmitsInProgress()) { hash = (37 * hash) + XMITSINPROGRESS_FIELD_NUMBER; hash = (53 * hash) + getXmitsInProgress(); } if (hasXceiverCount()) { hash = (37 * hash) + XCEIVERCOUNT_FIELD_NUMBER; hash = (53 * hash) + getXceiverCount(); } if (hasFailedVolumes()) { hash = (37 * hash) + FAILEDVOLUMES_FIELD_NUMBER; hash = (53 * hash) + getFailedVolumes(); } if (hasCacheCapacity()) { hash = (37 * hash) + CACHECAPACITY_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getCacheCapacity()); } if (hasCacheUsed()) { hash = (37 * hash) + CACHEUSED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getCacheUsed()); } if (hasVolumeFailureSummary()) { hash = (37 * hash) + VOLUMEFAILURESUMMARY_FIELD_NUMBER; hash = (53 * hash) + getVolumeFailureSummary().hashCode(); } if (hasRequestFullBlockReportLease()) { hash = (37 * hash) + REQUESTFULLBLOCKREPORTLEASE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean( getRequestFullBlockReportLease()); } if (getSlowPeersCount() > 0) { hash = (37 * hash) + SLOWPEERS_FIELD_NUMBER; hash = (53 * hash) + getSlowPeersList().hashCode(); } if (getSlowDisksCount() > 0) { hash = (37 * hash) + SLOWDISKS_FIELD_NUMBER; hash = (53 * hash) + getSlowDisksList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registration - datanode registration information
     * capacity - total storage capacity available at the datanode
     * dfsUsed - storage used by HDFS
     * remaining - remaining storage available for HDFS
     * blockPoolUsed - storage used by the block pool
     * xmitsInProgress - number of transfers from this datanode to others
     * xceiverCount - number of active transceiver threads
     * failedVolumes - number of failed volumes.  This is redundant with the
     *     information included in volumeFailureSummary, but the field is retained
     *     for backwards compatibility.
     * cacheCapacity - total cache capacity available at the datanode
     * cacheUsed - amount of cache used
     * volumeFailureSummary - info about volume failures
     * slowPeers - info about peer DataNodes that are suspected to be slow.
     * slowDisks - info about DataNode disks that are suspected to be slow.
     * blksMovementResults - status of the scheduled blocks movements
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.HeartbeatRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.HeartbeatRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistrationFieldBuilder(); getReportsFieldBuilder(); getVolumeFailureSummaryFieldBuilder(); getSlowPeersFieldBuilder(); getSlowDisksFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registrationBuilder_ == null) { registration_ = null; } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); if (reportsBuilder_ == null) { reports_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); } else { reportsBuilder_.clear(); } xmitsInProgress_ = 0; bitField0_ = (bitField0_ & ~0x00000004); xceiverCount_ = 0; bitField0_ = (bitField0_ & ~0x00000008); failedVolumes_ = 0; bitField0_ = (bitField0_ & ~0x00000010); cacheCapacity_ = 0L; bitField0_ = (bitField0_ & ~0x00000020); cacheUsed_ = 0L; bitField0_ = (bitField0_ & ~0x00000040); if (volumeFailureSummaryBuilder_ == null) { volumeFailureSummary_ = null; } else { volumeFailureSummaryBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000080); requestFullBlockReportLease_ = false; bitField0_ = (bitField0_ & ~0x00000100); if (slowPeersBuilder_ == null) { slowPeers_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000200); } else { slowPeersBuilder_.clear(); } if (slowDisksBuilder_ == null) { slowDisks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000400); } else { slowDisksBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registrationBuilder_ == null) { result.registration_ = registration_; } else { result.registration_ = registrationBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (reportsBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0)) { reports_ = java.util.Collections.unmodifiableList(reports_); bitField0_ = (bitField0_ & ~0x00000002); } result.reports_ = reports_; } else { result.reports_ = reportsBuilder_.build(); } if (((from_bitField0_ & 0x00000004) != 0)) { result.xmitsInProgress_ = xmitsInProgress_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000008) != 0)) { result.xceiverCount_ = xceiverCount_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000010) != 0)) { result.failedVolumes_ = failedVolumes_; to_bitField0_ |= 0x00000008; } if (((from_bitField0_ & 0x00000020) != 0)) { result.cacheCapacity_ = cacheCapacity_; to_bitField0_ |= 0x00000010; } if (((from_bitField0_ & 0x00000040) != 0)) { result.cacheUsed_ = cacheUsed_; to_bitField0_ |= 0x00000020; } if (((from_bitField0_ & 0x00000080) != 0)) { if (volumeFailureSummaryBuilder_ == null) { result.volumeFailureSummary_ = volumeFailureSummary_; } else { result.volumeFailureSummary_ = volumeFailureSummaryBuilder_.build(); } to_bitField0_ |= 0x00000040; } if (((from_bitField0_ & 0x00000100) != 0)) { result.requestFullBlockReportLease_ = requestFullBlockReportLease_; to_bitField0_ |= 0x00000080; } if (slowPeersBuilder_ == null) { if (((bitField0_ & 0x00000200) != 0)) { slowPeers_ = java.util.Collections.unmodifiableList(slowPeers_); bitField0_ = (bitField0_ & ~0x00000200); } result.slowPeers_ = slowPeers_; } else { result.slowPeers_ = slowPeersBuilder_.build(); } if (slowDisksBuilder_ == null) { if (((bitField0_ & 0x00000400) != 0)) { slowDisks_ = java.util.Collections.unmodifiableList(slowDisks_); bitField0_ = (bitField0_ & ~0x00000400); } result.slowDisks_ = slowDisks_; } else { result.slowDisks_ = slowDisksBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.getDefaultInstance()) return this; if (other.hasRegistration()) { mergeRegistration(other.getRegistration()); } if (reportsBuilder_ == null) { if (!other.reports_.isEmpty()) { if (reports_.isEmpty()) { reports_ = other.reports_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureReportsIsMutable(); reports_.addAll(other.reports_); } onChanged(); } } else { if (!other.reports_.isEmpty()) { if (reportsBuilder_.isEmpty()) { reportsBuilder_.dispose(); reportsBuilder_ = null; reports_ = other.reports_; bitField0_ = (bitField0_ & ~0x00000002); reportsBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getReportsFieldBuilder() : null; } else { reportsBuilder_.addAllMessages(other.reports_); } } } if (other.hasXmitsInProgress()) { setXmitsInProgress(other.getXmitsInProgress()); } if (other.hasXceiverCount()) { setXceiverCount(other.getXceiverCount()); } if (other.hasFailedVolumes()) { setFailedVolumes(other.getFailedVolumes()); } if (other.hasCacheCapacity()) { setCacheCapacity(other.getCacheCapacity()); } if (other.hasCacheUsed()) { setCacheUsed(other.getCacheUsed()); } if (other.hasVolumeFailureSummary()) { mergeVolumeFailureSummary(other.getVolumeFailureSummary()); } if (other.hasRequestFullBlockReportLease()) { setRequestFullBlockReportLease(other.getRequestFullBlockReportLease()); } if (slowPeersBuilder_ == null) { if (!other.slowPeers_.isEmpty()) { if (slowPeers_.isEmpty()) { slowPeers_ = other.slowPeers_; bitField0_ = (bitField0_ & ~0x00000200); } else { ensureSlowPeersIsMutable(); slowPeers_.addAll(other.slowPeers_); } onChanged(); } } else { if (!other.slowPeers_.isEmpty()) { if (slowPeersBuilder_.isEmpty()) { slowPeersBuilder_.dispose(); slowPeersBuilder_ = null; slowPeers_ = other.slowPeers_; bitField0_ = (bitField0_ & ~0x00000200); slowPeersBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getSlowPeersFieldBuilder() : null; } else { slowPeersBuilder_.addAllMessages(other.slowPeers_); } } } if (slowDisksBuilder_ == null) { if (!other.slowDisks_.isEmpty()) { if (slowDisks_.isEmpty()) { slowDisks_ = other.slowDisks_; bitField0_ = (bitField0_ & ~0x00000400); } else { ensureSlowDisksIsMutable(); slowDisks_.addAll(other.slowDisks_); } onChanged(); } } else { if (!other.slowDisks_.isEmpty()) { if (slowDisksBuilder_.isEmpty()) { slowDisksBuilder_.dispose(); slowDisksBuilder_ = null; slowDisks_ = other.slowDisks_; bitField0_ = (bitField0_ & ~0x00000400); slowDisksBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getSlowDisksFieldBuilder() : null; } else { slowDisksBuilder_.addAllMessages(other.slowDisks_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistration()) { return false; } if (!getRegistration().isInitialized()) { return false; } for (int i = 0; i < getReportsCount(); i++) { if (!getReports(i).isInitialized()) { return false; } } if (hasVolumeFailureSummary()) { if (!getVolumeFailureSummary().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registrationBuilder_; /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { if (registrationBuilder_ == null) { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } else { return registrationBuilder_.getMessage(); } } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registration_ = value; onChanged(); } else { registrationBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registrationBuilder_ == null) { registration_ = builderForValue.build(); onChanged(); } else { registrationBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder mergeRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registration_ != null && registration_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registration_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registration_).mergeFrom(value).buildPartial(); } else { registration_ = value; } onChanged(); } else { registrationBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder clearRegistration() { if (registrationBuilder_ == null) { registration_ = null; onChanged(); } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistrationBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistrationFieldBuilder().getBuilder(); } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { if (registrationBuilder_ != null) { return registrationBuilder_.getMessageOrBuilder(); } else { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } } /** *
       * Datanode info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistrationFieldBuilder() { if (registrationBuilder_ == null) { registrationBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistration(), getParentForChildren(), isClean()); registration_ = null; } return registrationBuilder_; } private java.util.List reports_ = java.util.Collections.emptyList(); private void ensureReportsIsMutable() { if (!((bitField0_ & 0x00000002) != 0)) { reports_ = new java.util.ArrayList(reports_); bitField0_ |= 0x00000002; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProtoOrBuilder> reportsBuilder_; /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public java.util.List getReportsList() { if (reportsBuilder_ == null) { return java.util.Collections.unmodifiableList(reports_); } else { return reportsBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public int getReportsCount() { if (reportsBuilder_ == null) { return reports_.size(); } else { return reportsBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto getReports(int index) { if (reportsBuilder_ == null) { return reports_.get(index); } else { return reportsBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder setReports( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto value) { if (reportsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureReportsIsMutable(); reports_.set(index, value); onChanged(); } else { reportsBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder setReports( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder builderForValue) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.set(index, builderForValue.build()); onChanged(); } else { reportsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder addReports(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto value) { if (reportsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureReportsIsMutable(); reports_.add(value); onChanged(); } else { reportsBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder addReports( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto value) { if (reportsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureReportsIsMutable(); reports_.add(index, value); onChanged(); } else { reportsBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder addReports( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder builderForValue) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.add(builderForValue.build()); onChanged(); } else { reportsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder addReports( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder builderForValue) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.add(index, builderForValue.build()); onChanged(); } else { reportsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder addAllReports( java.lang.Iterable values) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, reports_); onChanged(); } else { reportsBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder clearReports() { if (reportsBuilder_ == null) { reports_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); } else { reportsBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public Builder removeReports(int index) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.remove(index); onChanged(); } else { reportsBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder getReportsBuilder( int index) { return getReportsFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProtoOrBuilder getReportsOrBuilder( int index) { if (reportsBuilder_ == null) { return reports_.get(index); } else { return reportsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public java.util.List getReportsOrBuilderList() { if (reportsBuilder_ != null) { return reportsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(reports_); } } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder addReportsBuilder() { return getReportsFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder addReportsBuilder( int index) { return getReportsFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.StorageReportProto reports = 2; */ public java.util.List getReportsBuilderList() { return getReportsFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProtoOrBuilder> getReportsFieldBuilder() { if (reportsBuilder_ == null) { reportsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.StorageReportProtoOrBuilder>( reports_, ((bitField0_ & 0x00000002) != 0), getParentForChildren(), isClean()); reports_ = null; } return reportsBuilder_; } private int xmitsInProgress_ ; /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ public boolean hasXmitsInProgress() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ public int getXmitsInProgress() { return xmitsInProgress_; } /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ public Builder setXmitsInProgress(int value) { bitField0_ |= 0x00000004; xmitsInProgress_ = value; onChanged(); return this; } /** * optional uint32 xmitsInProgress = 3 [default = 0]; */ public Builder clearXmitsInProgress() { bitField0_ = (bitField0_ & ~0x00000004); xmitsInProgress_ = 0; onChanged(); return this; } private int xceiverCount_ ; /** * optional uint32 xceiverCount = 4 [default = 0]; */ public boolean hasXceiverCount() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint32 xceiverCount = 4 [default = 0]; */ public int getXceiverCount() { return xceiverCount_; } /** * optional uint32 xceiverCount = 4 [default = 0]; */ public Builder setXceiverCount(int value) { bitField0_ |= 0x00000008; xceiverCount_ = value; onChanged(); return this; } /** * optional uint32 xceiverCount = 4 [default = 0]; */ public Builder clearXceiverCount() { bitField0_ = (bitField0_ & ~0x00000008); xceiverCount_ = 0; onChanged(); return this; } private int failedVolumes_ ; /** * optional uint32 failedVolumes = 5 [default = 0]; */ public boolean hasFailedVolumes() { return ((bitField0_ & 0x00000010) != 0); } /** * optional uint32 failedVolumes = 5 [default = 0]; */ public int getFailedVolumes() { return failedVolumes_; } /** * optional uint32 failedVolumes = 5 [default = 0]; */ public Builder setFailedVolumes(int value) { bitField0_ |= 0x00000010; failedVolumes_ = value; onChanged(); return this; } /** * optional uint32 failedVolumes = 5 [default = 0]; */ public Builder clearFailedVolumes() { bitField0_ = (bitField0_ & ~0x00000010); failedVolumes_ = 0; onChanged(); return this; } private long cacheCapacity_ ; /** * optional uint64 cacheCapacity = 6 [default = 0]; */ public boolean hasCacheCapacity() { return ((bitField0_ & 0x00000020) != 0); } /** * optional uint64 cacheCapacity = 6 [default = 0]; */ public long getCacheCapacity() { return cacheCapacity_; } /** * optional uint64 cacheCapacity = 6 [default = 0]; */ public Builder setCacheCapacity(long value) { bitField0_ |= 0x00000020; cacheCapacity_ = value; onChanged(); return this; } /** * optional uint64 cacheCapacity = 6 [default = 0]; */ public Builder clearCacheCapacity() { bitField0_ = (bitField0_ & ~0x00000020); cacheCapacity_ = 0L; onChanged(); return this; } private long cacheUsed_ ; /** * optional uint64 cacheUsed = 7 [default = 0]; */ public boolean hasCacheUsed() { return ((bitField0_ & 0x00000040) != 0); } /** * optional uint64 cacheUsed = 7 [default = 0]; */ public long getCacheUsed() { return cacheUsed_; } /** * optional uint64 cacheUsed = 7 [default = 0]; */ public Builder setCacheUsed(long value) { bitField0_ |= 0x00000040; cacheUsed_ = value; onChanged(); return this; } /** * optional uint64 cacheUsed = 7 [default = 0]; */ public Builder clearCacheUsed() { bitField0_ = (bitField0_ & ~0x00000040); cacheUsed_ = 0L; onChanged(); return this; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto volumeFailureSummary_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder> volumeFailureSummaryBuilder_; /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public boolean hasVolumeFailureSummary() { return ((bitField0_ & 0x00000080) != 0); } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto getVolumeFailureSummary() { if (volumeFailureSummaryBuilder_ == null) { return volumeFailureSummary_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance() : volumeFailureSummary_; } else { return volumeFailureSummaryBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public Builder setVolumeFailureSummary(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto value) { if (volumeFailureSummaryBuilder_ == null) { if (value == null) { throw new NullPointerException(); } volumeFailureSummary_ = value; onChanged(); } else { volumeFailureSummaryBuilder_.setMessage(value); } bitField0_ |= 0x00000080; return this; } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public Builder setVolumeFailureSummary( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder builderForValue) { if (volumeFailureSummaryBuilder_ == null) { volumeFailureSummary_ = builderForValue.build(); onChanged(); } else { volumeFailureSummaryBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000080; return this; } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public Builder mergeVolumeFailureSummary(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto value) { if (volumeFailureSummaryBuilder_ == null) { if (((bitField0_ & 0x00000080) != 0) && volumeFailureSummary_ != null && volumeFailureSummary_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance()) { volumeFailureSummary_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.newBuilder(volumeFailureSummary_).mergeFrom(value).buildPartial(); } else { volumeFailureSummary_ = value; } onChanged(); } else { volumeFailureSummaryBuilder_.mergeFrom(value); } bitField0_ |= 0x00000080; return this; } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public Builder clearVolumeFailureSummary() { if (volumeFailureSummaryBuilder_ == null) { volumeFailureSummary_ = null; onChanged(); } else { volumeFailureSummaryBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000080); return this; } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder getVolumeFailureSummaryBuilder() { bitField0_ |= 0x00000080; onChanged(); return getVolumeFailureSummaryFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder getVolumeFailureSummaryOrBuilder() { if (volumeFailureSummaryBuilder_ != null) { return volumeFailureSummaryBuilder_.getMessageOrBuilder(); } else { return volumeFailureSummary_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.getDefaultInstance() : volumeFailureSummary_; } } /** * optional .hadoop.hdfs.datanode.VolumeFailureSummaryProto volumeFailureSummary = 8; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder> getVolumeFailureSummaryFieldBuilder() { if (volumeFailureSummaryBuilder_ == null) { volumeFailureSummaryBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.VolumeFailureSummaryProtoOrBuilder>( getVolumeFailureSummary(), getParentForChildren(), isClean()); volumeFailureSummary_ = null; } return volumeFailureSummaryBuilder_; } private boolean requestFullBlockReportLease_ ; /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ public boolean hasRequestFullBlockReportLease() { return ((bitField0_ & 0x00000100) != 0); } /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ public boolean getRequestFullBlockReportLease() { return requestFullBlockReportLease_; } /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ public Builder setRequestFullBlockReportLease(boolean value) { bitField0_ |= 0x00000100; requestFullBlockReportLease_ = value; onChanged(); return this; } /** * optional bool requestFullBlockReportLease = 9 [default = false]; */ public Builder clearRequestFullBlockReportLease() { bitField0_ = (bitField0_ & ~0x00000100); requestFullBlockReportLease_ = false; onChanged(); return this; } private java.util.List slowPeers_ = java.util.Collections.emptyList(); private void ensureSlowPeersIsMutable() { if (!((bitField0_ & 0x00000200) != 0)) { slowPeers_ = new java.util.ArrayList(slowPeers_); bitField0_ |= 0x00000200; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder> slowPeersBuilder_; /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public java.util.List getSlowPeersList() { if (slowPeersBuilder_ == null) { return java.util.Collections.unmodifiableList(slowPeers_); } else { return slowPeersBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public int getSlowPeersCount() { if (slowPeersBuilder_ == null) { return slowPeers_.size(); } else { return slowPeersBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto getSlowPeers(int index) { if (slowPeersBuilder_ == null) { return slowPeers_.get(index); } else { return slowPeersBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder setSlowPeers( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto value) { if (slowPeersBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSlowPeersIsMutable(); slowPeers_.set(index, value); onChanged(); } else { slowPeersBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder setSlowPeers( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder builderForValue) { if (slowPeersBuilder_ == null) { ensureSlowPeersIsMutable(); slowPeers_.set(index, builderForValue.build()); onChanged(); } else { slowPeersBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder addSlowPeers(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto value) { if (slowPeersBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSlowPeersIsMutable(); slowPeers_.add(value); onChanged(); } else { slowPeersBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder addSlowPeers( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto value) { if (slowPeersBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSlowPeersIsMutable(); slowPeers_.add(index, value); onChanged(); } else { slowPeersBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder addSlowPeers( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder builderForValue) { if (slowPeersBuilder_ == null) { ensureSlowPeersIsMutable(); slowPeers_.add(builderForValue.build()); onChanged(); } else { slowPeersBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder addSlowPeers( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder builderForValue) { if (slowPeersBuilder_ == null) { ensureSlowPeersIsMutable(); slowPeers_.add(index, builderForValue.build()); onChanged(); } else { slowPeersBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder addAllSlowPeers( java.lang.Iterable values) { if (slowPeersBuilder_ == null) { ensureSlowPeersIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, slowPeers_); onChanged(); } else { slowPeersBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder clearSlowPeers() { if (slowPeersBuilder_ == null) { slowPeers_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000200); onChanged(); } else { slowPeersBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public Builder removeSlowPeers(int index) { if (slowPeersBuilder_ == null) { ensureSlowPeersIsMutable(); slowPeers_.remove(index); onChanged(); } else { slowPeersBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder getSlowPeersBuilder( int index) { return getSlowPeersFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder getSlowPeersOrBuilder( int index) { if (slowPeersBuilder_ == null) { return slowPeers_.get(index); } else { return slowPeersBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public java.util.List getSlowPeersOrBuilderList() { if (slowPeersBuilder_ != null) { return slowPeersBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(slowPeers_); } } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder addSlowPeersBuilder() { return getSlowPeersFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder addSlowPeersBuilder( int index) { return getSlowPeersFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.SlowPeerReportProto slowPeers = 10; */ public java.util.List getSlowPeersBuilderList() { return getSlowPeersFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder> getSlowPeersFieldBuilder() { if (slowPeersBuilder_ == null) { slowPeersBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder>( slowPeers_, ((bitField0_ & 0x00000200) != 0), getParentForChildren(), isClean()); slowPeers_ = null; } return slowPeersBuilder_; } private java.util.List slowDisks_ = java.util.Collections.emptyList(); private void ensureSlowDisksIsMutable() { if (!((bitField0_ & 0x00000400) != 0)) { slowDisks_ = new java.util.ArrayList(slowDisks_); bitField0_ |= 0x00000400; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder> slowDisksBuilder_; /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public java.util.List getSlowDisksList() { if (slowDisksBuilder_ == null) { return java.util.Collections.unmodifiableList(slowDisks_); } else { return slowDisksBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public int getSlowDisksCount() { if (slowDisksBuilder_ == null) { return slowDisks_.size(); } else { return slowDisksBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto getSlowDisks(int index) { if (slowDisksBuilder_ == null) { return slowDisks_.get(index); } else { return slowDisksBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder setSlowDisks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto value) { if (slowDisksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSlowDisksIsMutable(); slowDisks_.set(index, value); onChanged(); } else { slowDisksBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder setSlowDisks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder builderForValue) { if (slowDisksBuilder_ == null) { ensureSlowDisksIsMutable(); slowDisks_.set(index, builderForValue.build()); onChanged(); } else { slowDisksBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder addSlowDisks(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto value) { if (slowDisksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSlowDisksIsMutable(); slowDisks_.add(value); onChanged(); } else { slowDisksBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder addSlowDisks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto value) { if (slowDisksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSlowDisksIsMutable(); slowDisks_.add(index, value); onChanged(); } else { slowDisksBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder addSlowDisks( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder builderForValue) { if (slowDisksBuilder_ == null) { ensureSlowDisksIsMutable(); slowDisks_.add(builderForValue.build()); onChanged(); } else { slowDisksBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder addSlowDisks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder builderForValue) { if (slowDisksBuilder_ == null) { ensureSlowDisksIsMutable(); slowDisks_.add(index, builderForValue.build()); onChanged(); } else { slowDisksBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder addAllSlowDisks( java.lang.Iterable values) { if (slowDisksBuilder_ == null) { ensureSlowDisksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, slowDisks_); onChanged(); } else { slowDisksBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder clearSlowDisks() { if (slowDisksBuilder_ == null) { slowDisks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000400); onChanged(); } else { slowDisksBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public Builder removeSlowDisks(int index) { if (slowDisksBuilder_ == null) { ensureSlowDisksIsMutable(); slowDisks_.remove(index); onChanged(); } else { slowDisksBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder getSlowDisksBuilder( int index) { return getSlowDisksFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder getSlowDisksOrBuilder( int index) { if (slowDisksBuilder_ == null) { return slowDisks_.get(index); } else { return slowDisksBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public java.util.List getSlowDisksOrBuilderList() { if (slowDisksBuilder_ != null) { return slowDisksBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(slowDisks_); } } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder addSlowDisksBuilder() { return getSlowDisksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder addSlowDisksBuilder( int index) { return getSlowDisksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.SlowDiskReportProto slowDisks = 11; */ public java.util.List getSlowDisksBuilderList() { return getSlowDisksFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder> getSlowDisksFieldBuilder() { if (slowDisksBuilder_ == null) { slowDisksBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder>( slowDisks_, ((bitField0_ & 0x00000400) != 0), getParentForChildren(), isClean()); slowDisks_ = null; } return slowDisksBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.HeartbeatRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.HeartbeatRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public HeartbeatRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new HeartbeatRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface HeartbeatResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.HeartbeatResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ java.util.List getCmdsList(); /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmds(int index); /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ int getCmdsCount(); /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ java.util.List getCmdsOrBuilderList(); /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdsOrBuilder( int index); /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ boolean hasHaStatus(); /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto getHaStatus(); /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProtoOrBuilder getHaStatusOrBuilder(); /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ boolean hasRollingUpgradeStatus(); /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto getRollingUpgradeStatus(); /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder getRollingUpgradeStatusOrBuilder(); /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ boolean hasRollingUpgradeStatusV2(); /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto getRollingUpgradeStatusV2(); /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder getRollingUpgradeStatusV2OrBuilder(); /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ boolean hasFullBlockReportLeaseId(); /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ long getFullBlockReportLeaseId(); } /** *
   **
   * cmds - Commands from namenode to datanode.
   * haStatus - Status (from an HA perspective) of the NN sending this response
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.HeartbeatResponseProto} */ public static final class HeartbeatResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.HeartbeatResponseProto) HeartbeatResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use HeartbeatResponseProto.newBuilder() to construct. private HeartbeatResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private HeartbeatResponseProto() { cmds_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private HeartbeatResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { cmds_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } cmds_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.PARSER, extensionRegistry)); break; } case 18: { org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = haStatus_.toBuilder(); } haStatus_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(haStatus_); haStatus_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 26: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = rollingUpgradeStatus_.toBuilder(); } rollingUpgradeStatus_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(rollingUpgradeStatus_); rollingUpgradeStatus_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 34: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder subBuilder = null; if (((bitField0_ & 0x00000004) != 0)) { subBuilder = rollingUpgradeStatusV2_.toBuilder(); } rollingUpgradeStatusV2_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(rollingUpgradeStatusV2_); rollingUpgradeStatusV2_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } case 40: { bitField0_ |= 0x00000008; fullBlockReportLeaseId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { cmds_ = java.util.Collections.unmodifiableList(cmds_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.Builder.class); } private int bitField0_; public static final int CMDS_FIELD_NUMBER = 1; private java.util.List cmds_; /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public java.util.List getCmdsList() { return cmds_; } /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public java.util.List getCmdsOrBuilderList() { return cmds_; } /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public int getCmdsCount() { return cmds_.size(); } /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmds(int index) { return cmds_.get(index); } /** *
     * Returned commands can be null
     * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdsOrBuilder( int index) { return cmds_.get(index); } public static final int HASTATUS_FIELD_NUMBER = 2; private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto haStatus_; /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public boolean hasHaStatus() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto getHaStatus() { return haStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.getDefaultInstance() : haStatus_; } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProtoOrBuilder getHaStatusOrBuilder() { return haStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.getDefaultInstance() : haStatus_; } public static final int ROLLINGUPGRADESTATUS_FIELD_NUMBER = 3; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto rollingUpgradeStatus_; /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public boolean hasRollingUpgradeStatus() { return ((bitField0_ & 0x00000002) != 0); } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto getRollingUpgradeStatus() { return rollingUpgradeStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatus_; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder getRollingUpgradeStatusOrBuilder() { return rollingUpgradeStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatus_; } public static final int ROLLINGUPGRADESTATUSV2_FIELD_NUMBER = 4; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto rollingUpgradeStatusV2_; /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public boolean hasRollingUpgradeStatusV2() { return ((bitField0_ & 0x00000004) != 0); } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto getRollingUpgradeStatusV2() { return rollingUpgradeStatusV2_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatusV2_; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder getRollingUpgradeStatusV2OrBuilder() { return rollingUpgradeStatusV2_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatusV2_; } public static final int FULLBLOCKREPORTLEASEID_FIELD_NUMBER = 5; private long fullBlockReportLeaseId_; /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ public boolean hasFullBlockReportLeaseId() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ public long getFullBlockReportLeaseId() { return fullBlockReportLeaseId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasHaStatus()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getCmdsCount(); i++) { if (!getCmds(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } if (!getHaStatus().isInitialized()) { memoizedIsInitialized = 0; return false; } if (hasRollingUpgradeStatus()) { if (!getRollingUpgradeStatus().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasRollingUpgradeStatusV2()) { if (!getRollingUpgradeStatusV2().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < cmds_.size(); i++) { output.writeMessage(1, cmds_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(2, getHaStatus()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(3, getRollingUpgradeStatus()); } if (((bitField0_ & 0x00000004) != 0)) { output.writeMessage(4, getRollingUpgradeStatusV2()); } if (((bitField0_ & 0x00000008) != 0)) { output.writeUInt64(5, fullBlockReportLeaseId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < cmds_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, cmds_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(2, getHaStatus()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, getRollingUpgradeStatus()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(4, getRollingUpgradeStatusV2()); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(5, fullBlockReportLeaseId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto) obj; if (!getCmdsList() .equals(other.getCmdsList())) return false; if (hasHaStatus() != other.hasHaStatus()) return false; if (hasHaStatus()) { if (!getHaStatus() .equals(other.getHaStatus())) return false; } if (hasRollingUpgradeStatus() != other.hasRollingUpgradeStatus()) return false; if (hasRollingUpgradeStatus()) { if (!getRollingUpgradeStatus() .equals(other.getRollingUpgradeStatus())) return false; } if (hasRollingUpgradeStatusV2() != other.hasRollingUpgradeStatusV2()) return false; if (hasRollingUpgradeStatusV2()) { if (!getRollingUpgradeStatusV2() .equals(other.getRollingUpgradeStatusV2())) return false; } if (hasFullBlockReportLeaseId() != other.hasFullBlockReportLeaseId()) return false; if (hasFullBlockReportLeaseId()) { if (getFullBlockReportLeaseId() != other.getFullBlockReportLeaseId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getCmdsCount() > 0) { hash = (37 * hash) + CMDS_FIELD_NUMBER; hash = (53 * hash) + getCmdsList().hashCode(); } if (hasHaStatus()) { hash = (37 * hash) + HASTATUS_FIELD_NUMBER; hash = (53 * hash) + getHaStatus().hashCode(); } if (hasRollingUpgradeStatus()) { hash = (37 * hash) + ROLLINGUPGRADESTATUS_FIELD_NUMBER; hash = (53 * hash) + getRollingUpgradeStatus().hashCode(); } if (hasRollingUpgradeStatusV2()) { hash = (37 * hash) + ROLLINGUPGRADESTATUSV2_FIELD_NUMBER; hash = (53 * hash) + getRollingUpgradeStatusV2().hashCode(); } if (hasFullBlockReportLeaseId()) { hash = (37 * hash) + FULLBLOCKREPORTLEASEID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getFullBlockReportLeaseId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * cmds - Commands from namenode to datanode.
     * haStatus - Status (from an HA perspective) of the NN sending this response
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.HeartbeatResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.HeartbeatResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getCmdsFieldBuilder(); getHaStatusFieldBuilder(); getRollingUpgradeStatusFieldBuilder(); getRollingUpgradeStatusV2FieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (cmdsBuilder_ == null) { cmds_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { cmdsBuilder_.clear(); } if (haStatusBuilder_ == null) { haStatus_ = null; } else { haStatusBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); if (rollingUpgradeStatusBuilder_ == null) { rollingUpgradeStatus_ = null; } else { rollingUpgradeStatusBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); if (rollingUpgradeStatusV2Builder_ == null) { rollingUpgradeStatusV2_ = null; } else { rollingUpgradeStatusV2Builder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); fullBlockReportLeaseId_ = 0L; bitField0_ = (bitField0_ & ~0x00000010); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (cmdsBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { cmds_ = java.util.Collections.unmodifiableList(cmds_); bitField0_ = (bitField0_ & ~0x00000001); } result.cmds_ = cmds_; } else { result.cmds_ = cmdsBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { if (haStatusBuilder_ == null) { result.haStatus_ = haStatus_; } else { result.haStatus_ = haStatusBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000004) != 0)) { if (rollingUpgradeStatusBuilder_ == null) { result.rollingUpgradeStatus_ = rollingUpgradeStatus_; } else { result.rollingUpgradeStatus_ = rollingUpgradeStatusBuilder_.build(); } to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000008) != 0)) { if (rollingUpgradeStatusV2Builder_ == null) { result.rollingUpgradeStatusV2_ = rollingUpgradeStatusV2_; } else { result.rollingUpgradeStatusV2_ = rollingUpgradeStatusV2Builder_.build(); } to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000010) != 0)) { result.fullBlockReportLeaseId_ = fullBlockReportLeaseId_; to_bitField0_ |= 0x00000008; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance()) return this; if (cmdsBuilder_ == null) { if (!other.cmds_.isEmpty()) { if (cmds_.isEmpty()) { cmds_ = other.cmds_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureCmdsIsMutable(); cmds_.addAll(other.cmds_); } onChanged(); } } else { if (!other.cmds_.isEmpty()) { if (cmdsBuilder_.isEmpty()) { cmdsBuilder_.dispose(); cmdsBuilder_ = null; cmds_ = other.cmds_; bitField0_ = (bitField0_ & ~0x00000001); cmdsBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getCmdsFieldBuilder() : null; } else { cmdsBuilder_.addAllMessages(other.cmds_); } } } if (other.hasHaStatus()) { mergeHaStatus(other.getHaStatus()); } if (other.hasRollingUpgradeStatus()) { mergeRollingUpgradeStatus(other.getRollingUpgradeStatus()); } if (other.hasRollingUpgradeStatusV2()) { mergeRollingUpgradeStatusV2(other.getRollingUpgradeStatusV2()); } if (other.hasFullBlockReportLeaseId()) { setFullBlockReportLeaseId(other.getFullBlockReportLeaseId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasHaStatus()) { return false; } for (int i = 0; i < getCmdsCount(); i++) { if (!getCmds(i).isInitialized()) { return false; } } if (!getHaStatus().isInitialized()) { return false; } if (hasRollingUpgradeStatus()) { if (!getRollingUpgradeStatus().isInitialized()) { return false; } } if (hasRollingUpgradeStatusV2()) { if (!getRollingUpgradeStatusV2().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List cmds_ = java.util.Collections.emptyList(); private void ensureCmdsIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { cmds_ = new java.util.ArrayList(cmds_); bitField0_ |= 0x00000001; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder> cmdsBuilder_; /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public java.util.List getCmdsList() { if (cmdsBuilder_ == null) { return java.util.Collections.unmodifiableList(cmds_); } else { return cmdsBuilder_.getMessageList(); } } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public int getCmdsCount() { if (cmdsBuilder_ == null) { return cmds_.size(); } else { return cmdsBuilder_.getCount(); } } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmds(int index) { if (cmdsBuilder_ == null) { return cmds_.get(index); } else { return cmdsBuilder_.getMessage(index); } } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder setCmds( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureCmdsIsMutable(); cmds_.set(index, value); onChanged(); } else { cmdsBuilder_.setMessage(index, value); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder setCmds( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder builderForValue) { if (cmdsBuilder_ == null) { ensureCmdsIsMutable(); cmds_.set(index, builderForValue.build()); onChanged(); } else { cmdsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder addCmds(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureCmdsIsMutable(); cmds_.add(value); onChanged(); } else { cmdsBuilder_.addMessage(value); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder addCmds( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureCmdsIsMutable(); cmds_.add(index, value); onChanged(); } else { cmdsBuilder_.addMessage(index, value); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder addCmds( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder builderForValue) { if (cmdsBuilder_ == null) { ensureCmdsIsMutable(); cmds_.add(builderForValue.build()); onChanged(); } else { cmdsBuilder_.addMessage(builderForValue.build()); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder addCmds( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder builderForValue) { if (cmdsBuilder_ == null) { ensureCmdsIsMutable(); cmds_.add(index, builderForValue.build()); onChanged(); } else { cmdsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder addAllCmds( java.lang.Iterable values) { if (cmdsBuilder_ == null) { ensureCmdsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, cmds_); onChanged(); } else { cmdsBuilder_.addAllMessages(values); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder clearCmds() { if (cmdsBuilder_ == null) { cmds_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { cmdsBuilder_.clear(); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public Builder removeCmds(int index) { if (cmdsBuilder_ == null) { ensureCmdsIsMutable(); cmds_.remove(index); onChanged(); } else { cmdsBuilder_.remove(index); } return this; } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder getCmdsBuilder( int index) { return getCmdsFieldBuilder().getBuilder(index); } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdsOrBuilder( int index) { if (cmdsBuilder_ == null) { return cmds_.get(index); } else { return cmdsBuilder_.getMessageOrBuilder(index); } } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public java.util.List getCmdsOrBuilderList() { if (cmdsBuilder_ != null) { return cmdsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(cmds_); } } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder addCmdsBuilder() { return getCmdsFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance()); } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder addCmdsBuilder( int index) { return getCmdsFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance()); } /** *
       * Returned commands can be null
       * 
* * repeated .hadoop.hdfs.datanode.DatanodeCommandProto cmds = 1; */ public java.util.List getCmdsBuilderList() { return getCmdsFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder> getCmdsFieldBuilder() { if (cmdsBuilder_ == null) { cmdsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder>( cmds_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); cmds_ = null; } return cmdsBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto haStatus_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProtoOrBuilder> haStatusBuilder_; /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public boolean hasHaStatus() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto getHaStatus() { if (haStatusBuilder_ == null) { return haStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.getDefaultInstance() : haStatus_; } else { return haStatusBuilder_.getMessage(); } } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public Builder setHaStatus(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto value) { if (haStatusBuilder_ == null) { if (value == null) { throw new NullPointerException(); } haStatus_ = value; onChanged(); } else { haStatusBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public Builder setHaStatus( org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.Builder builderForValue) { if (haStatusBuilder_ == null) { haStatus_ = builderForValue.build(); onChanged(); } else { haStatusBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public Builder mergeHaStatus(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto value) { if (haStatusBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && haStatus_ != null && haStatus_ != org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.getDefaultInstance()) { haStatus_ = org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.newBuilder(haStatus_).mergeFrom(value).buildPartial(); } else { haStatus_ = value; } onChanged(); } else { haStatusBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public Builder clearHaStatus() { if (haStatusBuilder_ == null) { haStatus_ = null; onChanged(); } else { haStatusBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.Builder getHaStatusBuilder() { bitField0_ |= 0x00000002; onChanged(); return getHaStatusFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProtoOrBuilder getHaStatusOrBuilder() { if (haStatusBuilder_ != null) { return haStatusBuilder_.getMessageOrBuilder(); } else { return haStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.getDefaultInstance() : haStatus_; } } /** * required .hadoop.hdfs.NNHAStatusHeartbeatProto haStatus = 2; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProtoOrBuilder> getHaStatusFieldBuilder() { if (haStatusBuilder_ == null) { haStatusBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NNHAStatusHeartbeatProtoOrBuilder>( getHaStatus(), getParentForChildren(), isClean()); haStatus_ = null; } return haStatusBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto rollingUpgradeStatus_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder> rollingUpgradeStatusBuilder_; /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public boolean hasRollingUpgradeStatus() { return ((bitField0_ & 0x00000004) != 0); } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto getRollingUpgradeStatus() { if (rollingUpgradeStatusBuilder_ == null) { return rollingUpgradeStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatus_; } else { return rollingUpgradeStatusBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public Builder setRollingUpgradeStatus(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto value) { if (rollingUpgradeStatusBuilder_ == null) { if (value == null) { throw new NullPointerException(); } rollingUpgradeStatus_ = value; onChanged(); } else { rollingUpgradeStatusBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public Builder setRollingUpgradeStatus( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder builderForValue) { if (rollingUpgradeStatusBuilder_ == null) { rollingUpgradeStatus_ = builderForValue.build(); onChanged(); } else { rollingUpgradeStatusBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public Builder mergeRollingUpgradeStatus(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto value) { if (rollingUpgradeStatusBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0) && rollingUpgradeStatus_ != null && rollingUpgradeStatus_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance()) { rollingUpgradeStatus_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.newBuilder(rollingUpgradeStatus_).mergeFrom(value).buildPartial(); } else { rollingUpgradeStatus_ = value; } onChanged(); } else { rollingUpgradeStatusBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public Builder clearRollingUpgradeStatus() { if (rollingUpgradeStatusBuilder_ == null) { rollingUpgradeStatus_ = null; onChanged(); } else { rollingUpgradeStatusBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder getRollingUpgradeStatusBuilder() { bitField0_ |= 0x00000004; onChanged(); return getRollingUpgradeStatusFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder getRollingUpgradeStatusOrBuilder() { if (rollingUpgradeStatusBuilder_ != null) { return rollingUpgradeStatusBuilder_.getMessageOrBuilder(); } else { return rollingUpgradeStatus_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatus_; } } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatus = 3; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder> getRollingUpgradeStatusFieldBuilder() { if (rollingUpgradeStatusBuilder_ == null) { rollingUpgradeStatusBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder>( getRollingUpgradeStatus(), getParentForChildren(), isClean()); rollingUpgradeStatus_ = null; } return rollingUpgradeStatusBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto rollingUpgradeStatusV2_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder> rollingUpgradeStatusV2Builder_; /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public boolean hasRollingUpgradeStatusV2() { return ((bitField0_ & 0x00000008) != 0); } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto getRollingUpgradeStatusV2() { if (rollingUpgradeStatusV2Builder_ == null) { return rollingUpgradeStatusV2_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatusV2_; } else { return rollingUpgradeStatusV2Builder_.getMessage(); } } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public Builder setRollingUpgradeStatusV2(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto value) { if (rollingUpgradeStatusV2Builder_ == null) { if (value == null) { throw new NullPointerException(); } rollingUpgradeStatusV2_ = value; onChanged(); } else { rollingUpgradeStatusV2Builder_.setMessage(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public Builder setRollingUpgradeStatusV2( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder builderForValue) { if (rollingUpgradeStatusV2Builder_ == null) { rollingUpgradeStatusV2_ = builderForValue.build(); onChanged(); } else { rollingUpgradeStatusV2Builder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public Builder mergeRollingUpgradeStatusV2(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto value) { if (rollingUpgradeStatusV2Builder_ == null) { if (((bitField0_ & 0x00000008) != 0) && rollingUpgradeStatusV2_ != null && rollingUpgradeStatusV2_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance()) { rollingUpgradeStatusV2_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.newBuilder(rollingUpgradeStatusV2_).mergeFrom(value).buildPartial(); } else { rollingUpgradeStatusV2_ = value; } onChanged(); } else { rollingUpgradeStatusV2Builder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public Builder clearRollingUpgradeStatusV2() { if (rollingUpgradeStatusV2Builder_ == null) { rollingUpgradeStatusV2_ = null; onChanged(); } else { rollingUpgradeStatusV2Builder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); return this; } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder getRollingUpgradeStatusV2Builder() { bitField0_ |= 0x00000008; onChanged(); return getRollingUpgradeStatusV2FieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder getRollingUpgradeStatusV2OrBuilder() { if (rollingUpgradeStatusV2Builder_ != null) { return rollingUpgradeStatusV2Builder_.getMessageOrBuilder(); } else { return rollingUpgradeStatusV2_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.getDefaultInstance() : rollingUpgradeStatusV2_; } } /** * optional .hadoop.hdfs.RollingUpgradeStatusProto rollingUpgradeStatusV2 = 4; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder> getRollingUpgradeStatusV2FieldBuilder() { if (rollingUpgradeStatusV2Builder_ == null) { rollingUpgradeStatusV2Builder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RollingUpgradeStatusProtoOrBuilder>( getRollingUpgradeStatusV2(), getParentForChildren(), isClean()); rollingUpgradeStatusV2_ = null; } return rollingUpgradeStatusV2Builder_; } private long fullBlockReportLeaseId_ ; /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ public boolean hasFullBlockReportLeaseId() { return ((bitField0_ & 0x00000010) != 0); } /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ public long getFullBlockReportLeaseId() { return fullBlockReportLeaseId_; } /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ public Builder setFullBlockReportLeaseId(long value) { bitField0_ |= 0x00000010; fullBlockReportLeaseId_ = value; onChanged(); return this; } /** * optional uint64 fullBlockReportLeaseId = 5 [default = 0]; */ public Builder clearFullBlockReportLeaseId() { bitField0_ = (bitField0_ & ~0x00000010); fullBlockReportLeaseId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.HeartbeatResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.HeartbeatResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public HeartbeatResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new HeartbeatResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockReportRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockReportRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ boolean hasRegistration(); /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration(); /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder(); /** * required string blockPoolId = 2; */ boolean hasBlockPoolId(); /** * required string blockPoolId = 2; */ java.lang.String getBlockPoolId(); /** * required string blockPoolId = 2; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes(); /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ java.util.List getReportsList(); /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto getReports(int index); /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ int getReportsCount(); /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ java.util.List getReportsOrBuilderList(); /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder getReportsOrBuilder( int index); /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ boolean hasContext(); /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto getContext(); /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder getContextOrBuilder(); } /** *
   **
   * registration - datanode registration information
   * blockPoolID  - block pool ID of the reported blocks
   * blocks       - each block is represented as multiple longs in the array.
   *                first long represents block ID
   *                second long represents length
   *                third long represents gen stamp
   *                fourth long (if under construction) represents replica state
   * context      - An optional field containing information about the context
   *                of this block report.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReportRequestProto} */ public static final class BlockReportRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockReportRequestProto) BlockReportRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockReportRequestProto.newBuilder() to construct. private BlockReportRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockReportRequestProto() { blockPoolId_ = ""; reports_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockReportRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registration_.toBuilder(); } registration_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registration_); registration_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; blockPoolId_ = bs; break; } case 26: { if (!((mutable_bitField0_ & 0x00000004) != 0)) { reports_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000004; } reports_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.PARSER, extensionRegistry)); break; } case 34: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder subBuilder = null; if (((bitField0_ & 0x00000004) != 0)) { subBuilder = context_.toBuilder(); } context_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(context_); context_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { reports_ = java.util.Collections.unmodifiableList(reports_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.Builder.class); } private int bitField0_; public static final int REGISTRATION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } public static final int BLOCKPOOLID_FIELD_NUMBER = 2; private volatile java.lang.Object blockPoolId_; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int REPORTS_FIELD_NUMBER = 3; private java.util.List reports_; /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public java.util.List getReportsList() { return reports_; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public java.util.List getReportsOrBuilderList() { return reports_; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public int getReportsCount() { return reports_.size(); } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto getReports(int index) { return reports_.get(index); } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder getReportsOrBuilder( int index) { return reports_.get(index); } public static final int CONTEXT_FIELD_NUMBER = 4; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto context_; /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public boolean hasContext() { return ((bitField0_ & 0x00000004) != 0); } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto getContext() { return context_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance() : context_; } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder getContextOrBuilder() { return context_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance() : context_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistration()) { memoizedIsInitialized = 0; return false; } if (!hasBlockPoolId()) { memoizedIsInitialized = 0; return false; } if (!getRegistration().isInitialized()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getReportsCount(); i++) { if (!getReports(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasContext()) { if (!getContext().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistration()); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, blockPoolId_); } for (int i = 0; i < reports_.size(); i++) { output.writeMessage(3, reports_.get(i)); } if (((bitField0_ & 0x00000004) != 0)) { output.writeMessage(4, getContext()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistration()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, blockPoolId_); } for (int i = 0; i < reports_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, reports_.get(i)); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(4, getContext()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto) obj; if (hasRegistration() != other.hasRegistration()) return false; if (hasRegistration()) { if (!getRegistration() .equals(other.getRegistration())) return false; } if (hasBlockPoolId() != other.hasBlockPoolId()) return false; if (hasBlockPoolId()) { if (!getBlockPoolId() .equals(other.getBlockPoolId())) return false; } if (!getReportsList() .equals(other.getReportsList())) return false; if (hasContext() != other.hasContext()) return false; if (hasContext()) { if (!getContext() .equals(other.getContext())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistration()) { hash = (37 * hash) + REGISTRATION_FIELD_NUMBER; hash = (53 * hash) + getRegistration().hashCode(); } if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } if (getReportsCount() > 0) { hash = (37 * hash) + REPORTS_FIELD_NUMBER; hash = (53 * hash) + getReportsList().hashCode(); } if (hasContext()) { hash = (37 * hash) + CONTEXT_FIELD_NUMBER; hash = (53 * hash) + getContext().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registration - datanode registration information
     * blockPoolID  - block pool ID of the reported blocks
     * blocks       - each block is represented as multiple longs in the array.
     *                first long represents block ID
     *                second long represents length
     *                third long represents gen stamp
     *                fourth long (if under construction) represents replica state
     * context      - An optional field containing information about the context
     *                of this block report.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReportRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockReportRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistrationFieldBuilder(); getReportsFieldBuilder(); getContextFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registrationBuilder_ == null) { registration_ = null; } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); if (reportsBuilder_ == null) { reports_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); } else { reportsBuilder_.clear(); } if (contextBuilder_ == null) { context_ = null; } else { contextBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registrationBuilder_ == null) { result.registration_ = registration_; } else { result.registration_ = registrationBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.blockPoolId_ = blockPoolId_; if (reportsBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0)) { reports_ = java.util.Collections.unmodifiableList(reports_); bitField0_ = (bitField0_ & ~0x00000004); } result.reports_ = reports_; } else { result.reports_ = reportsBuilder_.build(); } if (((from_bitField0_ & 0x00000008) != 0)) { if (contextBuilder_ == null) { result.context_ = context_; } else { result.context_ = contextBuilder_.build(); } to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.getDefaultInstance()) return this; if (other.hasRegistration()) { mergeRegistration(other.getRegistration()); } if (other.hasBlockPoolId()) { bitField0_ |= 0x00000002; blockPoolId_ = other.blockPoolId_; onChanged(); } if (reportsBuilder_ == null) { if (!other.reports_.isEmpty()) { if (reports_.isEmpty()) { reports_ = other.reports_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureReportsIsMutable(); reports_.addAll(other.reports_); } onChanged(); } } else { if (!other.reports_.isEmpty()) { if (reportsBuilder_.isEmpty()) { reportsBuilder_.dispose(); reportsBuilder_ = null; reports_ = other.reports_; bitField0_ = (bitField0_ & ~0x00000004); reportsBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getReportsFieldBuilder() : null; } else { reportsBuilder_.addAllMessages(other.reports_); } } } if (other.hasContext()) { mergeContext(other.getContext()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistration()) { return false; } if (!hasBlockPoolId()) { return false; } if (!getRegistration().isInitialized()) { return false; } for (int i = 0; i < getReportsCount(); i++) { if (!getReports(i).isInitialized()) { return false; } } if (hasContext()) { if (!getContext().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registrationBuilder_; /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { if (registrationBuilder_ == null) { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } else { return registrationBuilder_.getMessage(); } } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registration_ = value; onChanged(); } else { registrationBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registrationBuilder_ == null) { registration_ = builderForValue.build(); onChanged(); } else { registrationBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder mergeRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registration_ != null && registration_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registration_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registration_).mergeFrom(value).buildPartial(); } else { registration_ = value; } onChanged(); } else { registrationBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder clearRegistration() { if (registrationBuilder_ == null) { registration_ = null; onChanged(); } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistrationBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistrationFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { if (registrationBuilder_ != null) { return registrationBuilder_.getMessageOrBuilder(); } else { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistrationFieldBuilder() { if (registrationBuilder_ == null) { registrationBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistration(), getParentForChildren(), isClean()); registration_ = null; } return registrationBuilder_; } private java.lang.Object blockPoolId_ = ""; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * required string blockPoolId = 2; */ public Builder setBlockPoolId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000002); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder setBlockPoolIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } private java.util.List reports_ = java.util.Collections.emptyList(); private void ensureReportsIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { reports_ = new java.util.ArrayList(reports_); bitField0_ |= 0x00000004; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder> reportsBuilder_; /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public java.util.List getReportsList() { if (reportsBuilder_ == null) { return java.util.Collections.unmodifiableList(reports_); } else { return reportsBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public int getReportsCount() { if (reportsBuilder_ == null) { return reports_.size(); } else { return reportsBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto getReports(int index) { if (reportsBuilder_ == null) { return reports_.get(index); } else { return reportsBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder setReports( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto value) { if (reportsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureReportsIsMutable(); reports_.set(index, value); onChanged(); } else { reportsBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder setReports( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder builderForValue) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.set(index, builderForValue.build()); onChanged(); } else { reportsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder addReports(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto value) { if (reportsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureReportsIsMutable(); reports_.add(value); onChanged(); } else { reportsBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder addReports( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto value) { if (reportsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureReportsIsMutable(); reports_.add(index, value); onChanged(); } else { reportsBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder addReports( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder builderForValue) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.add(builderForValue.build()); onChanged(); } else { reportsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder addReports( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder builderForValue) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.add(index, builderForValue.build()); onChanged(); } else { reportsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder addAllReports( java.lang.Iterable values) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, reports_); onChanged(); } else { reportsBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder clearReports() { if (reportsBuilder_ == null) { reports_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); } else { reportsBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public Builder removeReports(int index) { if (reportsBuilder_ == null) { ensureReportsIsMutable(); reports_.remove(index); onChanged(); } else { reportsBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder getReportsBuilder( int index) { return getReportsFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder getReportsOrBuilder( int index) { if (reportsBuilder_ == null) { return reports_.get(index); } else { return reportsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public java.util.List getReportsOrBuilderList() { if (reportsBuilder_ != null) { return reportsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(reports_); } } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder addReportsBuilder() { return getReportsFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder addReportsBuilder( int index) { return getReportsFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.StorageBlockReportProto reports = 3; */ public java.util.List getReportsBuilderList() { return getReportsFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder> getReportsFieldBuilder() { if (reportsBuilder_ == null) { reportsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder>( reports_, ((bitField0_ & 0x00000004) != 0), getParentForChildren(), isClean()); reports_ = null; } return reportsBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto context_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder> contextBuilder_; /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public boolean hasContext() { return ((bitField0_ & 0x00000008) != 0); } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto getContext() { if (contextBuilder_ == null) { return context_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance() : context_; } else { return contextBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public Builder setContext(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto value) { if (contextBuilder_ == null) { if (value == null) { throw new NullPointerException(); } context_ = value; onChanged(); } else { contextBuilder_.setMessage(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public Builder setContext( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder builderForValue) { if (contextBuilder_ == null) { context_ = builderForValue.build(); onChanged(); } else { contextBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public Builder mergeContext(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto value) { if (contextBuilder_ == null) { if (((bitField0_ & 0x00000008) != 0) && context_ != null && context_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance()) { context_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.newBuilder(context_).mergeFrom(value).buildPartial(); } else { context_ = value; } onChanged(); } else { contextBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public Builder clearContext() { if (contextBuilder_ == null) { context_ = null; onChanged(); } else { contextBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); return this; } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder getContextBuilder() { bitField0_ |= 0x00000008; onChanged(); return getContextFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder getContextOrBuilder() { if (contextBuilder_ != null) { return contextBuilder_.getMessageOrBuilder(); } else { return context_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance() : context_; } } /** * optional .hadoop.hdfs.datanode.BlockReportContextProto context = 4; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder> getContextFieldBuilder() { if (contextBuilder_ == null) { contextBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder>( getContext(), getParentForChildren(), isClean()); context_ = null; } return contextBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockReportRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockReportRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockReportRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockReportRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockReportContextProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockReportContextProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * The total number of RPCs this block report is broken into.
     * 
* * required int32 totalRpcs = 1; */ boolean hasTotalRpcs(); /** *
     * The total number of RPCs this block report is broken into.
     * 
* * required int32 totalRpcs = 1; */ int getTotalRpcs(); /** *
     * The index of the current RPC (zero-based)
     * 
* * required int32 curRpc = 2; */ boolean hasCurRpc(); /** *
     * The index of the current RPC (zero-based)
     * 
* * required int32 curRpc = 2; */ int getCurRpc(); /** *
     * The unique 64-bit ID of this block report
     * 
* * required int64 id = 3; */ boolean hasId(); /** *
     * The unique 64-bit ID of this block report
     * 
* * required int64 id = 3; */ long getId(); /** *
     * The block report lease ID, or 0 if we are sending without a lease to
     * bypass rate-limiting.
     * 
* * optional uint64 leaseId = 4 [default = 0]; */ boolean hasLeaseId(); /** *
     * The block report lease ID, or 0 if we are sending without a lease to
     * bypass rate-limiting.
     * 
* * optional uint64 leaseId = 4 [default = 0]; */ long getLeaseId(); /** *
     * True if the reported blocks are sorted by increasing block IDs
     * 
* * optional bool sorted = 5 [default = false]; */ boolean hasSorted(); /** *
     * True if the reported blocks are sorted by increasing block IDs
     * 
* * optional bool sorted = 5 [default = false]; */ boolean getSorted(); } /** * Protobuf type {@code hadoop.hdfs.datanode.BlockReportContextProto} */ public static final class BlockReportContextProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockReportContextProto) BlockReportContextProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockReportContextProto.newBuilder() to construct. private BlockReportContextProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockReportContextProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockReportContextProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; totalRpcs_ = input.readInt32(); break; } case 16: { bitField0_ |= 0x00000002; curRpc_ = input.readInt32(); break; } case 24: { bitField0_ |= 0x00000004; id_ = input.readInt64(); break; } case 32: { bitField0_ |= 0x00000008; leaseId_ = input.readUInt64(); break; } case 40: { bitField0_ |= 0x00000010; sorted_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportContextProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportContextProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder.class); } private int bitField0_; public static final int TOTALRPCS_FIELD_NUMBER = 1; private int totalRpcs_; /** *
     * The total number of RPCs this block report is broken into.
     * 
* * required int32 totalRpcs = 1; */ public boolean hasTotalRpcs() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * The total number of RPCs this block report is broken into.
     * 
* * required int32 totalRpcs = 1; */ public int getTotalRpcs() { return totalRpcs_; } public static final int CURRPC_FIELD_NUMBER = 2; private int curRpc_; /** *
     * The index of the current RPC (zero-based)
     * 
* * required int32 curRpc = 2; */ public boolean hasCurRpc() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * The index of the current RPC (zero-based)
     * 
* * required int32 curRpc = 2; */ public int getCurRpc() { return curRpc_; } public static final int ID_FIELD_NUMBER = 3; private long id_; /** *
     * The unique 64-bit ID of this block report
     * 
* * required int64 id = 3; */ public boolean hasId() { return ((bitField0_ & 0x00000004) != 0); } /** *
     * The unique 64-bit ID of this block report
     * 
* * required int64 id = 3; */ public long getId() { return id_; } public static final int LEASEID_FIELD_NUMBER = 4; private long leaseId_; /** *
     * The block report lease ID, or 0 if we are sending without a lease to
     * bypass rate-limiting.
     * 
* * optional uint64 leaseId = 4 [default = 0]; */ public boolean hasLeaseId() { return ((bitField0_ & 0x00000008) != 0); } /** *
     * The block report lease ID, or 0 if we are sending without a lease to
     * bypass rate-limiting.
     * 
* * optional uint64 leaseId = 4 [default = 0]; */ public long getLeaseId() { return leaseId_; } public static final int SORTED_FIELD_NUMBER = 5; private boolean sorted_; /** *
     * True if the reported blocks are sorted by increasing block IDs
     * 
* * optional bool sorted = 5 [default = false]; */ public boolean hasSorted() { return ((bitField0_ & 0x00000010) != 0); } /** *
     * True if the reported blocks are sorted by increasing block IDs
     * 
* * optional bool sorted = 5 [default = false]; */ public boolean getSorted() { return sorted_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTotalRpcs()) { memoizedIsInitialized = 0; return false; } if (!hasCurRpc()) { memoizedIsInitialized = 0; return false; } if (!hasId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeInt32(1, totalRpcs_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeInt32(2, curRpc_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeInt64(3, id_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeUInt64(4, leaseId_); } if (((bitField0_ & 0x00000010) != 0)) { output.writeBool(5, sorted_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeInt32Size(1, totalRpcs_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeInt32Size(2, curRpc_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeInt64Size(3, id_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(4, leaseId_); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeBoolSize(5, sorted_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto) obj; if (hasTotalRpcs() != other.hasTotalRpcs()) return false; if (hasTotalRpcs()) { if (getTotalRpcs() != other.getTotalRpcs()) return false; } if (hasCurRpc() != other.hasCurRpc()) return false; if (hasCurRpc()) { if (getCurRpc() != other.getCurRpc()) return false; } if (hasId() != other.hasId()) return false; if (hasId()) { if (getId() != other.getId()) return false; } if (hasLeaseId() != other.hasLeaseId()) return false; if (hasLeaseId()) { if (getLeaseId() != other.getLeaseId()) return false; } if (hasSorted() != other.hasSorted()) return false; if (hasSorted()) { if (getSorted() != other.getSorted()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTotalRpcs()) { hash = (37 * hash) + TOTALRPCS_FIELD_NUMBER; hash = (53 * hash) + getTotalRpcs(); } if (hasCurRpc()) { hash = (37 * hash) + CURRPC_FIELD_NUMBER; hash = (53 * hash) + getCurRpc(); } if (hasId()) { hash = (37 * hash) + ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getId()); } if (hasLeaseId()) { hash = (37 * hash) + LEASEID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getLeaseId()); } if (hasSorted()) { hash = (37 * hash) + SORTED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean( getSorted()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hadoop.hdfs.datanode.BlockReportContextProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockReportContextProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportContextProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportContextProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); totalRpcs_ = 0; bitField0_ = (bitField0_ & ~0x00000001); curRpc_ = 0; bitField0_ = (bitField0_ & ~0x00000002); id_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); leaseId_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); sorted_ = false; bitField0_ = (bitField0_ & ~0x00000010); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportContextProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.totalRpcs_ = totalRpcs_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.curRpc_ = curRpc_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.id_ = id_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.leaseId_ = leaseId_; to_bitField0_ |= 0x00000008; } if (((from_bitField0_ & 0x00000010) != 0)) { result.sorted_ = sorted_; to_bitField0_ |= 0x00000010; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto.getDefaultInstance()) return this; if (other.hasTotalRpcs()) { setTotalRpcs(other.getTotalRpcs()); } if (other.hasCurRpc()) { setCurRpc(other.getCurRpc()); } if (other.hasId()) { setId(other.getId()); } if (other.hasLeaseId()) { setLeaseId(other.getLeaseId()); } if (other.hasSorted()) { setSorted(other.getSorted()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTotalRpcs()) { return false; } if (!hasCurRpc()) { return false; } if (!hasId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int totalRpcs_ ; /** *
       * The total number of RPCs this block report is broken into.
       * 
* * required int32 totalRpcs = 1; */ public boolean hasTotalRpcs() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * The total number of RPCs this block report is broken into.
       * 
* * required int32 totalRpcs = 1; */ public int getTotalRpcs() { return totalRpcs_; } /** *
       * The total number of RPCs this block report is broken into.
       * 
* * required int32 totalRpcs = 1; */ public Builder setTotalRpcs(int value) { bitField0_ |= 0x00000001; totalRpcs_ = value; onChanged(); return this; } /** *
       * The total number of RPCs this block report is broken into.
       * 
* * required int32 totalRpcs = 1; */ public Builder clearTotalRpcs() { bitField0_ = (bitField0_ & ~0x00000001); totalRpcs_ = 0; onChanged(); return this; } private int curRpc_ ; /** *
       * The index of the current RPC (zero-based)
       * 
* * required int32 curRpc = 2; */ public boolean hasCurRpc() { return ((bitField0_ & 0x00000002) != 0); } /** *
       * The index of the current RPC (zero-based)
       * 
* * required int32 curRpc = 2; */ public int getCurRpc() { return curRpc_; } /** *
       * The index of the current RPC (zero-based)
       * 
* * required int32 curRpc = 2; */ public Builder setCurRpc(int value) { bitField0_ |= 0x00000002; curRpc_ = value; onChanged(); return this; } /** *
       * The index of the current RPC (zero-based)
       * 
* * required int32 curRpc = 2; */ public Builder clearCurRpc() { bitField0_ = (bitField0_ & ~0x00000002); curRpc_ = 0; onChanged(); return this; } private long id_ ; /** *
       * The unique 64-bit ID of this block report
       * 
* * required int64 id = 3; */ public boolean hasId() { return ((bitField0_ & 0x00000004) != 0); } /** *
       * The unique 64-bit ID of this block report
       * 
* * required int64 id = 3; */ public long getId() { return id_; } /** *
       * The unique 64-bit ID of this block report
       * 
* * required int64 id = 3; */ public Builder setId(long value) { bitField0_ |= 0x00000004; id_ = value; onChanged(); return this; } /** *
       * The unique 64-bit ID of this block report
       * 
* * required int64 id = 3; */ public Builder clearId() { bitField0_ = (bitField0_ & ~0x00000004); id_ = 0L; onChanged(); return this; } private long leaseId_ ; /** *
       * The block report lease ID, or 0 if we are sending without a lease to
       * bypass rate-limiting.
       * 
* * optional uint64 leaseId = 4 [default = 0]; */ public boolean hasLeaseId() { return ((bitField0_ & 0x00000008) != 0); } /** *
       * The block report lease ID, or 0 if we are sending without a lease to
       * bypass rate-limiting.
       * 
* * optional uint64 leaseId = 4 [default = 0]; */ public long getLeaseId() { return leaseId_; } /** *
       * The block report lease ID, or 0 if we are sending without a lease to
       * bypass rate-limiting.
       * 
* * optional uint64 leaseId = 4 [default = 0]; */ public Builder setLeaseId(long value) { bitField0_ |= 0x00000008; leaseId_ = value; onChanged(); return this; } /** *
       * The block report lease ID, or 0 if we are sending without a lease to
       * bypass rate-limiting.
       * 
* * optional uint64 leaseId = 4 [default = 0]; */ public Builder clearLeaseId() { bitField0_ = (bitField0_ & ~0x00000008); leaseId_ = 0L; onChanged(); return this; } private boolean sorted_ ; /** *
       * True if the reported blocks are sorted by increasing block IDs
       * 
* * optional bool sorted = 5 [default = false]; */ public boolean hasSorted() { return ((bitField0_ & 0x00000010) != 0); } /** *
       * True if the reported blocks are sorted by increasing block IDs
       * 
* * optional bool sorted = 5 [default = false]; */ public boolean getSorted() { return sorted_; } /** *
       * True if the reported blocks are sorted by increasing block IDs
       * 
* * optional bool sorted = 5 [default = false]; */ public Builder setSorted(boolean value) { bitField0_ |= 0x00000010; sorted_ = value; onChanged(); return this; } /** *
       * True if the reported blocks are sorted by increasing block IDs
       * 
* * optional bool sorted = 5 [default = false]; */ public Builder clearSorted() { bitField0_ = (bitField0_ & ~0x00000010); sorted_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockReportContextProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockReportContextProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockReportContextProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockReportContextProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportContextProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface StorageBlockReportProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.StorageBlockReportProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Storage
     * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ boolean hasStorage(); /** *
     * Storage
     * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto getStorage(); /** *
     * Storage
     * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder getStorageOrBuilder(); /** * repeated uint64 blocks = 2 [packed = true]; */ java.util.List getBlocksList(); /** * repeated uint64 blocks = 2 [packed = true]; */ int getBlocksCount(); /** * repeated uint64 blocks = 2 [packed = true]; */ long getBlocks(int index); /** * optional uint64 numberOfBlocks = 3; */ boolean hasNumberOfBlocks(); /** * optional uint64 numberOfBlocks = 3; */ long getNumberOfBlocks(); /** * repeated bytes blocksBuffers = 4; */ java.util.List getBlocksBuffersList(); /** * repeated bytes blocksBuffers = 4; */ int getBlocksBuffersCount(); /** * repeated bytes blocksBuffers = 4; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlocksBuffers(int index); } /** *
   **
   * Report of blocks in a storage
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.StorageBlockReportProto} */ public static final class StorageBlockReportProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.StorageBlockReportProto) StorageBlockReportProtoOrBuilder { private static final long serialVersionUID = 0L; // Use StorageBlockReportProto.newBuilder() to construct. private StorageBlockReportProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private StorageBlockReportProto() { blocks_ = emptyLongList(); blocksBuffers_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private StorageBlockReportProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = storage_.toBuilder(); } storage_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(storage_); storage_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { if (!((mutable_bitField0_ & 0x00000002) != 0)) { blocks_ = newLongList(); mutable_bitField0_ |= 0x00000002; } blocks_.addLong(input.readUInt64()); break; } case 18: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000002) != 0) && input.getBytesUntilLimit() > 0) { blocks_ = newLongList(); mutable_bitField0_ |= 0x00000002; } while (input.getBytesUntilLimit() > 0) { blocks_.addLong(input.readUInt64()); } input.popLimit(limit); break; } case 24: { bitField0_ |= 0x00000002; numberOfBlocks_ = input.readUInt64(); break; } case 34: { if (!((mutable_bitField0_ & 0x00000008) != 0)) { blocksBuffers_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000008; } blocksBuffers_.add(input.readBytes()); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) != 0)) { blocks_.makeImmutable(); // C } if (((mutable_bitField0_ & 0x00000008) != 0)) { blocksBuffers_ = java.util.Collections.unmodifiableList(blocksBuffers_); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder.class); } private int bitField0_; public static final int STORAGE_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto storage_; /** *
     * Storage
     * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public boolean hasStorage() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Storage
     * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto getStorage() { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } /** *
     * Storage
     * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder getStorageOrBuilder() { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } public static final int BLOCKS_FIELD_NUMBER = 2; private org.apache.hadoop.thirdparty.protobuf.Internal.LongList blocks_; /** * repeated uint64 blocks = 2 [packed = true]; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated uint64 blocks = 2 [packed = true]; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated uint64 blocks = 2 [packed = true]; */ public long getBlocks(int index) { return blocks_.getLong(index); } private int blocksMemoizedSerializedSize = -1; public static final int NUMBEROFBLOCKS_FIELD_NUMBER = 3; private long numberOfBlocks_; /** * optional uint64 numberOfBlocks = 3; */ public boolean hasNumberOfBlocks() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 numberOfBlocks = 3; */ public long getNumberOfBlocks() { return numberOfBlocks_; } public static final int BLOCKSBUFFERS_FIELD_NUMBER = 4; private java.util.List blocksBuffers_; /** * repeated bytes blocksBuffers = 4; */ public java.util.List getBlocksBuffersList() { return blocksBuffers_; } /** * repeated bytes blocksBuffers = 4; */ public int getBlocksBuffersCount() { return blocksBuffers_.size(); } /** * repeated bytes blocksBuffers = 4; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlocksBuffers(int index) { return blocksBuffers_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasStorage()) { memoizedIsInitialized = 0; return false; } if (!getStorage().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getStorage()); } if (getBlocksList().size() > 0) { output.writeUInt32NoTag(18); output.writeUInt32NoTag(blocksMemoizedSerializedSize); } for (int i = 0; i < blocks_.size(); i++) { output.writeUInt64NoTag(blocks_.getLong(i)); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(3, numberOfBlocks_); } for (int i = 0; i < blocksBuffers_.size(); i++) { output.writeBytes(4, blocksBuffers_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getStorage()); } { int dataSize = 0; for (int i = 0; i < blocks_.size(); i++) { dataSize += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64SizeNoTag(blocks_.getLong(i)); } size += dataSize; if (!getBlocksList().isEmpty()) { size += 1; size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeInt32SizeNoTag(dataSize); } blocksMemoizedSerializedSize = dataSize; } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(3, numberOfBlocks_); } { int dataSize = 0; for (int i = 0; i < blocksBuffers_.size(); i++) { dataSize += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeBytesSizeNoTag(blocksBuffers_.get(i)); } size += dataSize; size += 1 * getBlocksBuffersList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto) obj; if (hasStorage() != other.hasStorage()) return false; if (hasStorage()) { if (!getStorage() .equals(other.getStorage())) return false; } if (!getBlocksList() .equals(other.getBlocksList())) return false; if (hasNumberOfBlocks() != other.hasNumberOfBlocks()) return false; if (hasNumberOfBlocks()) { if (getNumberOfBlocks() != other.getNumberOfBlocks()) return false; } if (!getBlocksBuffersList() .equals(other.getBlocksBuffersList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasStorage()) { hash = (37 * hash) + STORAGE_FIELD_NUMBER; hash = (53 * hash) + getStorage().hashCode(); } if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } if (hasNumberOfBlocks()) { hash = (37 * hash) + NUMBEROFBLOCKS_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getNumberOfBlocks()); } if (getBlocksBuffersCount() > 0) { hash = (37 * hash) + BLOCKSBUFFERS_FIELD_NUMBER; hash = (53 * hash) + getBlocksBuffersList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Report of blocks in a storage
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.StorageBlockReportProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.StorageBlockReportProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getStorageFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (storageBuilder_ == null) { storage_ = null; } else { storageBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); blocks_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000002); numberOfBlocks_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); blocksBuffers_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (storageBuilder_ == null) { result.storage_ = storage_; } else { result.storage_ = storageBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((bitField0_ & 0x00000002) != 0)) { blocks_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000002); } result.blocks_ = blocks_; if (((from_bitField0_ & 0x00000004) != 0)) { result.numberOfBlocks_ = numberOfBlocks_; to_bitField0_ |= 0x00000002; } if (((bitField0_ & 0x00000008) != 0)) { blocksBuffers_ = java.util.Collections.unmodifiableList(blocksBuffers_); bitField0_ = (bitField0_ & ~0x00000008); } result.blocksBuffers_ = blocksBuffers_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto.getDefaultInstance()) return this; if (other.hasStorage()) { mergeStorage(other.getStorage()); } if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } if (other.hasNumberOfBlocks()) { setNumberOfBlocks(other.getNumberOfBlocks()); } if (!other.blocksBuffers_.isEmpty()) { if (blocksBuffers_.isEmpty()) { blocksBuffers_ = other.blocksBuffers_; bitField0_ = (bitField0_ & ~0x00000008); } else { ensureBlocksBuffersIsMutable(); blocksBuffers_.addAll(other.blocksBuffers_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasStorage()) { return false; } if (!getStorage().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto storage_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder> storageBuilder_; /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public boolean hasStorage() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto getStorage() { if (storageBuilder_ == null) { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } else { return storageBuilder_.getMessage(); } } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public Builder setStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto value) { if (storageBuilder_ == null) { if (value == null) { throw new NullPointerException(); } storage_ = value; onChanged(); } else { storageBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public Builder setStorage( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder builderForValue) { if (storageBuilder_ == null) { storage_ = builderForValue.build(); onChanged(); } else { storageBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public Builder mergeStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto value) { if (storageBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && storage_ != null && storage_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance()) { storage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.newBuilder(storage_).mergeFrom(value).buildPartial(); } else { storage_ = value; } onChanged(); } else { storageBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public Builder clearStorage() { if (storageBuilder_ == null) { storage_ = null; onChanged(); } else { storageBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder getStorageBuilder() { bitField0_ |= 0x00000001; onChanged(); return getStorageFieldBuilder().getBuilder(); } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder getStorageOrBuilder() { if (storageBuilder_ != null) { return storageBuilder_.getMessageOrBuilder(); } else { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } } /** *
       * Storage
       * 
* * required .hadoop.hdfs.DatanodeStorageProto storage = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder> getStorageFieldBuilder() { if (storageBuilder_ == null) { storageBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder>( getStorage(), getParentForChildren(), isClean()); storage_ = null; } return storageBuilder_; } private org.apache.hadoop.thirdparty.protobuf.Internal.LongList blocks_ = emptyLongList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000002) != 0)) { blocks_ = mutableCopy(blocks_); bitField0_ |= 0x00000002; } } /** * repeated uint64 blocks = 2 [packed = true]; */ public java.util.List getBlocksList() { return ((bitField0_ & 0x00000002) != 0) ? java.util.Collections.unmodifiableList(blocks_) : blocks_; } /** * repeated uint64 blocks = 2 [packed = true]; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated uint64 blocks = 2 [packed = true]; */ public long getBlocks(int index) { return blocks_.getLong(index); } /** * repeated uint64 blocks = 2 [packed = true]; */ public Builder setBlocks( int index, long value) { ensureBlocksIsMutable(); blocks_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 blocks = 2 [packed = true]; */ public Builder addBlocks(long value) { ensureBlocksIsMutable(); blocks_.addLong(value); onChanged(); return this; } /** * repeated uint64 blocks = 2 [packed = true]; */ public Builder addAllBlocks( java.lang.Iterable values) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); return this; } /** * repeated uint64 blocks = 2 [packed = true]; */ public Builder clearBlocks() { blocks_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); return this; } private long numberOfBlocks_ ; /** * optional uint64 numberOfBlocks = 3; */ public boolean hasNumberOfBlocks() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 numberOfBlocks = 3; */ public long getNumberOfBlocks() { return numberOfBlocks_; } /** * optional uint64 numberOfBlocks = 3; */ public Builder setNumberOfBlocks(long value) { bitField0_ |= 0x00000004; numberOfBlocks_ = value; onChanged(); return this; } /** * optional uint64 numberOfBlocks = 3; */ public Builder clearNumberOfBlocks() { bitField0_ = (bitField0_ & ~0x00000004); numberOfBlocks_ = 0L; onChanged(); return this; } private java.util.List blocksBuffers_ = java.util.Collections.emptyList(); private void ensureBlocksBuffersIsMutable() { if (!((bitField0_ & 0x00000008) != 0)) { blocksBuffers_ = new java.util.ArrayList(blocksBuffers_); bitField0_ |= 0x00000008; } } /** * repeated bytes blocksBuffers = 4; */ public java.util.List getBlocksBuffersList() { return ((bitField0_ & 0x00000008) != 0) ? java.util.Collections.unmodifiableList(blocksBuffers_) : blocksBuffers_; } /** * repeated bytes blocksBuffers = 4; */ public int getBlocksBuffersCount() { return blocksBuffers_.size(); } /** * repeated bytes blocksBuffers = 4; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlocksBuffers(int index) { return blocksBuffers_.get(index); } /** * repeated bytes blocksBuffers = 4; */ public Builder setBlocksBuffers( int index, org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureBlocksBuffersIsMutable(); blocksBuffers_.set(index, value); onChanged(); return this; } /** * repeated bytes blocksBuffers = 4; */ public Builder addBlocksBuffers(org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureBlocksBuffersIsMutable(); blocksBuffers_.add(value); onChanged(); return this; } /** * repeated bytes blocksBuffers = 4; */ public Builder addAllBlocksBuffers( java.lang.Iterable values) { ensureBlocksBuffersIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocksBuffers_); onChanged(); return this; } /** * repeated bytes blocksBuffers = 4; */ public Builder clearBlocksBuffers() { blocksBuffers_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.StorageBlockReportProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.StorageBlockReportProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public StorageBlockReportProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new StorageBlockReportProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageBlockReportProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockReportResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockReportResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ boolean hasCmd(); /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmd(); /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdOrBuilder(); } /** *
   **
   * cmd - Command from namenode to the datanode
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReportResponseProto} */ public static final class BlockReportResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockReportResponseProto) BlockReportResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockReportResponseProto.newBuilder() to construct. private BlockReportResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockReportResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockReportResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = cmd_.toBuilder(); } cmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(cmd_); cmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.Builder.class); } private int bitField0_; public static final int CMD_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto cmd_; /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public boolean hasCmd() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmd() { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdOrBuilder() { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasCmd()) { if (!getCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getCmd()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getCmd()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto) obj; if (hasCmd() != other.hasCmd()) return false; if (hasCmd()) { if (!getCmd() .equals(other.getCmd())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasCmd()) { hash = (37 * hash) + CMD_FIELD_NUMBER; hash = (53 * hash) + getCmd().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * cmd - Command from namenode to the datanode
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReportResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockReportResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getCmdFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (cmdBuilder_ == null) { cmd_ = null; } else { cmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (cmdBuilder_ == null) { result.cmd_ = cmd_; } else { result.cmd_ = cmdBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance()) return this; if (other.hasCmd()) { mergeCmd(other.getCmd()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasCmd()) { if (!getCmd().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto cmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder> cmdBuilder_; /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public boolean hasCmd() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmd() { if (cmdBuilder_ == null) { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } else { return cmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder setCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } cmd_ = value; onChanged(); } else { cmdBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder setCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder builderForValue) { if (cmdBuilder_ == null) { cmd_ = builderForValue.build(); onChanged(); } else { cmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder mergeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && cmd_ != null && cmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance()) { cmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.newBuilder(cmd_).mergeFrom(value).buildPartial(); } else { cmd_ = value; } onChanged(); } else { cmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder clearCmd() { if (cmdBuilder_ == null) { cmd_ = null; onChanged(); } else { cmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder getCmdBuilder() { bitField0_ |= 0x00000001; onChanged(); return getCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdOrBuilder() { if (cmdBuilder_ != null) { return cmdBuilder_.getMessageOrBuilder(); } else { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder> getCmdFieldBuilder() { if (cmdBuilder_ == null) { cmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder>( getCmd(), getParentForChildren(), isClean()); cmd_ = null; } return cmdBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockReportResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockReportResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockReportResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockReportResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CacheReportRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.CacheReportRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ boolean hasRegistration(); /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration(); /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder(); /** * required string blockPoolId = 2; */ boolean hasBlockPoolId(); /** * required string blockPoolId = 2; */ java.lang.String getBlockPoolId(); /** * required string blockPoolId = 2; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes(); /** * repeated uint64 blocks = 3 [packed = true]; */ java.util.List getBlocksList(); /** * repeated uint64 blocks = 3 [packed = true]; */ int getBlocksCount(); /** * repeated uint64 blocks = 3 [packed = true]; */ long getBlocks(int index); } /** *
   **
   * registration - datanode registration information
   * blockPoolId  - block pool ID of the reported blocks
   * blocks       - representation of blocks as longs for efficiency reasons
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.CacheReportRequestProto} */ public static final class CacheReportRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.CacheReportRequestProto) CacheReportRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use CacheReportRequestProto.newBuilder() to construct. private CacheReportRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CacheReportRequestProto() { blockPoolId_ = ""; blocks_ = emptyLongList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CacheReportRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registration_.toBuilder(); } registration_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registration_); registration_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; blockPoolId_ = bs; break; } case 24: { if (!((mutable_bitField0_ & 0x00000004) != 0)) { blocks_ = newLongList(); mutable_bitField0_ |= 0x00000004; } blocks_.addLong(input.readUInt64()); break; } case 26: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000004) != 0) && input.getBytesUntilLimit() > 0) { blocks_ = newLongList(); mutable_bitField0_ |= 0x00000004; } while (input.getBytesUntilLimit() > 0) { blocks_.addLong(input.readUInt64()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { blocks_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.Builder.class); } private int bitField0_; public static final int REGISTRATION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } public static final int BLOCKPOOLID_FIELD_NUMBER = 2; private volatile java.lang.Object blockPoolId_; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int BLOCKS_FIELD_NUMBER = 3; private org.apache.hadoop.thirdparty.protobuf.Internal.LongList blocks_; /** * repeated uint64 blocks = 3 [packed = true]; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated uint64 blocks = 3 [packed = true]; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated uint64 blocks = 3 [packed = true]; */ public long getBlocks(int index) { return blocks_.getLong(index); } private int blocksMemoizedSerializedSize = -1; private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistration()) { memoizedIsInitialized = 0; return false; } if (!hasBlockPoolId()) { memoizedIsInitialized = 0; return false; } if (!getRegistration().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistration()); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, blockPoolId_); } if (getBlocksList().size() > 0) { output.writeUInt32NoTag(26); output.writeUInt32NoTag(blocksMemoizedSerializedSize); } for (int i = 0; i < blocks_.size(); i++) { output.writeUInt64NoTag(blocks_.getLong(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistration()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, blockPoolId_); } { int dataSize = 0; for (int i = 0; i < blocks_.size(); i++) { dataSize += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64SizeNoTag(blocks_.getLong(i)); } size += dataSize; if (!getBlocksList().isEmpty()) { size += 1; size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeInt32SizeNoTag(dataSize); } blocksMemoizedSerializedSize = dataSize; } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto) obj; if (hasRegistration() != other.hasRegistration()) return false; if (hasRegistration()) { if (!getRegistration() .equals(other.getRegistration())) return false; } if (hasBlockPoolId() != other.hasBlockPoolId()) return false; if (hasBlockPoolId()) { if (!getBlockPoolId() .equals(other.getBlockPoolId())) return false; } if (!getBlocksList() .equals(other.getBlocksList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistration()) { hash = (37 * hash) + REGISTRATION_FIELD_NUMBER; hash = (53 * hash) + getRegistration().hashCode(); } if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registration - datanode registration information
     * blockPoolId  - block pool ID of the reported blocks
     * blocks       - representation of blocks as longs for efficiency reasons
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.CacheReportRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.CacheReportRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistrationFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registrationBuilder_ == null) { registration_ = null; } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); blocks_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registrationBuilder_ == null) { result.registration_ = registration_; } else { result.registration_ = registrationBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.blockPoolId_ = blockPoolId_; if (((bitField0_ & 0x00000004) != 0)) { blocks_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000004); } result.blocks_ = blocks_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.getDefaultInstance()) return this; if (other.hasRegistration()) { mergeRegistration(other.getRegistration()); } if (other.hasBlockPoolId()) { bitField0_ |= 0x00000002; blockPoolId_ = other.blockPoolId_; onChanged(); } if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistration()) { return false; } if (!hasBlockPoolId()) { return false; } if (!getRegistration().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registrationBuilder_; /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { if (registrationBuilder_ == null) { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } else { return registrationBuilder_.getMessage(); } } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registration_ = value; onChanged(); } else { registrationBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registrationBuilder_ == null) { registration_ = builderForValue.build(); onChanged(); } else { registrationBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder mergeRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registration_ != null && registration_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registration_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registration_).mergeFrom(value).buildPartial(); } else { registration_ = value; } onChanged(); } else { registrationBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder clearRegistration() { if (registrationBuilder_ == null) { registration_ = null; onChanged(); } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistrationBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistrationFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { if (registrationBuilder_ != null) { return registrationBuilder_.getMessageOrBuilder(); } else { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistrationFieldBuilder() { if (registrationBuilder_ == null) { registrationBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistration(), getParentForChildren(), isClean()); registration_ = null; } return registrationBuilder_; } private java.lang.Object blockPoolId_ = ""; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * required string blockPoolId = 2; */ public Builder setBlockPoolId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000002); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder setBlockPoolIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } private org.apache.hadoop.thirdparty.protobuf.Internal.LongList blocks_ = emptyLongList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { blocks_ = mutableCopy(blocks_); bitField0_ |= 0x00000004; } } /** * repeated uint64 blocks = 3 [packed = true]; */ public java.util.List getBlocksList() { return ((bitField0_ & 0x00000004) != 0) ? java.util.Collections.unmodifiableList(blocks_) : blocks_; } /** * repeated uint64 blocks = 3 [packed = true]; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated uint64 blocks = 3 [packed = true]; */ public long getBlocks(int index) { return blocks_.getLong(index); } /** * repeated uint64 blocks = 3 [packed = true]; */ public Builder setBlocks( int index, long value) { ensureBlocksIsMutable(); blocks_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 blocks = 3 [packed = true]; */ public Builder addBlocks(long value) { ensureBlocksIsMutable(); blocks_.addLong(value); onChanged(); return this; } /** * repeated uint64 blocks = 3 [packed = true]; */ public Builder addAllBlocks( java.lang.Iterable values) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); return this; } /** * repeated uint64 blocks = 3 [packed = true]; */ public Builder clearBlocks() { blocks_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.CacheReportRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.CacheReportRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public CacheReportRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new CacheReportRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CacheReportResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.CacheReportResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ boolean hasCmd(); /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmd(); /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdOrBuilder(); } /** * Protobuf type {@code hadoop.hdfs.datanode.CacheReportResponseProto} */ public static final class CacheReportResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.CacheReportResponseProto) CacheReportResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use CacheReportResponseProto.newBuilder() to construct. private CacheReportResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CacheReportResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CacheReportResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = cmd_.toBuilder(); } cmd_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(cmd_); cmd_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.Builder.class); } private int bitField0_; public static final int CMD_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto cmd_; /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public boolean hasCmd() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmd() { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdOrBuilder() { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasCmd()) { if (!getCmd().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getCmd()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getCmd()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto) obj; if (hasCmd() != other.hasCmd()) return false; if (hasCmd()) { if (!getCmd() .equals(other.getCmd())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasCmd()) { hash = (37 * hash) + CMD_FIELD_NUMBER; hash = (53 * hash) + getCmd().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hadoop.hdfs.datanode.CacheReportResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.CacheReportResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getCmdFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (cmdBuilder_ == null) { cmd_ = null; } else { cmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (cmdBuilder_ == null) { result.cmd_ = cmd_; } else { result.cmd_ = cmdBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance()) return this; if (other.hasCmd()) { mergeCmd(other.getCmd()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasCmd()) { if (!getCmd().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto cmd_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder> cmdBuilder_; /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public boolean hasCmd() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto getCmd() { if (cmdBuilder_ == null) { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } else { return cmdBuilder_.getMessage(); } } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder setCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } cmd_ = value; onChanged(); } else { cmdBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder setCmd( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder builderForValue) { if (cmdBuilder_ == null) { cmd_ = builderForValue.build(); onChanged(); } else { cmdBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder mergeCmd(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto value) { if (cmdBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && cmd_ != null && cmd_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance()) { cmd_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.newBuilder(cmd_).mergeFrom(value).buildPartial(); } else { cmd_ = value; } onChanged(); } else { cmdBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public Builder clearCmd() { if (cmdBuilder_ == null) { cmd_ = null; onChanged(); } else { cmdBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder getCmdBuilder() { bitField0_ |= 0x00000001; onChanged(); return getCmdFieldBuilder().getBuilder(); } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder getCmdOrBuilder() { if (cmdBuilder_ != null) { return cmdBuilder_.getMessageOrBuilder(); } else { return cmd_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.getDefaultInstance() : cmd_; } } /** * optional .hadoop.hdfs.datanode.DatanodeCommandProto cmd = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder> getCmdFieldBuilder() { if (cmdBuilder_ == null) { cmdBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeCommandProtoOrBuilder>( getCmd(), getParentForChildren(), isClean()); cmd_ = null; } return cmdBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.CacheReportResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.CacheReportResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public CacheReportResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new CacheReportResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ReceivedDeletedBlockInfoProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.BlockProto block = 1; */ boolean hasBlock(); /** * required .hadoop.hdfs.BlockProto block = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlock(); /** * required .hadoop.hdfs.BlockProto block = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlockOrBuilder(); /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ boolean hasStatus(); /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus getStatus(); /** * optional string deleteHint = 2; */ boolean hasDeleteHint(); /** * optional string deleteHint = 2; */ java.lang.String getDeleteHint(); /** * optional string deleteHint = 2; */ org.apache.hadoop.thirdparty.protobuf.ByteString getDeleteHintBytes(); } /** *
   **
   * Data structure to send received or deleted block information
   * from datanode to namenode.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto} */ public static final class ReceivedDeletedBlockInfoProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto) ReceivedDeletedBlockInfoProtoOrBuilder { private static final long serialVersionUID = 0L; // Use ReceivedDeletedBlockInfoProto.newBuilder() to construct. private ReceivedDeletedBlockInfoProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReceivedDeletedBlockInfoProto() { status_ = 1; deleteHint_ = ""; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReceivedDeletedBlockInfoProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = block_.toBuilder(); } block_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(block_); block_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; deleteHint_ = bs; break; } case 24: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus value = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { bitField0_ |= 0x00000002; status_ = rawValue; } break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder.class); } /** * Protobuf enum {@code hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus} */ public enum BlockStatus implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum { /** *
       * block being created
       * 
* * RECEIVING = 1; */ RECEIVING(1), /** *
       * block creation complete
       * 
* * RECEIVED = 2; */ RECEIVED(2), /** * DELETED = 3; */ DELETED(3), ; /** *
       * block being created
       * 
* * RECEIVING = 1; */ public static final int RECEIVING_VALUE = 1; /** *
       * block creation complete
       * 
* * RECEIVED = 2; */ public static final int RECEIVED_VALUE = 2; /** * DELETED = 3; */ public static final int DELETED_VALUE = 3; public final int getNumber() { return value; } /** * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static BlockStatus valueOf(int value) { return forNumber(value); } public static BlockStatus forNumber(int value) { switch (value) { case 1: return RECEIVING; case 2: return RECEIVED; case 3: return DELETED; default: return null; } } public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap< BlockStatus> internalValueMap = new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() { public BlockStatus findValueByNumber(int number) { return BlockStatus.forNumber(number); } }; public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.getDescriptor().getEnumTypes().get(0); } private static final BlockStatus[] VALUES = values(); public static BlockStatus valueOf( org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private BlockStatus(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus) } private int bitField0_; public static final int BLOCK_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto block_; /** * required .hadoop.hdfs.BlockProto block = 1; */ public boolean hasBlock() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.BlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlock() { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance() : block_; } /** * required .hadoop.hdfs.BlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlockOrBuilder() { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance() : block_; } public static final int STATUS_FIELD_NUMBER = 3; private int status_; /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ public boolean hasStatus() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus getStatus() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.valueOf(status_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.RECEIVING : result; } public static final int DELETEHINT_FIELD_NUMBER = 2; private volatile java.lang.Object deleteHint_; /** * optional string deleteHint = 2; */ public boolean hasDeleteHint() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string deleteHint = 2; */ public java.lang.String getDeleteHint() { java.lang.Object ref = deleteHint_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { deleteHint_ = s; } return s; } } /** * optional string deleteHint = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getDeleteHintBytes() { java.lang.Object ref = deleteHint_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); deleteHint_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasBlock()) { memoizedIsInitialized = 0; return false; } if (!hasStatus()) { memoizedIsInitialized = 0; return false; } if (!getBlock().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getBlock()); } if (((bitField0_ & 0x00000004) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, deleteHint_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeEnum(3, status_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getBlock()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, deleteHint_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeEnumSize(3, status_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto) obj; if (hasBlock() != other.hasBlock()) return false; if (hasBlock()) { if (!getBlock() .equals(other.getBlock())) return false; } if (hasStatus() != other.hasStatus()) return false; if (hasStatus()) { if (status_ != other.status_) return false; } if (hasDeleteHint() != other.hasDeleteHint()) return false; if (hasDeleteHint()) { if (!getDeleteHint() .equals(other.getDeleteHint())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasBlock()) { hash = (37 * hash) + BLOCK_FIELD_NUMBER; hash = (53 * hash) + getBlock().hashCode(); } if (hasStatus()) { hash = (37 * hash) + STATUS_FIELD_NUMBER; hash = (53 * hash) + status_; } if (hasDeleteHint()) { hash = (37 * hash) + DELETEHINT_FIELD_NUMBER; hash = (53 * hash) + getDeleteHint().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Data structure to send received or deleted block information
     * from datanode to namenode.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlockFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = null; } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); status_ = 1; bitField0_ = (bitField0_ & ~0x00000002); deleteHint_ = ""; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (blockBuilder_ == null) { result.block_ = block_; } else { result.block_ = blockBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.status_ = status_; if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.deleteHint_ = deleteHint_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.getDefaultInstance()) return this; if (other.hasBlock()) { mergeBlock(other.getBlock()); } if (other.hasStatus()) { setStatus(other.getStatus()); } if (other.hasDeleteHint()) { bitField0_ |= 0x00000004; deleteHint_ = other.deleteHint_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasBlock()) { return false; } if (!hasStatus()) { return false; } if (!getBlock().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto block_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> blockBuilder_; /** * required .hadoop.hdfs.BlockProto block = 1; */ public boolean hasBlock() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.BlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto getBlock() { if (blockBuilder_ == null) { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance() : block_; } else { return blockBuilder_.getMessage(); } } /** * required .hadoop.hdfs.BlockProto block = 1; */ public Builder setBlock(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { if (blockBuilder_ == null) { if (value == null) { throw new NullPointerException(); } block_ = value; onChanged(); } else { blockBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.BlockProto block = 1; */ public Builder setBlock( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder builderForValue) { if (blockBuilder_ == null) { block_ = builderForValue.build(); onChanged(); } else { blockBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.BlockProto block = 1; */ public Builder mergeBlock(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto value) { if (blockBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && block_ != null && block_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance()) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.newBuilder(block_).mergeFrom(value).buildPartial(); } else { block_ = value; } onChanged(); } else { blockBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.BlockProto block = 1; */ public Builder clearBlock() { if (blockBuilder_ == null) { block_ = null; onChanged(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hadoop.hdfs.BlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder getBlockBuilder() { bitField0_ |= 0x00000001; onChanged(); return getBlockFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.BlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder getBlockOrBuilder() { if (blockBuilder_ != null) { return blockBuilder_.getMessageOrBuilder(); } else { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.getDefaultInstance() : block_; } } /** * required .hadoop.hdfs.BlockProto block = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder> getBlockFieldBuilder() { if (blockBuilder_ == null) { blockBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.BlockProtoOrBuilder>( getBlock(), getParentForChildren(), isClean()); block_ = null; } return blockBuilder_; } private int status_ = 1; /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ public boolean hasStatus() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus getStatus() { @SuppressWarnings("deprecation") org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus result = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.valueOf(status_); return result == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus.RECEIVING : result; } /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ public Builder setStatus(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.BlockStatus value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; status_ = value.getNumber(); onChanged(); return this; } /** * required .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto.BlockStatus status = 3; */ public Builder clearStatus() { bitField0_ = (bitField0_ & ~0x00000002); status_ = 1; onChanged(); return this; } private java.lang.Object deleteHint_ = ""; /** * optional string deleteHint = 2; */ public boolean hasDeleteHint() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string deleteHint = 2; */ public java.lang.String getDeleteHint() { java.lang.Object ref = deleteHint_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { deleteHint_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string deleteHint = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getDeleteHintBytes() { java.lang.Object ref = deleteHint_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); deleteHint_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * optional string deleteHint = 2; */ public Builder setDeleteHint( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; deleteHint_ = value; onChanged(); return this; } /** * optional string deleteHint = 2; */ public Builder clearDeleteHint() { bitField0_ = (bitField0_ & ~0x00000004); deleteHint_ = getDefaultInstance().getDeleteHint(); onChanged(); return this; } /** * optional string deleteHint = 2; */ public Builder setDeleteHintBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; deleteHint_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public ReceivedDeletedBlockInfoProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new ReceivedDeletedBlockInfoProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface StorageReceivedDeletedBlocksProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated boolean hasStorageUuid(); /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated java.lang.String getStorageUuid(); /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated org.apache.hadoop.thirdparty.protobuf.ByteString getStorageUuidBytes(); /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ java.util.List getBlocksList(); /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto getBlocks(int index); /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ int getBlocksCount(); /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ java.util.List getBlocksOrBuilderList(); /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder getBlocksOrBuilder( int index); /** *
     * supersedes storageUuid.
     * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ boolean hasStorage(); /** *
     * supersedes storageUuid.
     * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto getStorage(); /** *
     * supersedes storageUuid.
     * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder getStorageOrBuilder(); } /** *
   **
   * List of blocks received and deleted for a storage.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto} */ public static final class StorageReceivedDeletedBlocksProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto) StorageReceivedDeletedBlocksProtoOrBuilder { private static final long serialVersionUID = 0L; // Use StorageReceivedDeletedBlocksProto.newBuilder() to construct. private StorageReceivedDeletedBlocksProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private StorageReceivedDeletedBlocksProto() { storageUuid_ = ""; blocks_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private StorageReceivedDeletedBlocksProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; storageUuid_ = bs; break; } case 18: { if (!((mutable_bitField0_ & 0x00000002) != 0)) { blocks_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000002; } blocks_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.PARSER, extensionRegistry)); break; } case 26: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = storage_.toBuilder(); } storage_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(storage_); storage_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder.class); } private int bitField0_; public static final int STORAGEUUID_FIELD_NUMBER = 1; private volatile java.lang.Object storageUuid_; /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public boolean hasStorageUuid() { return ((bitField0_ & 0x00000001) != 0); } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public java.lang.String getStorageUuid() { java.lang.Object ref = storageUuid_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { storageUuid_ = s; } return s; } } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public org.apache.hadoop.thirdparty.protobuf.ByteString getStorageUuidBytes() { java.lang.Object ref = storageUuid_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); storageUuid_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int BLOCKS_FIELD_NUMBER = 2; private java.util.List blocks_; /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public java.util.List getBlocksOrBuilderList() { return blocks_; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto getBlocks(int index) { return blocks_.get(index); } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder getBlocksOrBuilder( int index) { return blocks_.get(index); } public static final int STORAGE_FIELD_NUMBER = 3; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto storage_; /** *
     * supersedes storageUuid.
     * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public boolean hasStorage() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * supersedes storageUuid.
     * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto getStorage() { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } /** *
     * supersedes storageUuid.
     * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder getStorageOrBuilder() { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasStorageUuid()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasStorage()) { if (!getStorage().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, storageUuid_); } for (int i = 0; i < blocks_.size(); i++) { output.writeMessage(2, blocks_.get(i)); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(3, getStorage()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, storageUuid_); } for (int i = 0; i < blocks_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(2, blocks_.get(i)); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, getStorage()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto) obj; if (hasStorageUuid() != other.hasStorageUuid()) return false; if (hasStorageUuid()) { if (!getStorageUuid() .equals(other.getStorageUuid())) return false; } if (!getBlocksList() .equals(other.getBlocksList())) return false; if (hasStorage() != other.hasStorage()) return false; if (hasStorage()) { if (!getStorage() .equals(other.getStorage())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasStorageUuid()) { hash = (37 * hash) + STORAGEUUID_FIELD_NUMBER; hash = (53 * hash) + getStorageUuid().hashCode(); } if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } if (hasStorage()) { hash = (37 * hash) + STORAGE_FIELD_NUMBER; hash = (53 * hash) + getStorage().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * List of blocks received and deleted for a storage.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlocksFieldBuilder(); getStorageFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); storageUuid_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); } else { blocksBuilder_.clear(); } if (storageBuilder_ == null) { storage_ = null; } else { storageBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.storageUuid_ = storageUuid_; if (blocksBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); bitField0_ = (bitField0_ & ~0x00000002); } result.blocks_ = blocks_; } else { result.blocks_ = blocksBuilder_.build(); } if (((from_bitField0_ & 0x00000004) != 0)) { if (storageBuilder_ == null) { result.storage_ = storage_; } else { result.storage_ = storageBuilder_.build(); } to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.getDefaultInstance()) return this; if (other.hasStorageUuid()) { bitField0_ |= 0x00000001; storageUuid_ = other.storageUuid_; onChanged(); } if (blocksBuilder_ == null) { if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } } else { if (!other.blocks_.isEmpty()) { if (blocksBuilder_.isEmpty()) { blocksBuilder_.dispose(); blocksBuilder_ = null; blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000002); blocksBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBlocksFieldBuilder() : null; } else { blocksBuilder_.addAllMessages(other.blocks_); } } } if (other.hasStorage()) { mergeStorage(other.getStorage()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasStorageUuid()) { return false; } for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { return false; } } if (hasStorage()) { if (!getStorage().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object storageUuid_ = ""; /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public boolean hasStorageUuid() { return ((bitField0_ & 0x00000001) != 0); } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public java.lang.String getStorageUuid() { java.lang.Object ref = storageUuid_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { storageUuid_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public org.apache.hadoop.thirdparty.protobuf.ByteString getStorageUuidBytes() { java.lang.Object ref = storageUuid_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); storageUuid_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public Builder setStorageUuid( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; storageUuid_ = value; onChanged(); return this; } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public Builder clearStorageUuid() { bitField0_ = (bitField0_ & ~0x00000001); storageUuid_ = getDefaultInstance().getStorageUuid(); onChanged(); return this; } /** * required string storageUuid = 1 [deprecated = true]; */ @java.lang.Deprecated public Builder setStorageUuidBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; storageUuid_ = value; onChanged(); return this; } private java.util.List blocks_ = java.util.Collections.emptyList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000002) != 0)) { blocks_ = new java.util.ArrayList(blocks_); bitField0_ |= 0x00000002; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder> blocksBuilder_; /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public java.util.List getBlocksList() { if (blocksBuilder_ == null) { return java.util.Collections.unmodifiableList(blocks_); } else { return blocksBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public int getBlocksCount() { if (blocksBuilder_ == null) { return blocks_.size(); } else { return blocksBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto getBlocks(int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.set(index, value); onChanged(); } else { blocksBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.set(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(value); onChanged(); } else { blocksBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(index, value); onChanged(); } else { blocksBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder addBlocks( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder addAllBlocks( java.lang.Iterable values) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); } else { blocksBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder clearBlocks() { if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); } else { blocksBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public Builder removeBlocks(int index) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.remove(index); onChanged(); } else { blocksBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder getBlocksBuilder( int index) { return getBlocksFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder getBlocksOrBuilder( int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public java.util.List getBlocksOrBuilderList() { if (blocksBuilder_ != null) { return blocksBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(blocks_); } } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.ReceivedDeletedBlockInfoProto blocks = 2; */ public java.util.List getBlocksBuilderList() { return getBlocksFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder> getBlocksFieldBuilder() { if (blocksBuilder_ == null) { blocksBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReceivedDeletedBlockInfoProtoOrBuilder>( blocks_, ((bitField0_ & 0x00000002) != 0), getParentForChildren(), isClean()); blocks_ = null; } return blocksBuilder_; } private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto storage_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder> storageBuilder_; /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public boolean hasStorage() { return ((bitField0_ & 0x00000004) != 0); } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto getStorage() { if (storageBuilder_ == null) { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } else { return storageBuilder_.getMessage(); } } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public Builder setStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto value) { if (storageBuilder_ == null) { if (value == null) { throw new NullPointerException(); } storage_ = value; onChanged(); } else { storageBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public Builder setStorage( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder builderForValue) { if (storageBuilder_ == null) { storage_ = builderForValue.build(); onChanged(); } else { storageBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public Builder mergeStorage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto value) { if (storageBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0) && storage_ != null && storage_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance()) { storage_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.newBuilder(storage_).mergeFrom(value).buildPartial(); } else { storage_ = value; } onChanged(); } else { storageBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public Builder clearStorage() { if (storageBuilder_ == null) { storage_ = null; onChanged(); } else { storageBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder getStorageBuilder() { bitField0_ |= 0x00000004; onChanged(); return getStorageFieldBuilder().getBuilder(); } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder getStorageOrBuilder() { if (storageBuilder_ != null) { return storageBuilder_.getMessageOrBuilder(); } else { return storage_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.getDefaultInstance() : storage_; } } /** *
       * supersedes storageUuid.
       * 
* * optional .hadoop.hdfs.DatanodeStorageProto storage = 3; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder> getStorageFieldBuilder() { if (storageBuilder_ == null) { storageBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeStorageProtoOrBuilder>( getStorage(), getParentForChildren(), isClean()); storage_ = null; } return storageBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public StorageReceivedDeletedBlocksProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new StorageReceivedDeletedBlocksProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockReceivedAndDeletedRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ boolean hasRegistration(); /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration(); /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder(); /** * required string blockPoolId = 2; */ boolean hasBlockPoolId(); /** * required string blockPoolId = 2; */ java.lang.String getBlockPoolId(); /** * required string blockPoolId = 2; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes(); /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ java.util.List getBlocksList(); /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto getBlocks(int index); /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ int getBlocksCount(); /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ java.util.List getBlocksOrBuilderList(); /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder getBlocksOrBuilder( int index); } /** *
   **
   * registration - datanode registration information
   * blockPoolID  - block pool ID of the reported blocks
   * blocks       - Received/deleted block list
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto} */ public static final class BlockReceivedAndDeletedRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) BlockReceivedAndDeletedRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockReceivedAndDeletedRequestProto.newBuilder() to construct. private BlockReceivedAndDeletedRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockReceivedAndDeletedRequestProto() { blockPoolId_ = ""; blocks_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockReceivedAndDeletedRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registration_.toBuilder(); } registration_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registration_); registration_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; blockPoolId_ = bs; break; } case 26: { if (!((mutable_bitField0_ & 0x00000004) != 0)) { blocks_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000004; } blocks_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.Builder.class); } private int bitField0_; public static final int REGISTRATION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } public static final int BLOCKPOOLID_FIELD_NUMBER = 2; private volatile java.lang.Object blockPoolId_; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int BLOCKS_FIELD_NUMBER = 3; private java.util.List blocks_; /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public java.util.List getBlocksOrBuilderList() { return blocks_; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto getBlocks(int index) { return blocks_.get(index); } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder getBlocksOrBuilder( int index) { return blocks_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistration()) { memoizedIsInitialized = 0; return false; } if (!hasBlockPoolId()) { memoizedIsInitialized = 0; return false; } if (!getRegistration().isInitialized()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistration()); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, blockPoolId_); } for (int i = 0; i < blocks_.size(); i++) { output.writeMessage(3, blocks_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistration()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, blockPoolId_); } for (int i = 0; i < blocks_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(3, blocks_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto) obj; if (hasRegistration() != other.hasRegistration()) return false; if (hasRegistration()) { if (!getRegistration() .equals(other.getRegistration())) return false; } if (hasBlockPoolId() != other.hasBlockPoolId()) return false; if (hasBlockPoolId()) { if (!getBlockPoolId() .equals(other.getBlockPoolId())) return false; } if (!getBlocksList() .equals(other.getBlocksList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistration()) { hash = (37 * hash) + REGISTRATION_FIELD_NUMBER; hash = (53 * hash) + getRegistration().hashCode(); } if (hasBlockPoolId()) { hash = (37 * hash) + BLOCKPOOLID_FIELD_NUMBER; hash = (53 * hash) + getBlockPoolId().hashCode(); } if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registration - datanode registration information
     * blockPoolID  - block pool ID of the reported blocks
     * blocks       - Received/deleted block list
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistrationFieldBuilder(); getBlocksFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registrationBuilder_ == null) { registration_ = null; } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); blockPoolId_ = ""; bitField0_ = (bitField0_ & ~0x00000002); if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); } else { blocksBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registrationBuilder_ == null) { result.registration_ = registration_; } else { result.registration_ = registrationBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.blockPoolId_ = blockPoolId_; if (blocksBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); bitField0_ = (bitField0_ & ~0x00000004); } result.blocks_ = blocks_; } else { result.blocks_ = blocksBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.getDefaultInstance()) return this; if (other.hasRegistration()) { mergeRegistration(other.getRegistration()); } if (other.hasBlockPoolId()) { bitField0_ |= 0x00000002; blockPoolId_ = other.blockPoolId_; onChanged(); } if (blocksBuilder_ == null) { if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } } else { if (!other.blocks_.isEmpty()) { if (blocksBuilder_.isEmpty()) { blocksBuilder_.dispose(); blocksBuilder_ = null; blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000004); blocksBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBlocksFieldBuilder() : null; } else { blocksBuilder_.addAllMessages(other.blocks_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistration()) { return false; } if (!hasBlockPoolId()) { return false; } if (!getRegistration().isInitialized()) { return false; } for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registration_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registrationBuilder_; /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public boolean hasRegistration() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistration() { if (registrationBuilder_ == null) { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } else { return registrationBuilder_.getMessage(); } } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registration_ = value; onChanged(); } else { registrationBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder setRegistration( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registrationBuilder_ == null) { registration_ = builderForValue.build(); onChanged(); } else { registrationBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder mergeRegistration(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registrationBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registration_ != null && registration_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registration_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registration_).mergeFrom(value).buildPartial(); } else { registration_ = value; } onChanged(); } else { registrationBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public Builder clearRegistration() { if (registrationBuilder_ == null) { registration_ = null; onChanged(); } else { registrationBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistrationBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistrationFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistrationOrBuilder() { if (registrationBuilder_ != null) { return registrationBuilder_.getMessageOrBuilder(); } else { return registration_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registration_; } } /** * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registration = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistrationFieldBuilder() { if (registrationBuilder_ == null) { registrationBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistration(), getParentForChildren(), isClean()); registration_ = null; } return registrationBuilder_; } private java.lang.Object blockPoolId_ = ""; /** * required string blockPoolId = 2; */ public boolean hasBlockPoolId() { return ((bitField0_ & 0x00000002) != 0); } /** * required string blockPoolId = 2; */ public java.lang.String getBlockPoolId() { java.lang.Object ref = blockPoolId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { blockPoolId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string blockPoolId = 2; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBlockPoolIdBytes() { java.lang.Object ref = blockPoolId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); blockPoolId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * required string blockPoolId = 2; */ public Builder setBlockPoolId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder clearBlockPoolId() { bitField0_ = (bitField0_ & ~0x00000002); blockPoolId_ = getDefaultInstance().getBlockPoolId(); onChanged(); return this; } /** * required string blockPoolId = 2; */ public Builder setBlockPoolIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; blockPoolId_ = value; onChanged(); return this; } private java.util.List blocks_ = java.util.Collections.emptyList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { blocks_ = new java.util.ArrayList(blocks_); bitField0_ |= 0x00000004; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder> blocksBuilder_; /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public java.util.List getBlocksList() { if (blocksBuilder_ == null) { return java.util.Collections.unmodifiableList(blocks_); } else { return blocksBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public int getBlocksCount() { if (blocksBuilder_ == null) { return blocks_.size(); } else { return blocksBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto getBlocks(int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.set(index, value); onChanged(); } else { blocksBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.set(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(value); onChanged(); } else { blocksBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(index, value); onChanged(); } else { blocksBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder addBlocks( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder addAllBlocks( java.lang.Iterable values) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); } else { blocksBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder clearBlocks() { if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); } else { blocksBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public Builder removeBlocks(int index) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.remove(index); onChanged(); } else { blocksBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder getBlocksBuilder( int index) { return getBlocksFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder getBlocksOrBuilder( int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public java.util.List getBlocksOrBuilderList() { if (blocksBuilder_ != null) { return blocksBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(blocks_); } } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.datanode.StorageReceivedDeletedBlocksProto blocks = 3; */ public java.util.List getBlocksBuilderList() { return getBlocksFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder> getBlocksFieldBuilder() { if (blocksBuilder_ == null) { blocksBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.StorageReceivedDeletedBlocksProtoOrBuilder>( blocks_, ((bitField0_ & 0x00000004) != 0), getParentForChildren(), isClean()); blocks_ = null; } return blocksBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockReceivedAndDeletedRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockReceivedAndDeletedRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BlockReceivedAndDeletedResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { } /** *
   **
   * void response
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto} */ public static final class BlockReceivedAndDeletedResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto) BlockReceivedAndDeletedResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use BlockReceivedAndDeletedResponseProto.newBuilder() to construct. private BlockReceivedAndDeletedResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BlockReceivedAndDeletedResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BlockReceivedAndDeletedResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * void response
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public BlockReceivedAndDeletedResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new BlockReceivedAndDeletedResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ErrorReportRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.ErrorReportRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** *
     * Registartion info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ boolean hasRegistartion(); /** *
     * Registartion info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistartion(); /** *
     * Registartion info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistartionOrBuilder(); /** *
     * Error code
     * 
* * required uint32 errorCode = 2; */ boolean hasErrorCode(); /** *
     * Error code
     * 
* * required uint32 errorCode = 2; */ int getErrorCode(); /** *
     * Error message
     * 
* * required string msg = 3; */ boolean hasMsg(); /** *
     * Error message
     * 
* * required string msg = 3; */ java.lang.String getMsg(); /** *
     * Error message
     * 
* * required string msg = 3; */ org.apache.hadoop.thirdparty.protobuf.ByteString getMsgBytes(); } /** *
   **
   * registartion - Datanode reporting the error
   * errorCode - error code indicating the error
   * msg - Free text description of the error
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.ErrorReportRequestProto} */ public static final class ErrorReportRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.ErrorReportRequestProto) ErrorReportRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use ErrorReportRequestProto.newBuilder() to construct. private ErrorReportRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ErrorReportRequestProto() { msg_ = ""; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ErrorReportRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = registartion_.toBuilder(); } registartion_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(registartion_); registartion_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; errorCode_ = input.readUInt32(); break; } case 26: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; msg_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.Builder.class); } /** * Protobuf enum {@code hadoop.hdfs.datanode.ErrorReportRequestProto.ErrorCode} */ public enum ErrorCode implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum { /** *
       * Error report to be logged at the namenode
       * 
* * NOTIFY = 0; */ NOTIFY(0), /** *
       * DN has disk errors but still has valid volumes
       * 
* * DISK_ERROR = 1; */ DISK_ERROR(1), /** *
       * Command from namenode has invalid block ID
       * 
* * INVALID_BLOCK = 2; */ INVALID_BLOCK(2), /** *
       * No valid volumes left on datanode
       * 
* * FATAL_DISK_ERROR = 3; */ FATAL_DISK_ERROR(3), ; /** *
       * Error report to be logged at the namenode
       * 
* * NOTIFY = 0; */ public static final int NOTIFY_VALUE = 0; /** *
       * DN has disk errors but still has valid volumes
       * 
* * DISK_ERROR = 1; */ public static final int DISK_ERROR_VALUE = 1; /** *
       * Command from namenode has invalid block ID
       * 
* * INVALID_BLOCK = 2; */ public static final int INVALID_BLOCK_VALUE = 2; /** *
       * No valid volumes left on datanode
       * 
* * FATAL_DISK_ERROR = 3; */ public static final int FATAL_DISK_ERROR_VALUE = 3; public final int getNumber() { return value; } /** * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static ErrorCode valueOf(int value) { return forNumber(value); } public static ErrorCode forNumber(int value) { switch (value) { case 0: return NOTIFY; case 1: return DISK_ERROR; case 2: return INVALID_BLOCK; case 3: return FATAL_DISK_ERROR; default: return null; } } public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap< ErrorCode> internalValueMap = new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() { public ErrorCode findValueByNumber(int number) { return ErrorCode.forNumber(number); } }; public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.getDescriptor().getEnumTypes().get(0); } private static final ErrorCode[] VALUES = values(); public static ErrorCode valueOf( org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private ErrorCode(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hadoop.hdfs.datanode.ErrorReportRequestProto.ErrorCode) } private int bitField0_; public static final int REGISTARTION_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registartion_; /** *
     * Registartion info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public boolean hasRegistartion() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * Registartion info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistartion() { return registartion_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registartion_; } /** *
     * Registartion info
     * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistartionOrBuilder() { return registartion_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registartion_; } public static final int ERRORCODE_FIELD_NUMBER = 2; private int errorCode_; /** *
     * Error code
     * 
* * required uint32 errorCode = 2; */ public boolean hasErrorCode() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * Error code
     * 
* * required uint32 errorCode = 2; */ public int getErrorCode() { return errorCode_; } public static final int MSG_FIELD_NUMBER = 3; private volatile java.lang.Object msg_; /** *
     * Error message
     * 
* * required string msg = 3; */ public boolean hasMsg() { return ((bitField0_ & 0x00000004) != 0); } /** *
     * Error message
     * 
* * required string msg = 3; */ public java.lang.String getMsg() { java.lang.Object ref = msg_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { msg_ = s; } return s; } } /** *
     * Error message
     * 
* * required string msg = 3; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getMsgBytes() { java.lang.Object ref = msg_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); msg_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegistartion()) { memoizedIsInitialized = 0; return false; } if (!hasErrorCode()) { memoizedIsInitialized = 0; return false; } if (!hasMsg()) { memoizedIsInitialized = 0; return false; } if (!getRegistartion().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegistartion()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt32(2, errorCode_); } if (((bitField0_ & 0x00000004) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 3, msg_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getRegistartion()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt32Size(2, errorCode_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(3, msg_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto) obj; if (hasRegistartion() != other.hasRegistartion()) return false; if (hasRegistartion()) { if (!getRegistartion() .equals(other.getRegistartion())) return false; } if (hasErrorCode() != other.hasErrorCode()) return false; if (hasErrorCode()) { if (getErrorCode() != other.getErrorCode()) return false; } if (hasMsg() != other.hasMsg()) return false; if (hasMsg()) { if (!getMsg() .equals(other.getMsg())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegistartion()) { hash = (37 * hash) + REGISTARTION_FIELD_NUMBER; hash = (53 * hash) + getRegistartion().hashCode(); } if (hasErrorCode()) { hash = (37 * hash) + ERRORCODE_FIELD_NUMBER; hash = (53 * hash) + getErrorCode(); } if (hasMsg()) { hash = (37 * hash) + MSG_FIELD_NUMBER; hash = (53 * hash) + getMsg().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * registartion - Datanode reporting the error
     * errorCode - error code indicating the error
     * msg - Free text description of the error
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.ErrorReportRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.ErrorReportRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegistartionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (registartionBuilder_ == null) { registartion_ = null; } else { registartionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); errorCode_ = 0; bitField0_ = (bitField0_ & ~0x00000002); msg_ = ""; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (registartionBuilder_ == null) { result.registartion_ = registartion_; } else { result.registartion_ = registartionBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.errorCode_ = errorCode_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.msg_ = msg_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.getDefaultInstance()) return this; if (other.hasRegistartion()) { mergeRegistartion(other.getRegistartion()); } if (other.hasErrorCode()) { setErrorCode(other.getErrorCode()); } if (other.hasMsg()) { bitField0_ |= 0x00000004; msg_ = other.msg_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegistartion()) { return false; } if (!hasErrorCode()) { return false; } if (!hasMsg()) { return false; } if (!getRegistartion().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto registartion_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> registartionBuilder_; /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public boolean hasRegistartion() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto getRegistartion() { if (registartionBuilder_ == null) { return registartion_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registartion_; } else { return registartionBuilder_.getMessage(); } } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public Builder setRegistartion(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registartionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } registartion_ = value; onChanged(); } else { registartionBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public Builder setRegistartion( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder builderForValue) { if (registartionBuilder_ == null) { registartion_ = builderForValue.build(); onChanged(); } else { registartionBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public Builder mergeRegistartion(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto value) { if (registartionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && registartion_ != null && registartion_ != org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance()) { registartion_ = org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.newBuilder(registartion_).mergeFrom(value).buildPartial(); } else { registartion_ = value; } onChanged(); } else { registartionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public Builder clearRegistartion() { if (registartionBuilder_ == null) { registartion_ = null; onChanged(); } else { registartionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder getRegistartionBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegistartionFieldBuilder().getBuilder(); } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder getRegistartionOrBuilder() { if (registartionBuilder_ != null) { return registartionBuilder_.getMessageOrBuilder(); } else { return registartion_ == null ? org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.getDefaultInstance() : registartion_; } } /** *
       * Registartion info
       * 
* * required .hadoop.hdfs.datanode.DatanodeRegistrationProto registartion = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder> getRegistartionFieldBuilder() { if (registartionBuilder_ == null) { registartionBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProto.Builder, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeRegistrationProtoOrBuilder>( getRegistartion(), getParentForChildren(), isClean()); registartion_ = null; } return registartionBuilder_; } private int errorCode_ ; /** *
       * Error code
       * 
* * required uint32 errorCode = 2; */ public boolean hasErrorCode() { return ((bitField0_ & 0x00000002) != 0); } /** *
       * Error code
       * 
* * required uint32 errorCode = 2; */ public int getErrorCode() { return errorCode_; } /** *
       * Error code
       * 
* * required uint32 errorCode = 2; */ public Builder setErrorCode(int value) { bitField0_ |= 0x00000002; errorCode_ = value; onChanged(); return this; } /** *
       * Error code
       * 
* * required uint32 errorCode = 2; */ public Builder clearErrorCode() { bitField0_ = (bitField0_ & ~0x00000002); errorCode_ = 0; onChanged(); return this; } private java.lang.Object msg_ = ""; /** *
       * Error message
       * 
* * required string msg = 3; */ public boolean hasMsg() { return ((bitField0_ & 0x00000004) != 0); } /** *
       * Error message
       * 
* * required string msg = 3; */ public java.lang.String getMsg() { java.lang.Object ref = msg_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { msg_ = s; } return s; } else { return (java.lang.String) ref; } } /** *
       * Error message
       * 
* * required string msg = 3; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getMsgBytes() { java.lang.Object ref = msg_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); msg_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** *
       * Error message
       * 
* * required string msg = 3; */ public Builder setMsg( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; msg_ = value; onChanged(); return this; } /** *
       * Error message
       * 
* * required string msg = 3; */ public Builder clearMsg() { bitField0_ = (bitField0_ & ~0x00000004); msg_ = getDefaultInstance().getMsg(); onChanged(); return this; } /** *
       * Error message
       * 
* * required string msg = 3; */ public Builder setMsgBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; msg_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.ErrorReportRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.ErrorReportRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public ErrorReportRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new ErrorReportRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ErrorReportResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.ErrorReportResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { } /** *
   **
   * void response
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.ErrorReportResponseProto} */ public static final class ErrorReportResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.ErrorReportResponseProto) ErrorReportResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use ErrorReportResponseProto.newBuilder() to construct. private ErrorReportResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ErrorReportResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ErrorReportResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * void response
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.ErrorReportResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.ErrorReportResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.ErrorReportResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.ErrorReportResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public ErrorReportResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new ErrorReportResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ReportBadBlocksRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.ReportBadBlocksRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ java.util.List getBlocksList(); /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto getBlocks(int index); /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ int getBlocksCount(); /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ java.util.List getBlocksOrBuilderList(); /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProtoOrBuilder getBlocksOrBuilder( int index); } /** *
   **
   * blocks - list of blocks that are reported as corrupt
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.ReportBadBlocksRequestProto} */ public static final class ReportBadBlocksRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.ReportBadBlocksRequestProto) ReportBadBlocksRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use ReportBadBlocksRequestProto.newBuilder() to construct. private ReportBadBlocksRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReportBadBlocksRequestProto() { blocks_ = java.util.Collections.emptyList(); } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReportBadBlocksRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { blocks_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } blocks_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.Builder.class); } public static final int BLOCKS_FIELD_NUMBER = 1; private java.util.List blocks_; /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public java.util.List getBlocksList() { return blocks_; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public java.util.List getBlocksOrBuilderList() { return blocks_; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public int getBlocksCount() { return blocks_.size(); } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto getBlocks(int index) { return blocks_.get(index); } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProtoOrBuilder getBlocksOrBuilder( int index) { return blocks_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < blocks_.size(); i++) { output.writeMessage(1, blocks_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < blocks_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, blocks_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto) obj; if (!getBlocksList() .equals(other.getBlocksList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getBlocksCount() > 0) { hash = (37 * hash) + BLOCKS_FIELD_NUMBER; hash = (53 * hash) + getBlocksList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * blocks - list of blocks that are reported as corrupt
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.ReportBadBlocksRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.ReportBadBlocksRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlocksFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { blocksBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto(this); int from_bitField0_ = bitField0_; if (blocksBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { blocks_ = java.util.Collections.unmodifiableList(blocks_); bitField0_ = (bitField0_ & ~0x00000001); } result.blocks_ = blocks_; } else { result.blocks_ = blocksBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.getDefaultInstance()) return this; if (blocksBuilder_ == null) { if (!other.blocks_.isEmpty()) { if (blocks_.isEmpty()) { blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureBlocksIsMutable(); blocks_.addAll(other.blocks_); } onChanged(); } } else { if (!other.blocks_.isEmpty()) { if (blocksBuilder_.isEmpty()) { blocksBuilder_.dispose(); blocksBuilder_ = null; blocks_ = other.blocks_; bitField0_ = (bitField0_ & ~0x00000001); blocksBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBlocksFieldBuilder() : null; } else { blocksBuilder_.addAllMessages(other.blocks_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getBlocksCount(); i++) { if (!getBlocks(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List blocks_ = java.util.Collections.emptyList(); private void ensureBlocksIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { blocks_ = new java.util.ArrayList(blocks_); bitField0_ |= 0x00000001; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProtoOrBuilder> blocksBuilder_; /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public java.util.List getBlocksList() { if (blocksBuilder_ == null) { return java.util.Collections.unmodifiableList(blocks_); } else { return blocksBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public int getBlocksCount() { if (blocksBuilder_ == null) { return blocks_.size(); } else { return blocksBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto getBlocks(int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.set(index, value); onChanged(); } else { blocksBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder setBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.set(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder addBlocks(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(value); onChanged(); } else { blocksBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto value) { if (blocksBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBlocksIsMutable(); blocks_.add(index, value); onChanged(); } else { blocksBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder addBlocks( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder addBlocks( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder builderForValue) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.add(index, builderForValue.build()); onChanged(); } else { blocksBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder addAllBlocks( java.lang.Iterable values) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, blocks_); onChanged(); } else { blocksBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder clearBlocks() { if (blocksBuilder_ == null) { blocks_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { blocksBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public Builder removeBlocks(int index) { if (blocksBuilder_ == null) { ensureBlocksIsMutable(); blocks_.remove(index); onChanged(); } else { blocksBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder getBlocksBuilder( int index) { return getBlocksFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProtoOrBuilder getBlocksOrBuilder( int index) { if (blocksBuilder_ == null) { return blocks_.get(index); } else { return blocksBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public java.util.List getBlocksOrBuilderList() { if (blocksBuilder_ != null) { return blocksBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(blocks_); } } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder addBlocksBuilder() { return getBlocksFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder addBlocksBuilder( int index) { return getBlocksFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.LocatedBlockProto blocks = 1; */ public java.util.List getBlocksBuilderList() { return getBlocksFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProtoOrBuilder> getBlocksFieldBuilder() { if (blocksBuilder_ == null) { blocksBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.LocatedBlockProtoOrBuilder>( blocks_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); blocks_ = null; } return blocksBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.ReportBadBlocksRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.ReportBadBlocksRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public ReportBadBlocksRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new ReportBadBlocksRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ReportBadBlocksResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.ReportBadBlocksResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { } /** *
   **
   * void response
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.ReportBadBlocksResponseProto} */ public static final class ReportBadBlocksResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.ReportBadBlocksResponseProto) ReportBadBlocksResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use ReportBadBlocksResponseProto.newBuilder() to construct. private ReportBadBlocksResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReportBadBlocksResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReportBadBlocksResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * void response
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.ReportBadBlocksResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.ReportBadBlocksResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.ReportBadBlocksResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.ReportBadBlocksResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public ReportBadBlocksResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new ReportBadBlocksResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CommitBlockSynchronizationRequestProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ boolean hasBlock(); /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto getBlock(); /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProtoOrBuilder getBlockOrBuilder(); /** * required uint64 newGenStamp = 2; */ boolean hasNewGenStamp(); /** * required uint64 newGenStamp = 2; */ long getNewGenStamp(); /** * required uint64 newLength = 3; */ boolean hasNewLength(); /** * required uint64 newLength = 3; */ long getNewLength(); /** * required bool closeFile = 4; */ boolean hasCloseFile(); /** * required bool closeFile = 4; */ boolean getCloseFile(); /** * required bool deleteBlock = 5; */ boolean hasDeleteBlock(); /** * required bool deleteBlock = 5; */ boolean getDeleteBlock(); /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ java.util.List getNewTaragetsList(); /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto getNewTaragets(int index); /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ int getNewTaragetsCount(); /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ java.util.List getNewTaragetsOrBuilderList(); /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder getNewTaragetsOrBuilder( int index); /** * repeated string newTargetStorages = 7; */ java.util.List getNewTargetStoragesList(); /** * repeated string newTargetStorages = 7; */ int getNewTargetStoragesCount(); /** * repeated string newTargetStorages = 7; */ java.lang.String getNewTargetStorages(int index); /** * repeated string newTargetStorages = 7; */ org.apache.hadoop.thirdparty.protobuf.ByteString getNewTargetStoragesBytes(int index); } /** *
   **
   * Commit block synchronization request during lease recovery
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto} */ public static final class CommitBlockSynchronizationRequestProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) CommitBlockSynchronizationRequestProtoOrBuilder { private static final long serialVersionUID = 0L; // Use CommitBlockSynchronizationRequestProto.newBuilder() to construct. private CommitBlockSynchronizationRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CommitBlockSynchronizationRequestProto() { newTaragets_ = java.util.Collections.emptyList(); newTargetStorages_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CommitBlockSynchronizationRequestProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = block_.toBuilder(); } block_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(block_); block_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; newGenStamp_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; newLength_ = input.readUInt64(); break; } case 32: { bitField0_ |= 0x00000008; closeFile_ = input.readBool(); break; } case 40: { bitField0_ |= 0x00000010; deleteBlock_ = input.readBool(); break; } case 50: { if (!((mutable_bitField0_ & 0x00000020) != 0)) { newTaragets_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000020; } newTaragets_.add( input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.PARSER, extensionRegistry)); break; } case 58: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); if (!((mutable_bitField0_ & 0x00000040) != 0)) { newTargetStorages_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000040; } newTargetStorages_.add(bs); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000020) != 0)) { newTaragets_ = java.util.Collections.unmodifiableList(newTaragets_); } if (((mutable_bitField0_ & 0x00000040) != 0)) { newTargetStorages_ = newTargetStorages_.getUnmodifiableView(); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.Builder.class); } private int bitField0_; public static final int BLOCK_FIELD_NUMBER = 1; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto block_; /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public boolean hasBlock() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto getBlock() { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance() : block_; } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProtoOrBuilder getBlockOrBuilder() { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance() : block_; } public static final int NEWGENSTAMP_FIELD_NUMBER = 2; private long newGenStamp_; /** * required uint64 newGenStamp = 2; */ public boolean hasNewGenStamp() { return ((bitField0_ & 0x00000002) != 0); } /** * required uint64 newGenStamp = 2; */ public long getNewGenStamp() { return newGenStamp_; } public static final int NEWLENGTH_FIELD_NUMBER = 3; private long newLength_; /** * required uint64 newLength = 3; */ public boolean hasNewLength() { return ((bitField0_ & 0x00000004) != 0); } /** * required uint64 newLength = 3; */ public long getNewLength() { return newLength_; } public static final int CLOSEFILE_FIELD_NUMBER = 4; private boolean closeFile_; /** * required bool closeFile = 4; */ public boolean hasCloseFile() { return ((bitField0_ & 0x00000008) != 0); } /** * required bool closeFile = 4; */ public boolean getCloseFile() { return closeFile_; } public static final int DELETEBLOCK_FIELD_NUMBER = 5; private boolean deleteBlock_; /** * required bool deleteBlock = 5; */ public boolean hasDeleteBlock() { return ((bitField0_ & 0x00000010) != 0); } /** * required bool deleteBlock = 5; */ public boolean getDeleteBlock() { return deleteBlock_; } public static final int NEWTARAGETS_FIELD_NUMBER = 6; private java.util.List newTaragets_; /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public java.util.List getNewTaragetsList() { return newTaragets_; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public java.util.List getNewTaragetsOrBuilderList() { return newTaragets_; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public int getNewTaragetsCount() { return newTaragets_.size(); } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto getNewTaragets(int index) { return newTaragets_.get(index); } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder getNewTaragetsOrBuilder( int index) { return newTaragets_.get(index); } public static final int NEWTARGETSTORAGES_FIELD_NUMBER = 7; private org.apache.hadoop.thirdparty.protobuf.LazyStringList newTargetStorages_; /** * repeated string newTargetStorages = 7; */ public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList getNewTargetStoragesList() { return newTargetStorages_; } /** * repeated string newTargetStorages = 7; */ public int getNewTargetStoragesCount() { return newTargetStorages_.size(); } /** * repeated string newTargetStorages = 7; */ public java.lang.String getNewTargetStorages(int index) { return newTargetStorages_.get(index); } /** * repeated string newTargetStorages = 7; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getNewTargetStoragesBytes(int index) { return newTargetStorages_.getByteString(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasBlock()) { memoizedIsInitialized = 0; return false; } if (!hasNewGenStamp()) { memoizedIsInitialized = 0; return false; } if (!hasNewLength()) { memoizedIsInitialized = 0; return false; } if (!hasCloseFile()) { memoizedIsInitialized = 0; return false; } if (!hasDeleteBlock()) { memoizedIsInitialized = 0; return false; } if (!getBlock().isInitialized()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getNewTaragetsCount(); i++) { if (!getNewTaragets(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getBlock()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, newGenStamp_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, newLength_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeBool(4, closeFile_); } if (((bitField0_ & 0x00000010) != 0)) { output.writeBool(5, deleteBlock_); } for (int i = 0; i < newTaragets_.size(); i++) { output.writeMessage(6, newTaragets_.get(i)); } for (int i = 0; i < newTargetStorages_.size(); i++) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 7, newTargetStorages_.getRaw(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(1, getBlock()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(2, newGenStamp_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeUInt64Size(3, newLength_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeBoolSize(4, closeFile_); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeBoolSize(5, deleteBlock_); } for (int i = 0; i < newTaragets_.size(); i++) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeMessageSize(6, newTaragets_.get(i)); } { int dataSize = 0; for (int i = 0; i < newTargetStorages_.size(); i++) { dataSize += computeStringSizeNoTag(newTargetStorages_.getRaw(i)); } size += dataSize; size += 1 * getNewTargetStoragesList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto) obj; if (hasBlock() != other.hasBlock()) return false; if (hasBlock()) { if (!getBlock() .equals(other.getBlock())) return false; } if (hasNewGenStamp() != other.hasNewGenStamp()) return false; if (hasNewGenStamp()) { if (getNewGenStamp() != other.getNewGenStamp()) return false; } if (hasNewLength() != other.hasNewLength()) return false; if (hasNewLength()) { if (getNewLength() != other.getNewLength()) return false; } if (hasCloseFile() != other.hasCloseFile()) return false; if (hasCloseFile()) { if (getCloseFile() != other.getCloseFile()) return false; } if (hasDeleteBlock() != other.hasDeleteBlock()) return false; if (hasDeleteBlock()) { if (getDeleteBlock() != other.getDeleteBlock()) return false; } if (!getNewTaragetsList() .equals(other.getNewTaragetsList())) return false; if (!getNewTargetStoragesList() .equals(other.getNewTargetStoragesList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasBlock()) { hash = (37 * hash) + BLOCK_FIELD_NUMBER; hash = (53 * hash) + getBlock().hashCode(); } if (hasNewGenStamp()) { hash = (37 * hash) + NEWGENSTAMP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getNewGenStamp()); } if (hasNewLength()) { hash = (37 * hash) + NEWLENGTH_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( getNewLength()); } if (hasCloseFile()) { hash = (37 * hash) + CLOSEFILE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean( getCloseFile()); } if (hasDeleteBlock()) { hash = (37 * hash) + DELETEBLOCK_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashBoolean( getDeleteBlock()); } if (getNewTaragetsCount() > 0) { hash = (37 * hash) + NEWTARAGETS_FIELD_NUMBER; hash = (53 * hash) + getNewTaragetsList().hashCode(); } if (getNewTargetStoragesCount() > 0) { hash = (37 * hash) + NEWTARGETSTORAGES_FIELD_NUMBER; hash = (53 * hash) + getNewTargetStoragesList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Commit block synchronization request during lease recovery
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBlockFieldBuilder(); getNewTaragetsFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (blockBuilder_ == null) { block_ = null; } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); newGenStamp_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); newLength_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); closeFile_ = false; bitField0_ = (bitField0_ & ~0x00000008); deleteBlock_ = false; bitField0_ = (bitField0_ & ~0x00000010); if (newTaragetsBuilder_ == null) { newTaragets_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); } else { newTaragetsBuilder_.clear(); } newTargetStorages_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000040); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (blockBuilder_ == null) { result.block_ = block_; } else { result.block_ = blockBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.newGenStamp_ = newGenStamp_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.newLength_ = newLength_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.closeFile_ = closeFile_; to_bitField0_ |= 0x00000008; } if (((from_bitField0_ & 0x00000010) != 0)) { result.deleteBlock_ = deleteBlock_; to_bitField0_ |= 0x00000010; } if (newTaragetsBuilder_ == null) { if (((bitField0_ & 0x00000020) != 0)) { newTaragets_ = java.util.Collections.unmodifiableList(newTaragets_); bitField0_ = (bitField0_ & ~0x00000020); } result.newTaragets_ = newTaragets_; } else { result.newTaragets_ = newTaragetsBuilder_.build(); } if (((bitField0_ & 0x00000040) != 0)) { newTargetStorages_ = newTargetStorages_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000040); } result.newTargetStorages_ = newTargetStorages_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.getDefaultInstance()) return this; if (other.hasBlock()) { mergeBlock(other.getBlock()); } if (other.hasNewGenStamp()) { setNewGenStamp(other.getNewGenStamp()); } if (other.hasNewLength()) { setNewLength(other.getNewLength()); } if (other.hasCloseFile()) { setCloseFile(other.getCloseFile()); } if (other.hasDeleteBlock()) { setDeleteBlock(other.getDeleteBlock()); } if (newTaragetsBuilder_ == null) { if (!other.newTaragets_.isEmpty()) { if (newTaragets_.isEmpty()) { newTaragets_ = other.newTaragets_; bitField0_ = (bitField0_ & ~0x00000020); } else { ensureNewTaragetsIsMutable(); newTaragets_.addAll(other.newTaragets_); } onChanged(); } } else { if (!other.newTaragets_.isEmpty()) { if (newTaragetsBuilder_.isEmpty()) { newTaragetsBuilder_.dispose(); newTaragetsBuilder_ = null; newTaragets_ = other.newTaragets_; bitField0_ = (bitField0_ & ~0x00000020); newTaragetsBuilder_ = org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getNewTaragetsFieldBuilder() : null; } else { newTaragetsBuilder_.addAllMessages(other.newTaragets_); } } } if (!other.newTargetStorages_.isEmpty()) { if (newTargetStorages_.isEmpty()) { newTargetStorages_ = other.newTargetStorages_; bitField0_ = (bitField0_ & ~0x00000040); } else { ensureNewTargetStoragesIsMutable(); newTargetStorages_.addAll(other.newTargetStorages_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasBlock()) { return false; } if (!hasNewGenStamp()) { return false; } if (!hasNewLength()) { return false; } if (!hasCloseFile()) { return false; } if (!hasDeleteBlock()) { return false; } if (!getBlock().isInitialized()) { return false; } for (int i = 0; i < getNewTaragetsCount(); i++) { if (!getNewTaragets(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto block_; private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProtoOrBuilder> blockBuilder_; /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public boolean hasBlock() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto getBlock() { if (blockBuilder_ == null) { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance() : block_; } else { return blockBuilder_.getMessage(); } } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public Builder setBlock(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto value) { if (blockBuilder_ == null) { if (value == null) { throw new NullPointerException(); } block_ = value; onChanged(); } else { blockBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public Builder setBlock( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.Builder builderForValue) { if (blockBuilder_ == null) { block_ = builderForValue.build(); onChanged(); } else { blockBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public Builder mergeBlock(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto value) { if (blockBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && block_ != null && block_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance()) { block_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.newBuilder(block_).mergeFrom(value).buildPartial(); } else { block_ = value; } onChanged(); } else { blockBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public Builder clearBlock() { if (blockBuilder_ == null) { block_ = null; onChanged(); } else { blockBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.Builder getBlockBuilder() { bitField0_ |= 0x00000001; onChanged(); return getBlockFieldBuilder().getBuilder(); } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProtoOrBuilder getBlockOrBuilder() { if (blockBuilder_ != null) { return blockBuilder_.getMessageOrBuilder(); } else { return block_ == null ? org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.getDefaultInstance() : block_; } } /** * required .hadoop.hdfs.ExtendedBlockProto block = 1; */ private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProtoOrBuilder> getBlockFieldBuilder() { if (blockBuilder_ == null) { blockBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.ExtendedBlockProtoOrBuilder>( getBlock(), getParentForChildren(), isClean()); block_ = null; } return blockBuilder_; } private long newGenStamp_ ; /** * required uint64 newGenStamp = 2; */ public boolean hasNewGenStamp() { return ((bitField0_ & 0x00000002) != 0); } /** * required uint64 newGenStamp = 2; */ public long getNewGenStamp() { return newGenStamp_; } /** * required uint64 newGenStamp = 2; */ public Builder setNewGenStamp(long value) { bitField0_ |= 0x00000002; newGenStamp_ = value; onChanged(); return this; } /** * required uint64 newGenStamp = 2; */ public Builder clearNewGenStamp() { bitField0_ = (bitField0_ & ~0x00000002); newGenStamp_ = 0L; onChanged(); return this; } private long newLength_ ; /** * required uint64 newLength = 3; */ public boolean hasNewLength() { return ((bitField0_ & 0x00000004) != 0); } /** * required uint64 newLength = 3; */ public long getNewLength() { return newLength_; } /** * required uint64 newLength = 3; */ public Builder setNewLength(long value) { bitField0_ |= 0x00000004; newLength_ = value; onChanged(); return this; } /** * required uint64 newLength = 3; */ public Builder clearNewLength() { bitField0_ = (bitField0_ & ~0x00000004); newLength_ = 0L; onChanged(); return this; } private boolean closeFile_ ; /** * required bool closeFile = 4; */ public boolean hasCloseFile() { return ((bitField0_ & 0x00000008) != 0); } /** * required bool closeFile = 4; */ public boolean getCloseFile() { return closeFile_; } /** * required bool closeFile = 4; */ public Builder setCloseFile(boolean value) { bitField0_ |= 0x00000008; closeFile_ = value; onChanged(); return this; } /** * required bool closeFile = 4; */ public Builder clearCloseFile() { bitField0_ = (bitField0_ & ~0x00000008); closeFile_ = false; onChanged(); return this; } private boolean deleteBlock_ ; /** * required bool deleteBlock = 5; */ public boolean hasDeleteBlock() { return ((bitField0_ & 0x00000010) != 0); } /** * required bool deleteBlock = 5; */ public boolean getDeleteBlock() { return deleteBlock_; } /** * required bool deleteBlock = 5; */ public Builder setDeleteBlock(boolean value) { bitField0_ |= 0x00000010; deleteBlock_ = value; onChanged(); return this; } /** * required bool deleteBlock = 5; */ public Builder clearDeleteBlock() { bitField0_ = (bitField0_ & ~0x00000010); deleteBlock_ = false; onChanged(); return this; } private java.util.List newTaragets_ = java.util.Collections.emptyList(); private void ensureNewTaragetsIsMutable() { if (!((bitField0_ & 0x00000020) != 0)) { newTaragets_ = new java.util.ArrayList(newTaragets_); bitField0_ |= 0x00000020; } } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder> newTaragetsBuilder_; /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public java.util.List getNewTaragetsList() { if (newTaragetsBuilder_ == null) { return java.util.Collections.unmodifiableList(newTaragets_); } else { return newTaragetsBuilder_.getMessageList(); } } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public int getNewTaragetsCount() { if (newTaragetsBuilder_ == null) { return newTaragets_.size(); } else { return newTaragetsBuilder_.getCount(); } } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto getNewTaragets(int index) { if (newTaragetsBuilder_ == null) { return newTaragets_.get(index); } else { return newTaragetsBuilder_.getMessage(index); } } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder setNewTaragets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto value) { if (newTaragetsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureNewTaragetsIsMutable(); newTaragets_.set(index, value); onChanged(); } else { newTaragetsBuilder_.setMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder setNewTaragets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder builderForValue) { if (newTaragetsBuilder_ == null) { ensureNewTaragetsIsMutable(); newTaragets_.set(index, builderForValue.build()); onChanged(); } else { newTaragetsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder addNewTaragets(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto value) { if (newTaragetsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureNewTaragetsIsMutable(); newTaragets_.add(value); onChanged(); } else { newTaragetsBuilder_.addMessage(value); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder addNewTaragets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto value) { if (newTaragetsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureNewTaragetsIsMutable(); newTaragets_.add(index, value); onChanged(); } else { newTaragetsBuilder_.addMessage(index, value); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder addNewTaragets( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder builderForValue) { if (newTaragetsBuilder_ == null) { ensureNewTaragetsIsMutable(); newTaragets_.add(builderForValue.build()); onChanged(); } else { newTaragetsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder addNewTaragets( int index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder builderForValue) { if (newTaragetsBuilder_ == null) { ensureNewTaragetsIsMutable(); newTaragets_.add(index, builderForValue.build()); onChanged(); } else { newTaragetsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder addAllNewTaragets( java.lang.Iterable values) { if (newTaragetsBuilder_ == null) { ensureNewTaragetsIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, newTaragets_); onChanged(); } else { newTaragetsBuilder_.addAllMessages(values); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder clearNewTaragets() { if (newTaragetsBuilder_ == null) { newTaragets_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); onChanged(); } else { newTaragetsBuilder_.clear(); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public Builder removeNewTaragets(int index) { if (newTaragetsBuilder_ == null) { ensureNewTaragetsIsMutable(); newTaragets_.remove(index); onChanged(); } else { newTaragetsBuilder_.remove(index); } return this; } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder getNewTaragetsBuilder( int index) { return getNewTaragetsFieldBuilder().getBuilder(index); } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder getNewTaragetsOrBuilder( int index) { if (newTaragetsBuilder_ == null) { return newTaragets_.get(index); } else { return newTaragetsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public java.util.List getNewTaragetsOrBuilderList() { if (newTaragetsBuilder_ != null) { return newTaragetsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(newTaragets_); } } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder addNewTaragetsBuilder() { return getNewTaragetsFieldBuilder().addBuilder( org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder addNewTaragetsBuilder( int index) { return getNewTaragetsFieldBuilder().addBuilder( index, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.getDefaultInstance()); } /** * repeated .hadoop.hdfs.DatanodeIDProto newTaragets = 6; */ public java.util.List getNewTaragetsBuilderList() { return getNewTaragetsFieldBuilder().getBuilderList(); } private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder> getNewTaragetsFieldBuilder() { if (newTaragetsBuilder_ == null) { newTaragetsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.DatanodeIDProtoOrBuilder>( newTaragets_, ((bitField0_ & 0x00000020) != 0), getParentForChildren(), isClean()); newTaragets_ = null; } return newTaragetsBuilder_; } private org.apache.hadoop.thirdparty.protobuf.LazyStringList newTargetStorages_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; private void ensureNewTargetStoragesIsMutable() { if (!((bitField0_ & 0x00000040) != 0)) { newTargetStorages_ = new org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList(newTargetStorages_); bitField0_ |= 0x00000040; } } /** * repeated string newTargetStorages = 7; */ public org.apache.hadoop.thirdparty.protobuf.ProtocolStringList getNewTargetStoragesList() { return newTargetStorages_.getUnmodifiableView(); } /** * repeated string newTargetStorages = 7; */ public int getNewTargetStoragesCount() { return newTargetStorages_.size(); } /** * repeated string newTargetStorages = 7; */ public java.lang.String getNewTargetStorages(int index) { return newTargetStorages_.get(index); } /** * repeated string newTargetStorages = 7; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getNewTargetStoragesBytes(int index) { return newTargetStorages_.getByteString(index); } /** * repeated string newTargetStorages = 7; */ public Builder setNewTargetStorages( int index, java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureNewTargetStoragesIsMutable(); newTargetStorages_.set(index, value); onChanged(); return this; } /** * repeated string newTargetStorages = 7; */ public Builder addNewTargetStorages( java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureNewTargetStoragesIsMutable(); newTargetStorages_.add(value); onChanged(); return this; } /** * repeated string newTargetStorages = 7; */ public Builder addAllNewTargetStorages( java.lang.Iterable values) { ensureNewTargetStoragesIsMutable(); org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll( values, newTargetStorages_); onChanged(); return this; } /** * repeated string newTargetStorages = 7; */ public Builder clearNewTargetStorages() { newTargetStorages_ = org.apache.hadoop.thirdparty.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000040); onChanged(); return this; } /** * repeated string newTargetStorages = 7; */ public Builder addNewTargetStoragesBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureNewTargetStoragesIsMutable(); newTargetStorages_.add(value); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public CommitBlockSynchronizationRequestProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new CommitBlockSynchronizationRequestProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CommitBlockSynchronizationResponseProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { } /** *
   **
   * void response
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto} */ public static final class CommitBlockSynchronizationResponseProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto) CommitBlockSynchronizationResponseProtoOrBuilder { private static final long serialVersionUID = 0L; // Use CommitBlockSynchronizationResponseProto.newBuilder() to construct. private CommitBlockSynchronizationResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CommitBlockSynchronizationResponseProto() { } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CommitBlockSynchronizationResponseProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * void response
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public CommitBlockSynchronizationResponseProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new CommitBlockSynchronizationResponseProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SlowPeerReportProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.SlowPeerReportProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * optional string dataNodeId = 1; */ boolean hasDataNodeId(); /** * optional string dataNodeId = 1; */ java.lang.String getDataNodeId(); /** * optional string dataNodeId = 1; */ org.apache.hadoop.thirdparty.protobuf.ByteString getDataNodeIdBytes(); /** * optional double aggregateLatency = 2; */ boolean hasAggregateLatency(); /** * optional double aggregateLatency = 2; */ double getAggregateLatency(); } /** *
   **
   * Information about a single slow peer that may be reported by
   * the DataNode to the NameNode as part of the heartbeat request.
   * The message includes the peer's DataNodeId and its
   * aggregate packet latency as observed by the reporting DataNode.
   * (DataNodeId must be transmitted as a string for protocol compability
   *  with earlier versions of Hadoop).
   * The exact choice of the aggregate is opaque to the NameNode but it
   * _should_ be chosen consistenly by all DataNodes in the cluster.
   * Examples of aggregates are 90th percentile (good) and mean (not so
   * good).
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.SlowPeerReportProto} */ public static final class SlowPeerReportProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.SlowPeerReportProto) SlowPeerReportProtoOrBuilder { private static final long serialVersionUID = 0L; // Use SlowPeerReportProto.newBuilder() to construct. private SlowPeerReportProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SlowPeerReportProto() { dataNodeId_ = ""; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SlowPeerReportProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; dataNodeId_ = bs; break; } case 17: { bitField0_ |= 0x00000002; aggregateLatency_ = input.readDouble(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder.class); } private int bitField0_; public static final int DATANODEID_FIELD_NUMBER = 1; private volatile java.lang.Object dataNodeId_; /** * optional string dataNodeId = 1; */ public boolean hasDataNodeId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string dataNodeId = 1; */ public java.lang.String getDataNodeId() { java.lang.Object ref = dataNodeId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { dataNodeId_ = s; } return s; } } /** * optional string dataNodeId = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getDataNodeIdBytes() { java.lang.Object ref = dataNodeId_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); dataNodeId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int AGGREGATELATENCY_FIELD_NUMBER = 2; private double aggregateLatency_; /** * optional double aggregateLatency = 2; */ public boolean hasAggregateLatency() { return ((bitField0_ & 0x00000002) != 0); } /** * optional double aggregateLatency = 2; */ public double getAggregateLatency() { return aggregateLatency_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, dataNodeId_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeDouble(2, aggregateLatency_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, dataNodeId_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeDoubleSize(2, aggregateLatency_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto) obj; if (hasDataNodeId() != other.hasDataNodeId()) return false; if (hasDataNodeId()) { if (!getDataNodeId() .equals(other.getDataNodeId())) return false; } if (hasAggregateLatency() != other.hasAggregateLatency()) return false; if (hasAggregateLatency()) { if (java.lang.Double.doubleToLongBits(getAggregateLatency()) != java.lang.Double.doubleToLongBits( other.getAggregateLatency())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasDataNodeId()) { hash = (37 * hash) + DATANODEID_FIELD_NUMBER; hash = (53 * hash) + getDataNodeId().hashCode(); } if (hasAggregateLatency()) { hash = (37 * hash) + AGGREGATELATENCY_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( java.lang.Double.doubleToLongBits(getAggregateLatency())); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Information about a single slow peer that may be reported by
     * the DataNode to the NameNode as part of the heartbeat request.
     * The message includes the peer's DataNodeId and its
     * aggregate packet latency as observed by the reporting DataNode.
     * (DataNodeId must be transmitted as a string for protocol compability
     *  with earlier versions of Hadoop).
     * The exact choice of the aggregate is opaque to the NameNode but it
     * _should_ be chosen consistenly by all DataNodes in the cluster.
     * Examples of aggregates are 90th percentile (good) and mean (not so
     * good).
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.SlowPeerReportProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.SlowPeerReportProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); dataNodeId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); aggregateLatency_ = 0D; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.dataNodeId_ = dataNodeId_; if (((from_bitField0_ & 0x00000002) != 0)) { result.aggregateLatency_ = aggregateLatency_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto.getDefaultInstance()) return this; if (other.hasDataNodeId()) { bitField0_ |= 0x00000001; dataNodeId_ = other.dataNodeId_; onChanged(); } if (other.hasAggregateLatency()) { setAggregateLatency(other.getAggregateLatency()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object dataNodeId_ = ""; /** * optional string dataNodeId = 1; */ public boolean hasDataNodeId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string dataNodeId = 1; */ public java.lang.String getDataNodeId() { java.lang.Object ref = dataNodeId_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { dataNodeId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string dataNodeId = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getDataNodeIdBytes() { java.lang.Object ref = dataNodeId_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); dataNodeId_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * optional string dataNodeId = 1; */ public Builder setDataNodeId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; dataNodeId_ = value; onChanged(); return this; } /** * optional string dataNodeId = 1; */ public Builder clearDataNodeId() { bitField0_ = (bitField0_ & ~0x00000001); dataNodeId_ = getDefaultInstance().getDataNodeId(); onChanged(); return this; } /** * optional string dataNodeId = 1; */ public Builder setDataNodeIdBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; dataNodeId_ = value; onChanged(); return this; } private double aggregateLatency_ ; /** * optional double aggregateLatency = 2; */ public boolean hasAggregateLatency() { return ((bitField0_ & 0x00000002) != 0); } /** * optional double aggregateLatency = 2; */ public double getAggregateLatency() { return aggregateLatency_; } /** * optional double aggregateLatency = 2; */ public Builder setAggregateLatency(double value) { bitField0_ |= 0x00000002; aggregateLatency_ = value; onChanged(); return this; } /** * optional double aggregateLatency = 2; */ public Builder clearAggregateLatency() { bitField0_ = (bitField0_ & ~0x00000002); aggregateLatency_ = 0D; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.SlowPeerReportProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.SlowPeerReportProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public SlowPeerReportProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new SlowPeerReportProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowPeerReportProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SlowDiskReportProtoOrBuilder extends // @@protoc_insertion_point(interface_extends:hadoop.hdfs.datanode.SlowDiskReportProto) org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder { /** * optional string basePath = 1; */ boolean hasBasePath(); /** * optional string basePath = 1; */ java.lang.String getBasePath(); /** * optional string basePath = 1; */ org.apache.hadoop.thirdparty.protobuf.ByteString getBasePathBytes(); /** * optional double meanMetadataOpLatency = 2; */ boolean hasMeanMetadataOpLatency(); /** * optional double meanMetadataOpLatency = 2; */ double getMeanMetadataOpLatency(); /** * optional double meanReadIoLatency = 3; */ boolean hasMeanReadIoLatency(); /** * optional double meanReadIoLatency = 3; */ double getMeanReadIoLatency(); /** * optional double meanWriteIoLatency = 4; */ boolean hasMeanWriteIoLatency(); /** * optional double meanWriteIoLatency = 4; */ double getMeanWriteIoLatency(); } /** *
   **
   * Information about a single slow disk that may be reported by
   * the DataNode to the NameNode as part of the heartbeat request.
   * The message includes the disk's basePath, mean metadata op latency,
   * mean read io latency and mean write io latency as observed by the DataNode.
   * 
* * Protobuf type {@code hadoop.hdfs.datanode.SlowDiskReportProto} */ public static final class SlowDiskReportProto extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hadoop.hdfs.datanode.SlowDiskReportProto) SlowDiskReportProtoOrBuilder { private static final long serialVersionUID = 0L; // Use SlowDiskReportProto.newBuilder() to construct. private SlowDiskReportProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SlowDiskReportProto() { basePath_ = ""; } @java.lang.Override public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SlowDiskReportProto( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.thirdparty.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; basePath_ = bs; break; } case 17: { bitField0_ |= 0x00000002; meanMetadataOpLatency_ = input.readDouble(); break; } case 25: { bitField0_ |= 0x00000004; meanReadIoLatency_ = input.readDouble(); break; } case 33: { bitField0_ |= 0x00000008; meanWriteIoLatency_ = input.readDouble(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder.class); } private int bitField0_; public static final int BASEPATH_FIELD_NUMBER = 1; private volatile java.lang.Object basePath_; /** * optional string basePath = 1; */ public boolean hasBasePath() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string basePath = 1; */ public java.lang.String getBasePath() { java.lang.Object ref = basePath_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { basePath_ = s; } return s; } } /** * optional string basePath = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBasePathBytes() { java.lang.Object ref = basePath_; if (ref instanceof java.lang.String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); basePath_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } public static final int MEANMETADATAOPLATENCY_FIELD_NUMBER = 2; private double meanMetadataOpLatency_; /** * optional double meanMetadataOpLatency = 2; */ public boolean hasMeanMetadataOpLatency() { return ((bitField0_ & 0x00000002) != 0); } /** * optional double meanMetadataOpLatency = 2; */ public double getMeanMetadataOpLatency() { return meanMetadataOpLatency_; } public static final int MEANREADIOLATENCY_FIELD_NUMBER = 3; private double meanReadIoLatency_; /** * optional double meanReadIoLatency = 3; */ public boolean hasMeanReadIoLatency() { return ((bitField0_ & 0x00000004) != 0); } /** * optional double meanReadIoLatency = 3; */ public double getMeanReadIoLatency() { return meanReadIoLatency_; } public static final int MEANWRITEIOLATENCY_FIELD_NUMBER = 4; private double meanWriteIoLatency_; /** * optional double meanWriteIoLatency = 4; */ public boolean hasMeanWriteIoLatency() { return ((bitField0_ & 0x00000008) != 0); } /** * optional double meanWriteIoLatency = 4; */ public double getMeanWriteIoLatency() { return meanWriteIoLatency_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, basePath_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeDouble(2, meanMetadataOpLatency_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeDouble(3, meanReadIoLatency_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeDouble(4, meanWriteIoLatency_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, basePath_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeDoubleSize(2, meanMetadataOpLatency_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeDoubleSize(3, meanReadIoLatency_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream .computeDoubleSize(4, meanWriteIoLatency_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto)) { return super.equals(obj); } org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto other = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto) obj; if (hasBasePath() != other.hasBasePath()) return false; if (hasBasePath()) { if (!getBasePath() .equals(other.getBasePath())) return false; } if (hasMeanMetadataOpLatency() != other.hasMeanMetadataOpLatency()) return false; if (hasMeanMetadataOpLatency()) { if (java.lang.Double.doubleToLongBits(getMeanMetadataOpLatency()) != java.lang.Double.doubleToLongBits( other.getMeanMetadataOpLatency())) return false; } if (hasMeanReadIoLatency() != other.hasMeanReadIoLatency()) return false; if (hasMeanReadIoLatency()) { if (java.lang.Double.doubleToLongBits(getMeanReadIoLatency()) != java.lang.Double.doubleToLongBits( other.getMeanReadIoLatency())) return false; } if (hasMeanWriteIoLatency() != other.hasMeanWriteIoLatency()) return false; if (hasMeanWriteIoLatency()) { if (java.lang.Double.doubleToLongBits(getMeanWriteIoLatency()) != java.lang.Double.doubleToLongBits( other.getMeanWriteIoLatency())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasBasePath()) { hash = (37 * hash) + BASEPATH_FIELD_NUMBER; hash = (53 * hash) + getBasePath().hashCode(); } if (hasMeanMetadataOpLatency()) { hash = (37 * hash) + MEANMETADATAOPLATENCY_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( java.lang.Double.doubleToLongBits(getMeanMetadataOpLatency())); } if (hasMeanReadIoLatency()) { hash = (37 * hash) + MEANREADIOLATENCY_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( java.lang.Double.doubleToLongBits(getMeanReadIoLatency())); } if (hasMeanWriteIoLatency()) { hash = (37 * hash) + MEANWRITEIOLATENCY_FIELD_NUMBER; hash = (53 * hash) + org.apache.hadoop.thirdparty.protobuf.Internal.hashLong( java.lang.Double.doubleToLongBits(getMeanWriteIoLatency())); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( java.nio.ByteBuffer data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( java.nio.ByteBuffer data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.ByteString data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom(byte[] data) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( byte[] data, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseDelimitedFrom( java.io.InputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parseFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Information about a single slow disk that may be reported by
     * the DataNode to the NameNode as part of the heartbeat request.
     * The message includes the disk's basePath, mean metadata op latency,
     * mean read io latency and mean write io latency as observed by the DataNode.
     * 
* * Protobuf type {@code hadoop.hdfs.datanode.SlowDiskReportProto} */ public static final class Builder extends org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hadoop.hdfs.datanode.SlowDiskReportProto) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProtoOrBuilder { public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_descriptor; } @java.lang.Override protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.Builder.class); } // Construct using org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); basePath_ = ""; bitField0_ = (bitField0_ & ~0x00000001); meanMetadataOpLatency_ = 0D; bitField0_ = (bitField0_ & ~0x00000002); meanReadIoLatency_ = 0D; bitField0_ = (bitField0_ & ~0x00000004); meanWriteIoLatency_ = 0D; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_descriptor; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto getDefaultInstanceForType() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto build() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto buildPartial() { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto result = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.basePath_ = basePath_; if (((from_bitField0_ & 0x00000002) != 0)) { result.meanMetadataOpLatency_ = meanMetadataOpLatency_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.meanReadIoLatency_ = meanReadIoLatency_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.meanWriteIoLatency_ = meanWriteIoLatency_; to_bitField0_ |= 0x00000008; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) { if (other instanceof org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto) { return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto other) { if (other == org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto.getDefaultInstance()) return this; if (other.hasBasePath()) { bitField0_ |= 0x00000001; basePath_ = other.basePath_; onChanged(); } if (other.hasMeanMetadataOpLatency()) { setMeanMetadataOpLatency(other.getMeanMetadataOpLatency()); } if (other.hasMeanReadIoLatency()) { setMeanReadIoLatency(other.getMeanReadIoLatency()); } if (other.hasMeanWriteIoLatency()) { setMeanWriteIoLatency(other.getMeanWriteIoLatency()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object basePath_ = ""; /** * optional string basePath = 1; */ public boolean hasBasePath() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string basePath = 1; */ public java.lang.String getBasePath() { java.lang.Object ref = basePath_; if (!(ref instanceof java.lang.String)) { org.apache.hadoop.thirdparty.protobuf.ByteString bs = (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { basePath_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string basePath = 1; */ public org.apache.hadoop.thirdparty.protobuf.ByteString getBasePathBytes() { java.lang.Object ref = basePath_; if (ref instanceof String) { org.apache.hadoop.thirdparty.protobuf.ByteString b = org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); basePath_ = b; return b; } else { return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref; } } /** * optional string basePath = 1; */ public Builder setBasePath( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; basePath_ = value; onChanged(); return this; } /** * optional string basePath = 1; */ public Builder clearBasePath() { bitField0_ = (bitField0_ & ~0x00000001); basePath_ = getDefaultInstance().getBasePath(); onChanged(); return this; } /** * optional string basePath = 1; */ public Builder setBasePathBytes( org.apache.hadoop.thirdparty.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; basePath_ = value; onChanged(); return this; } private double meanMetadataOpLatency_ ; /** * optional double meanMetadataOpLatency = 2; */ public boolean hasMeanMetadataOpLatency() { return ((bitField0_ & 0x00000002) != 0); } /** * optional double meanMetadataOpLatency = 2; */ public double getMeanMetadataOpLatency() { return meanMetadataOpLatency_; } /** * optional double meanMetadataOpLatency = 2; */ public Builder setMeanMetadataOpLatency(double value) { bitField0_ |= 0x00000002; meanMetadataOpLatency_ = value; onChanged(); return this; } /** * optional double meanMetadataOpLatency = 2; */ public Builder clearMeanMetadataOpLatency() { bitField0_ = (bitField0_ & ~0x00000002); meanMetadataOpLatency_ = 0D; onChanged(); return this; } private double meanReadIoLatency_ ; /** * optional double meanReadIoLatency = 3; */ public boolean hasMeanReadIoLatency() { return ((bitField0_ & 0x00000004) != 0); } /** * optional double meanReadIoLatency = 3; */ public double getMeanReadIoLatency() { return meanReadIoLatency_; } /** * optional double meanReadIoLatency = 3; */ public Builder setMeanReadIoLatency(double value) { bitField0_ |= 0x00000004; meanReadIoLatency_ = value; onChanged(); return this; } /** * optional double meanReadIoLatency = 3; */ public Builder clearMeanReadIoLatency() { bitField0_ = (bitField0_ & ~0x00000004); meanReadIoLatency_ = 0D; onChanged(); return this; } private double meanWriteIoLatency_ ; /** * optional double meanWriteIoLatency = 4; */ public boolean hasMeanWriteIoLatency() { return ((bitField0_ & 0x00000008) != 0); } /** * optional double meanWriteIoLatency = 4; */ public double getMeanWriteIoLatency() { return meanWriteIoLatency_; } /** * optional double meanWriteIoLatency = 4; */ public Builder setMeanWriteIoLatency(double value) { bitField0_ |= 0x00000008; meanWriteIoLatency_ = value; onChanged(); return this; } /** * optional double meanWriteIoLatency = 4; */ public Builder clearMeanWriteIoLatency() { bitField0_ = (bitField0_ & ~0x00000008); meanWriteIoLatency_ = 0D; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hadoop.hdfs.datanode.SlowDiskReportProto) } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.SlowDiskReportProto) private static final org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto(); } public static org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() { @java.lang.Override public SlowDiskReportProto parsePartialFrom( org.apache.hadoop.thirdparty.protobuf.CodedInputStream input, org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException { return new SlowDiskReportProto(input, extensionRegistry); } }; public static org.apache.hadoop.thirdparty.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.SlowDiskReportProto getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } /** *
   **
   * Protocol used from datanode to the namenode
   * See the request and response for details of rpc call.
   * 
* * Protobuf service {@code hadoop.hdfs.datanode.DatanodeProtocolService} */ public static abstract class DatanodeProtocolService implements org.apache.hadoop.thirdparty.protobuf.Service { protected DatanodeProtocolService() {} public interface Interface { /** *
       **
       * Register a datanode at a namenode
       * 
* * rpc registerDatanode(.hadoop.hdfs.datanode.RegisterDatanodeRequestProto) returns (.hadoop.hdfs.datanode.RegisterDatanodeResponseProto); */ public abstract void registerDatanode( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Send heartbeat from datanode to namenode
       * 
* * rpc sendHeartbeat(.hadoop.hdfs.datanode.HeartbeatRequestProto) returns (.hadoop.hdfs.datanode.HeartbeatResponseProto); */ public abstract void sendHeartbeat( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Report blocks at a given datanode to the namenode
       * 
* * rpc blockReport(.hadoop.hdfs.datanode.BlockReportRequestProto) returns (.hadoop.hdfs.datanode.BlockReportResponseProto); */ public abstract void blockReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Report cached blocks at a datanode to the namenode
       * 
* * rpc cacheReport(.hadoop.hdfs.datanode.CacheReportRequestProto) returns (.hadoop.hdfs.datanode.CacheReportResponseProto); */ public abstract void cacheReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Incremental block report from the DN. This contains info about recently
       * received and deleted blocks, as well as when blocks start being
       * received.
       * 
* * rpc blockReceivedAndDeleted(.hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) returns (.hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto); */ public abstract void blockReceivedAndDeleted( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Report from a datanode of an error to the active namenode.
       * Used for debugging.
       * 
* * rpc errorReport(.hadoop.hdfs.datanode.ErrorReportRequestProto) returns (.hadoop.hdfs.datanode.ErrorReportResponseProto); */ public abstract void errorReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Request the version
       * 
* * rpc versionRequest(.hadoop.hdfs.VersionRequestProto) returns (.hadoop.hdfs.VersionResponseProto); */ public abstract void versionRequest( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Report corrupt blocks at the specified location
       * 
* * rpc reportBadBlocks(.hadoop.hdfs.datanode.ReportBadBlocksRequestProto) returns (.hadoop.hdfs.datanode.ReportBadBlocksResponseProto); */ public abstract void reportBadBlocks( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
       **
       * Commit block synchronization during lease recovery.
       * 
* * rpc commitBlockSynchronization(.hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) returns (.hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto); */ public abstract void commitBlockSynchronization( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); } public static org.apache.hadoop.thirdparty.protobuf.Service newReflectiveService( final Interface impl) { return new DatanodeProtocolService() { @java.lang.Override public void registerDatanode( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.registerDatanode(controller, request, done); } @java.lang.Override public void sendHeartbeat( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.sendHeartbeat(controller, request, done); } @java.lang.Override public void blockReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.blockReport(controller, request, done); } @java.lang.Override public void cacheReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.cacheReport(controller, request, done); } @java.lang.Override public void blockReceivedAndDeleted( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.blockReceivedAndDeleted(controller, request, done); } @java.lang.Override public void errorReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.errorReport(controller, request, done); } @java.lang.Override public void versionRequest( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.versionRequest(controller, request, done); } @java.lang.Override public void reportBadBlocks( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.reportBadBlocks(controller, request, done); } @java.lang.Override public void commitBlockSynchronization( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { impl.commitBlockSynchronization(controller, request, done); } }; } public static org.apache.hadoop.thirdparty.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { return new org.apache.hadoop.thirdparty.protobuf.BlockingService() { public final org.apache.hadoop.thirdparty.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final org.apache.hadoop.thirdparty.protobuf.Message callBlockingMethod( org.apache.hadoop.thirdparty.protobuf.Descriptors.MethodDescriptor method, org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.thirdparty.protobuf.Message request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + "wrong service type."); } switch(method.getIndex()) { case 0: return impl.registerDatanode(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto)request); case 1: return impl.sendHeartbeat(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto)request); case 2: return impl.blockReport(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto)request); case 3: return impl.cacheReport(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto)request); case 4: return impl.blockReceivedAndDeleted(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto)request); case 5: return impl.errorReport(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto)request); case 6: return impl.versionRequest(controller, (org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto)request); case 7: return impl.reportBadBlocks(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto)request); case 8: return impl.commitBlockSynchronization(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto)request); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hadoop.thirdparty.protobuf.Message getRequestPrototype( org.apache.hadoop.thirdparty.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.getDefaultInstance(); case 1: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.getDefaultInstance(); case 2: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.getDefaultInstance(); case 3: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.getDefaultInstance(); case 4: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.getDefaultInstance(); case 5: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.getDefaultInstance(); case 6: return org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto.getDefaultInstance(); case 7: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.getDefaultInstance(); case 8: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hadoop.thirdparty.protobuf.Message getResponsePrototype( org.apache.hadoop.thirdparty.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance(); case 1: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance(); case 2: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance(); case 3: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance(); case 4: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance(); case 5: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance(); case 6: return org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto.getDefaultInstance(); case 7: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance(); case 8: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } }; } /** *
     **
     * Register a datanode at a namenode
     * 
* * rpc registerDatanode(.hadoop.hdfs.datanode.RegisterDatanodeRequestProto) returns (.hadoop.hdfs.datanode.RegisterDatanodeResponseProto); */ public abstract void registerDatanode( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Send heartbeat from datanode to namenode
     * 
* * rpc sendHeartbeat(.hadoop.hdfs.datanode.HeartbeatRequestProto) returns (.hadoop.hdfs.datanode.HeartbeatResponseProto); */ public abstract void sendHeartbeat( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Report blocks at a given datanode to the namenode
     * 
* * rpc blockReport(.hadoop.hdfs.datanode.BlockReportRequestProto) returns (.hadoop.hdfs.datanode.BlockReportResponseProto); */ public abstract void blockReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Report cached blocks at a datanode to the namenode
     * 
* * rpc cacheReport(.hadoop.hdfs.datanode.CacheReportRequestProto) returns (.hadoop.hdfs.datanode.CacheReportResponseProto); */ public abstract void cacheReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Incremental block report from the DN. This contains info about recently
     * received and deleted blocks, as well as when blocks start being
     * received.
     * 
* * rpc blockReceivedAndDeleted(.hadoop.hdfs.datanode.BlockReceivedAndDeletedRequestProto) returns (.hadoop.hdfs.datanode.BlockReceivedAndDeletedResponseProto); */ public abstract void blockReceivedAndDeleted( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Report from a datanode of an error to the active namenode.
     * Used for debugging.
     * 
* * rpc errorReport(.hadoop.hdfs.datanode.ErrorReportRequestProto) returns (.hadoop.hdfs.datanode.ErrorReportResponseProto); */ public abstract void errorReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Request the version
     * 
* * rpc versionRequest(.hadoop.hdfs.VersionRequestProto) returns (.hadoop.hdfs.VersionResponseProto); */ public abstract void versionRequest( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Report corrupt blocks at the specified location
     * 
* * rpc reportBadBlocks(.hadoop.hdfs.datanode.ReportBadBlocksRequestProto) returns (.hadoop.hdfs.datanode.ReportBadBlocksResponseProto); */ public abstract void reportBadBlocks( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); /** *
     **
     * Commit block synchronization during lease recovery.
     * 
* * rpc commitBlockSynchronization(.hadoop.hdfs.datanode.CommitBlockSynchronizationRequestProto) returns (.hadoop.hdfs.datanode.CommitBlockSynchronizationResponseProto); */ public abstract void commitBlockSynchronization( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done); public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.getDescriptor().getServices().get(0); } public final org.apache.hadoop.thirdparty.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( org.apache.hadoop.thirdparty.protobuf.Descriptors.MethodDescriptor method, org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.thirdparty.protobuf.Message request, org.apache.hadoop.thirdparty.protobuf.RpcCallback< org.apache.hadoop.thirdparty.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + "service type."); } switch(method.getIndex()) { case 0: this.registerDatanode(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.sendHeartbeat(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 2: this.blockReport(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 3: this.cacheReport(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 4: this.blockReceivedAndDeleted(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 5: this.errorReport(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 6: this.versionRequest(controller, (org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 7: this.reportBadBlocks(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; case 8: this.commitBlockSynchronization(controller, (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto)request, org.apache.hadoop.thirdparty.protobuf.RpcUtil.specializeCallback( done)); return; default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hadoop.thirdparty.protobuf.Message getRequestPrototype( org.apache.hadoop.thirdparty.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto.getDefaultInstance(); case 1: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto.getDefaultInstance(); case 2: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto.getDefaultInstance(); case 3: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto.getDefaultInstance(); case 4: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto.getDefaultInstance(); case 5: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto.getDefaultInstance(); case 6: return org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto.getDefaultInstance(); case 7: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto.getDefaultInstance(); case 8: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hadoop.thirdparty.protobuf.Message getResponsePrototype( org.apache.hadoop.thirdparty.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance(); case 1: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance(); case 2: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance(); case 3: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance(); case 4: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance(); case 5: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance(); case 6: return org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto.getDefaultInstance(); case 7: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance(); case 8: return org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public static Stub newStub( org.apache.hadoop.thirdparty.protobuf.RpcChannel channel) { return new Stub(channel); } public static final class Stub extends org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.DatanodeProtocolService implements Interface { private Stub(org.apache.hadoop.thirdparty.protobuf.RpcChannel channel) { this.channel = channel; } private final org.apache.hadoop.thirdparty.protobuf.RpcChannel channel; public org.apache.hadoop.thirdparty.protobuf.RpcChannel getChannel() { return channel; } public void registerDatanode( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance())); } public void sendHeartbeat( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance())); } public void blockReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance())); } public void cacheReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance())); } public void blockReceivedAndDeleted( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance())); } public void errorReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(5), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance())); } public void versionRequest( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(6), controller, request, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto.getDefaultInstance())); } public void reportBadBlocks( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(7), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance())); } public void commitBlockSynchronization( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto request, org.apache.hadoop.thirdparty.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance(), org.apache.hadoop.thirdparty.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance())); } } public static BlockingInterface newBlockingStub( org.apache.hadoop.thirdparty.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto registerDatanode( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto sendHeartbeat( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto blockReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto cacheReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto blockReceivedAndDeleted( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto errorReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto versionRequest( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto reportBadBlocks( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto commitBlockSynchronization( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { private BlockingStub(org.apache.hadoop.thirdparty.protobuf.BlockingRpcChannel channel) { this.channel = channel; } private final org.apache.hadoop.thirdparty.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto registerDatanode( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.RegisterDatanodeResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto sendHeartbeat( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.HeartbeatResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto blockReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReportResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto cacheReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CacheReportResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto blockReceivedAndDeleted( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.BlockReceivedAndDeletedResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto errorReport( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(5), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ErrorReportResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto versionRequest( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(6), controller, request, org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.VersionResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto reportBadBlocks( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(7), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.ReportBadBlocksResponseProto.getDefaultInstance()); } public org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto commitBlockSynchronization( org.apache.hadoop.thirdparty.protobuf.RpcController controller, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationRequestProto request) throws org.apache.hadoop.thirdparty.protobuf.ServiceException { return (org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto) channel.callBlockingMethod( getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos.CommitBlockSynchronizationResponseProto.getDefaultInstance()); } } // @@protoc_insertion_point(class_scope:hadoop.hdfs.datanode.DatanodeProtocolService) } private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_RegisterCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_RegisterCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockReportContextProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockReportContextProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_fieldAccessorTable; private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_descriptor; private static final org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_fieldAccessorTable; public static org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } private static org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { "\n\026DatanodeProtocol.proto\022\024hadoop.hdfs.da" + "tanode\032\nhdfs.proto\032\023erasurecoding.proto\032" + "\020HdfsServer.proto\"\315\001\n\031DatanodeRegistrati" + "onProto\0220\n\ndatanodeID\030\001 \002(\0132\034.hadoop.hdf" + "s.DatanodeIDProto\0222\n\013storageInfo\030\002 \002(\0132\035" + ".hadoop.hdfs.StorageInfoProto\0221\n\004keys\030\003 " + "\002(\0132#.hadoop.hdfs.ExportedBlockKeysProto" + "\022\027\n\017softwareVersion\030\004 \002(\t\"\370\006\n\024DatanodeCo" + "mmandProto\022@\n\007cmdType\030\001 \002(\0162/.hadoop.hdf" + "s.datanode.DatanodeCommandProto.Type\022H\n\013" + "balancerCmd\030\002 \001(\01323.hadoop.hdfs.datanode" + ".BalancerBandwidthCommandProto\0227\n\006blkCmd" + "\030\003 \001(\0132\'.hadoop.hdfs.datanode.BlockComma" + "ndProto\022D\n\013recoveryCmd\030\004 \001(\0132/.hadoop.hd" + "fs.datanode.BlockRecoveryCommandProto\022?\n" + "\013finalizeCmd\030\005 \001(\0132*.hadoop.hdfs.datanod" + "e.FinalizeCommandProto\022A\n\014keyUpdateCmd\030\006" + " \001(\0132+.hadoop.hdfs.datanode.KeyUpdateCom" + "mandProto\022?\n\013registerCmd\030\007 \001(\0132*.hadoop." + "hdfs.datanode.RegisterCommandProto\022;\n\010bl" + "kIdCmd\030\010 \001(\0132).hadoop.hdfs.datanode.Bloc" + "kIdCommandProto\022W\n\026blkECReconstructionCm" + "d\030\t \001(\01327.hadoop.hdfs.datanode.BlockECRe" + "constructionCommandProto\"\371\001\n\004Type\022\034\n\030Bal" + "ancerBandwidthCommand\020\000\022\020\n\014BlockCommand\020" + "\001\022\030\n\024BlockRecoveryCommand\020\002\022\023\n\017FinalizeC" + "ommand\020\003\022\024\n\020KeyUpdateCommand\020\004\022\023\n\017Regist" + "erCommand\020\005\022\030\n\024UnusedUpgradeCommand\020\006\022\027\n" + "\023NullDatanodeCommand\020\007\022\022\n\016BlockIdCommand" + "\020\010\022 \n\034BlockECReconstructionCommand\020\t\"2\n\035" + "BalancerBandwidthCommandProto\022\021\n\tbandwid" + "th\030\001 \002(\004\"\361\002\n\021BlockCommandProto\022>\n\006action" + "\030\001 \002(\0162..hadoop.hdfs.datanode.BlockComma" + "ndProto.Action\022\023\n\013blockPoolId\030\002 \002(\t\022\'\n\006b" + "locks\030\003 \003(\0132\027.hadoop.hdfs.BlockProto\0220\n\007" + "targets\030\004 \003(\0132\037.hadoop.hdfs.DatanodeInfo" + "sProto\022:\n\022targetStorageUuids\030\005 \003(\0132\036.had" + "oop.hdfs.StorageUuidsProto\022:\n\022targetStor" + "ageTypes\030\006 \003(\0132\036.hadoop.hdfs.StorageType" + "sProto\"4\n\006Action\022\014\n\010TRANSFER\020\001\022\016\n\nINVALI" + "DATE\020\002\022\014\n\010SHUTDOWN\020\003\"\244\001\n\023BlockIdCommandP" + "roto\022@\n\006action\030\001 \002(\01620.hadoop.hdfs.datan" + "ode.BlockIdCommandProto.Action\022\023\n\013blockP" + "oolId\030\002 \002(\t\022\024\n\010blockIds\030\003 \003(\004B\002\020\001\" \n\006Act" + "ion\022\t\n\005CACHE\020\001\022\013\n\007UNCACHE\020\002\"N\n\031BlockReco" + "veryCommandProto\0221\n\006blocks\030\001 \003(\0132!.hadoo" + "p.hdfs.RecoveringBlockProto\"+\n\024FinalizeC" + "ommandProto\022\023\n\013blockPoolId\030\001 \002(\t\"J\n\025KeyU" + "pdateCommandProto\0221\n\004keys\030\001 \002(\0132#.hadoop" + ".hdfs.ExportedBlockKeysProto\"\026\n\024Register" + "CommandProto\"s\n!BlockECReconstructionCom" + "mandProto\022N\n\031blockECReconstructioninfo\030\001" + " \003(\0132+.hadoop.hdfs.BlockECReconstruction" + "InfoProto\"e\n\034RegisterDatanodeRequestProt" + "o\022E\n\014registration\030\001 \002(\0132/.hadoop.hdfs.da" + "tanode.DatanodeRegistrationProto\"f\n\035Regi" + "sterDatanodeResponseProto\022E\n\014registratio" + "n\030\001 \002(\0132/.hadoop.hdfs.datanode.DatanodeR" + "egistrationProto\"~\n\031VolumeFailureSummary" + "Proto\022\036\n\026failedStorageLocations\030\001 \003(\t\022\035\n" + "\025lastVolumeFailureDate\030\002 \002(\004\022\"\n\032estimate" + "dCapacityLostTotal\030\003 \002(\004\"\206\004\n\025HeartbeatRe" + "questProto\022E\n\014registration\030\001 \002(\0132/.hadoo" + "p.hdfs.datanode.DatanodeRegistrationProt" + "o\0220\n\007reports\030\002 \003(\0132\037.hadoop.hdfs.Storage" + "ReportProto\022\032\n\017xmitsInProgress\030\003 \001(\r:\0010\022" + "\027\n\014xceiverCount\030\004 \001(\r:\0010\022\030\n\rfailedVolume" + "s\030\005 \001(\r:\0010\022\030\n\rcacheCapacity\030\006 \001(\004:\0010\022\024\n\t" + "cacheUsed\030\007 \001(\004:\0010\022M\n\024volumeFailureSumma" + "ry\030\010 \001(\0132/.hadoop.hdfs.datanode.VolumeFa" + "ilureSummaryProto\022*\n\033requestFullBlockRep" + "ortLease\030\t \001(\010:\005false\022<\n\tslowPeers\030\n \003(\013" + "2).hadoop.hdfs.datanode.SlowPeerReportPr" + "oto\022<\n\tslowDisks\030\013 \003(\0132).hadoop.hdfs.dat" + "anode.SlowDiskReportProto\"\274\002\n\026HeartbeatR" + "esponseProto\0228\n\004cmds\030\001 \003(\0132*.hadoop.hdfs" + ".datanode.DatanodeCommandProto\0227\n\010haStat" + "us\030\002 \002(\0132%.hadoop.hdfs.NNHAStatusHeartbe" + "atProto\022D\n\024rollingUpgradeStatus\030\003 \001(\0132&." + "hadoop.hdfs.RollingUpgradeStatusProto\022F\n" + "\026rollingUpgradeStatusV2\030\004 \001(\0132&.hadoop.h" + "dfs.RollingUpgradeStatusProto\022!\n\026fullBlo" + "ckReportLeaseId\030\005 \001(\004:\0010\"\365\001\n\027BlockReport" + "RequestProto\022E\n\014registration\030\001 \002(\0132/.had" + "oop.hdfs.datanode.DatanodeRegistrationPr" + "oto\022\023\n\013blockPoolId\030\002 \002(\t\022>\n\007reports\030\003 \003(" + "\0132-.hadoop.hdfs.datanode.StorageBlockRep" + "ortProto\022>\n\007context\030\004 \001(\0132-.hadoop.hdfs." + "datanode.BlockReportContextProto\"s\n\027Bloc" + "kReportContextProto\022\021\n\ttotalRpcs\030\001 \002(\005\022\016" + "\n\006curRpc\030\002 \002(\005\022\n\n\002id\030\003 \002(\003\022\022\n\007leaseId\030\004 " + "\001(\004:\0010\022\025\n\006sorted\030\005 \001(\010:\005false\"\220\001\n\027Storag" + "eBlockReportProto\0222\n\007storage\030\001 \002(\0132!.had" + "oop.hdfs.DatanodeStorageProto\022\022\n\006blocks\030" + "\002 \003(\004B\002\020\001\022\026\n\016numberOfBlocks\030\003 \001(\004\022\025\n\rblo" + "cksBuffers\030\004 \003(\014\"S\n\030BlockReportResponseP" + "roto\0227\n\003cmd\030\001 \001(\0132*.hadoop.hdfs.datanode" + ".DatanodeCommandProto\"\211\001\n\027CacheReportReq" + "uestProto\022E\n\014registration\030\001 \002(\0132/.hadoop" + ".hdfs.datanode.DatanodeRegistrationProto" + "\022\023\n\013blockPoolId\030\002 \002(\t\022\022\n\006blocks\030\003 \003(\004B\002\020" + "\001\"S\n\030CacheReportResponseProto\0227\n\003cmd\030\001 \001" + "(\0132*.hadoop.hdfs.datanode.DatanodeComman" + "dProto\"\345\001\n\035ReceivedDeletedBlockInfoProto" + "\022&\n\005block\030\001 \002(\0132\027.hadoop.hdfs.BlockProto" + "\022O\n\006status\030\003 \002(\0162?.hadoop.hdfs.datanode." + "ReceivedDeletedBlockInfoProto.BlockStatu" + "s\022\022\n\ndeleteHint\030\002 \001(\t\"7\n\013BlockStatus\022\r\n\t" + "RECEIVING\020\001\022\014\n\010RECEIVED\020\002\022\013\n\007DELETED\020\003\"\265" + "\001\n!StorageReceivedDeletedBlocksProto\022\027\n\013" + "storageUuid\030\001 \002(\tB\002\030\001\022C\n\006blocks\030\002 \003(\01323." + "hadoop.hdfs.datanode.ReceivedDeletedBloc" + "kInfoProto\0222\n\007storage\030\003 \001(\0132!.hadoop.hdf" + "s.DatanodeStorageProto\"\312\001\n#BlockReceived" + "AndDeletedRequestProto\022E\n\014registration\030\001" + " \002(\0132/.hadoop.hdfs.datanode.DatanodeRegi" + "strationProto\022\023\n\013blockPoolId\030\002 \002(\t\022G\n\006bl" + "ocks\030\003 \003(\01327.hadoop.hdfs.datanode.Storag" + "eReceivedDeletedBlocksProto\"&\n$BlockRece" + "ivedAndDeletedResponseProto\"\322\001\n\027ErrorRep" + "ortRequestProto\022E\n\014registartion\030\001 \002(\0132/." + "hadoop.hdfs.datanode.DatanodeRegistratio" + "nProto\022\021\n\terrorCode\030\002 \002(\r\022\013\n\003msg\030\003 \002(\t\"P" + "\n\tErrorCode\022\n\n\006NOTIFY\020\000\022\016\n\nDISK_ERROR\020\001\022" + "\021\n\rINVALID_BLOCK\020\002\022\024\n\020FATAL_DISK_ERROR\020\003" + "\"\032\n\030ErrorReportResponseProto\"M\n\033ReportBa" + "dBlocksRequestProto\022.\n\006blocks\030\001 \003(\0132\036.ha" + "doop.hdfs.LocatedBlockProto\"\036\n\034ReportBad" + "BlocksResponseProto\"\366\001\n&CommitBlockSynch" + "ronizationRequestProto\022.\n\005block\030\001 \002(\0132\037." + "hadoop.hdfs.ExtendedBlockProto\022\023\n\013newGen" + "Stamp\030\002 \002(\004\022\021\n\tnewLength\030\003 \002(\004\022\021\n\tcloseF" + "ile\030\004 \002(\010\022\023\n\013deleteBlock\030\005 \002(\010\0221\n\013newTar" + "agets\030\006 \003(\0132\034.hadoop.hdfs.DatanodeIDProt" + "o\022\031\n\021newTargetStorages\030\007 \003(\t\")\n\'CommitBl" + "ockSynchronizationResponseProto\"C\n\023SlowP" + "eerReportProto\022\022\n\ndataNodeId\030\001 \001(\t\022\030\n\020ag" + "gregateLatency\030\002 \001(\001\"}\n\023SlowDiskReportPr" + "oto\022\020\n\010basePath\030\001 \001(\t\022\035\n\025meanMetadataOpL" + "atency\030\002 \001(\001\022\031\n\021meanReadIoLatency\030\003 \001(\001\022" + "\032\n\022meanWriteIoLatency\030\004 \001(\0012\314\010\n\027Datanode" + "ProtocolService\022{\n\020registerDatanode\0222.ha" + "doop.hdfs.datanode.RegisterDatanodeReque" + "stProto\0323.hadoop.hdfs.datanode.RegisterD" + "atanodeResponseProto\022j\n\rsendHeartbeat\022+." + "hadoop.hdfs.datanode.HeartbeatRequestPro" + "to\032,.hadoop.hdfs.datanode.HeartbeatRespo" + "nseProto\022l\n\013blockReport\022-.hadoop.hdfs.da" + "tanode.BlockReportRequestProto\032..hadoop." + "hdfs.datanode.BlockReportResponseProto\022l" + "\n\013cacheReport\022-.hadoop.hdfs.datanode.Cac" + "heReportRequestProto\032..hadoop.hdfs.datan" + "ode.CacheReportResponseProto\022\220\001\n\027blockRe" + "ceivedAndDeleted\0229.hadoop.hdfs.datanode." + "BlockReceivedAndDeletedRequestProto\032:.ha" + "doop.hdfs.datanode.BlockReceivedAndDelet" + "edResponseProto\022l\n\013errorReport\022-.hadoop." + "hdfs.datanode.ErrorReportRequestProto\032.." + "hadoop.hdfs.datanode.ErrorReportResponse" + "Proto\022U\n\016versionRequest\022 .hadoop.hdfs.Ve" + "rsionRequestProto\032!.hadoop.hdfs.VersionR" + "esponseProto\022x\n\017reportBadBlocks\0221.hadoop" + ".hdfs.datanode.ReportBadBlocksRequestPro" + "to\0322.hadoop.hdfs.datanode.ReportBadBlock" + "sResponseProto\022\231\001\n\032commitBlockSynchroniz" + "ation\022<.hadoop.hdfs.datanode.CommitBlock" + "SynchronizationRequestProto\032=.hadoop.hdf" + "s.datanode.CommitBlockSynchronizationRes" + "ponseProtoBE\n%org.apache.hadoop.hdfs.pro" + "tocol.protoB\026DatanodeProtocolProtos\210\001\001\240\001" + "\001" }; org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() { public org.apache.hadoop.thirdparty.protobuf.ExtensionRegistry assignDescriptors( org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor root) { descriptor = root; return null; } }; org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(), org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.getDescriptor(), org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.getDescriptor(), }, assigner); internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_DatanodeRegistrationProto_descriptor, new java.lang.String[] { "DatanodeID", "StorageInfo", "Keys", "SoftwareVersion", }); internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_DatanodeCommandProto_descriptor, new java.lang.String[] { "CmdType", "BalancerCmd", "BlkCmd", "RecoveryCmd", "FinalizeCmd", "KeyUpdateCmd", "RegisterCmd", "BlkIdCmd", "BlkECReconstructionCmd", }); internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BalancerBandwidthCommandProto_descriptor, new java.lang.String[] { "Bandwidth", }); internal_static_hadoop_hdfs_datanode_BlockCommandProto_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_hadoop_hdfs_datanode_BlockCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockCommandProto_descriptor, new java.lang.String[] { "Action", "BlockPoolId", "Blocks", "Targets", "TargetStorageUuids", "TargetStorageTypes", }); internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockIdCommandProto_descriptor, new java.lang.String[] { "Action", "BlockPoolId", "BlockIds", }); internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockRecoveryCommandProto_descriptor, new java.lang.String[] { "Blocks", }); internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_FinalizeCommandProto_descriptor, new java.lang.String[] { "BlockPoolId", }); internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_KeyUpdateCommandProto_descriptor, new java.lang.String[] { "Keys", }); internal_static_hadoop_hdfs_datanode_RegisterCommandProto_descriptor = getDescriptor().getMessageTypes().get(8); internal_static_hadoop_hdfs_datanode_RegisterCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_RegisterCommandProto_descriptor, new java.lang.String[] { }); internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_descriptor = getDescriptor().getMessageTypes().get(9); internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockECReconstructionCommandProto_descriptor, new java.lang.String[] { "BlockECReconstructioninfo", }); internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_descriptor = getDescriptor().getMessageTypes().get(10); internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_RegisterDatanodeRequestProto_descriptor, new java.lang.String[] { "Registration", }); internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_descriptor = getDescriptor().getMessageTypes().get(11); internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_RegisterDatanodeResponseProto_descriptor, new java.lang.String[] { "Registration", }); internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_descriptor = getDescriptor().getMessageTypes().get(12); internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_VolumeFailureSummaryProto_descriptor, new java.lang.String[] { "FailedStorageLocations", "LastVolumeFailureDate", "EstimatedCapacityLostTotal", }); internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_descriptor = getDescriptor().getMessageTypes().get(13); internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_HeartbeatRequestProto_descriptor, new java.lang.String[] { "Registration", "Reports", "XmitsInProgress", "XceiverCount", "FailedVolumes", "CacheCapacity", "CacheUsed", "VolumeFailureSummary", "RequestFullBlockReportLease", "SlowPeers", "SlowDisks", }); internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_descriptor = getDescriptor().getMessageTypes().get(14); internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_HeartbeatResponseProto_descriptor, new java.lang.String[] { "Cmds", "HaStatus", "RollingUpgradeStatus", "RollingUpgradeStatusV2", "FullBlockReportLeaseId", }); internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_descriptor = getDescriptor().getMessageTypes().get(15); internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockReportRequestProto_descriptor, new java.lang.String[] { "Registration", "BlockPoolId", "Reports", "Context", }); internal_static_hadoop_hdfs_datanode_BlockReportContextProto_descriptor = getDescriptor().getMessageTypes().get(16); internal_static_hadoop_hdfs_datanode_BlockReportContextProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockReportContextProto_descriptor, new java.lang.String[] { "TotalRpcs", "CurRpc", "Id", "LeaseId", "Sorted", }); internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_descriptor = getDescriptor().getMessageTypes().get(17); internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_StorageBlockReportProto_descriptor, new java.lang.String[] { "Storage", "Blocks", "NumberOfBlocks", "BlocksBuffers", }); internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_descriptor = getDescriptor().getMessageTypes().get(18); internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockReportResponseProto_descriptor, new java.lang.String[] { "Cmd", }); internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_descriptor = getDescriptor().getMessageTypes().get(19); internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_CacheReportRequestProto_descriptor, new java.lang.String[] { "Registration", "BlockPoolId", "Blocks", }); internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_descriptor = getDescriptor().getMessageTypes().get(20); internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_CacheReportResponseProto_descriptor, new java.lang.String[] { "Cmd", }); internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_descriptor = getDescriptor().getMessageTypes().get(21); internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_ReceivedDeletedBlockInfoProto_descriptor, new java.lang.String[] { "Block", "Status", "DeleteHint", }); internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_descriptor = getDescriptor().getMessageTypes().get(22); internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_StorageReceivedDeletedBlocksProto_descriptor, new java.lang.String[] { "StorageUuid", "Blocks", "Storage", }); internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_descriptor = getDescriptor().getMessageTypes().get(23); internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedRequestProto_descriptor, new java.lang.String[] { "Registration", "BlockPoolId", "Blocks", }); internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_descriptor = getDescriptor().getMessageTypes().get(24); internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_BlockReceivedAndDeletedResponseProto_descriptor, new java.lang.String[] { }); internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_descriptor = getDescriptor().getMessageTypes().get(25); internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_ErrorReportRequestProto_descriptor, new java.lang.String[] { "Registartion", "ErrorCode", "Msg", }); internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_descriptor = getDescriptor().getMessageTypes().get(26); internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_ErrorReportResponseProto_descriptor, new java.lang.String[] { }); internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_descriptor = getDescriptor().getMessageTypes().get(27); internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_ReportBadBlocksRequestProto_descriptor, new java.lang.String[] { "Blocks", }); internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_descriptor = getDescriptor().getMessageTypes().get(28); internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_ReportBadBlocksResponseProto_descriptor, new java.lang.String[] { }); internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_descriptor = getDescriptor().getMessageTypes().get(29); internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationRequestProto_descriptor, new java.lang.String[] { "Block", "NewGenStamp", "NewLength", "CloseFile", "DeleteBlock", "NewTaragets", "NewTargetStorages", }); internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_descriptor = getDescriptor().getMessageTypes().get(30); internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_CommitBlockSynchronizationResponseProto_descriptor, new java.lang.String[] { }); internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_descriptor = getDescriptor().getMessageTypes().get(31); internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_SlowPeerReportProto_descriptor, new java.lang.String[] { "DataNodeId", "AggregateLatency", }); internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_descriptor = getDescriptor().getMessageTypes().get(32); internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_fieldAccessorTable = new org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hadoop_hdfs_datanode_SlowDiskReportProto_descriptor, new java.lang.String[] { "BasePath", "MeanMetadataOpLatency", "MeanReadIoLatency", "MeanWriteIoLatency", }); org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(); org.apache.hadoop.hdfs.protocol.proto.ErasureCodingProtos.getDescriptor(); org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.getDescriptor(); } // @@protoc_insertion_point(outer_class_scope) }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy